INSERT Statement # INSERT statements are used to add rows to a table. Run an INSERT statement # Java Single INSERT statement can be executed through the executeSql() method of the TableEnvironment. The executeSql() method for INSERT statement will submit a Flink job immediately, and return a TableResult instance which associates the submitted job. Multiple INSERT statements can be executed through the addInsertSql() method of the StatementSet which can be created by the TableEnvironment.| nightlies.apache.org
Apache Kafka SQL Connector # Scan Source: Unbounded Sink: Streaming Append Mode The Kafka connector allows for reading data from and writing data into Kafka topics. Dependencies # Only available for stable versions. The Kafka connector is not part of the binary distribution. See how to link with it for cluster execution here. How to create a Kafka table # The example below shows how to create a Kafka table:| nightlies.apache.org
Windowing table-valued functions (Windowing TVFs) # Batch Streaming Windows are at the heart of processing infinite streams. Windows split the stream into “buckets” of finite size, over which we can apply computations. This document focuses on how windowing is performed in Flink SQL and how the programmer can benefit to the maximum from its offered functionality. Apache Flink provides several window table-valued functions (TVF) to divide the elements of your table into windows, including:| nightlies.apache.org
Introduction # The SQL Gateway is a service that enables multiple clients from the remote to execute SQL in concurrency. It provides an easy way to submit the Flink Job, look up the metadata, and analyze the data online. The SQL Gateway is composed of pluggable endpoints and the SqlGatewayService. The SqlGatewayService is a processor that is reused by the endpoints to handle the requests. The endpoint is an entry point that allows users to connect.| nightlies.apache.org
Name Last modified Size Description| nightlies.apache.org
Joins # Batch Streaming Flink SQL supports complex and flexible join operations over dynamic tables. There are several different types of joins to account for the wide variety of semantics queries may require. By default, the order of joins is not optimized. Tables are joined in the order in which they are specified in the FROM clause. You can tweak the performance of your join queries, by listing the tables with the lowest update frequency first and the tables with the highest update frequen...| nightlies.apache.org