| Interface and Description |
|---|
| org.apache.flink.table.sinks.AppendStreamTableSink
This interface has been replaced by
DynamicTableSink. The new interface
consumes internal data structures. See FLIP-95 for more information. |
| org.apache.flink.table.sinks.RetractStreamTableSink
This interface has been replaced by
DynamicTableSink. The new interface
consumes internal data structures. See FLIP-95 for more information. |
| org.apache.flink.table.connector.source.SourceFunctionProvider
This interface is based on the
SourceFunction API, which is due to be
removed. Use SourceProvider instead. |
| org.apache.flink.table.sinks.StreamTableSink
This interface has been replaced by
DynamicTableSink. The new interface
consumes internal data structures. See FLIP-95 for more information. |
| org.apache.flink.table.factories.StreamTableSinkFactory
This interface has been replaced by
DynamicTableSinkFactory. The new
interface creates instances of DynamicTableSink. See FLIP-95 for more information. |
| org.apache.flink.table.sources.StreamTableSource
This interface has been replaced by
DynamicTableSource. The new interface
produces internal data structures. See FLIP-95 for more information. |
| org.apache.flink.table.factories.StreamTableSourceFactory
This interface has been replaced by
DynamicTableSourceFactory. The new
interface creates instances of DynamicTableSource. See FLIP-95 for more information. |
| org.apache.flink.table.sinks.UpsertStreamTableSink
This interface has been replaced by
DynamicTableSink. The new interface
consumes internal data structures. See FLIP-95 for more information. |
| Class and Description |
|---|
| org.apache.flink.table.sinks.CsvAppendTableSinkFactory
The legacy CSV connector has been replaced by
FileSink. It is kept only to
support tests for the legacy connector stack. |
| org.apache.flink.table.sources.CsvAppendTableSourceFactory
The legacy CSV connector has been replaced by
FileSource. It is kept only to
support tests for the legacy connector stack. |
| org.apache.flink.table.sinks.CsvBatchTableSinkFactory
The legacy CSV connector has been replaced by
FileSink. It is kept only to
support tests for the legacy connector stack. |
| org.apache.flink.table.sources.CsvBatchTableSourceFactory
The legacy CSV connector has been replaced by
FileSource. It is kept only to
support tests for the legacy connector stack. |
| org.apache.flink.table.sinks.CsvTableSink
The legacy CSV connector has been replaced by
FileSink. It is kept only to
support tests for the legacy connector stack. |
| org.apache.flink.table.sinks.CsvTableSinkFactoryBase
The legacy CSV connector has been replaced by
FileSink. It is kept only to
support tests for the legacy connector stack. |
| org.apache.flink.table.sources.CsvTableSource
The legacy CSV connector has been replaced by
FileSource. It is kept only to
support tests for the legacy connector stack. |
| org.apache.flink.table.sources.CsvTableSourceFactoryBase
The legacy CSV connector has been replaced by
FileSource. It is kept only to
support tests for the legacy connector stack. |
| org.apache.flink.table.sources.InputFormatTableSource
This interface has been replaced by
DynamicTableSource. The new interface
produces internal data structures. See FLIP-95 for more information. |
| org.apache.flink.table.descriptors.OldCsvValidator
Use the RFC-compliant
Csv format in the dedicated flink-formats/flink-csv
module instead. |
| org.apache.flink.table.sinks.OutputFormatTableSink
This interface has been replaced by
DynamicTableSink. The new interface
consumes internal data structures. See FLIP-95 for more information. |
| org.apache.flink.table.descriptors.RowtimeValidator
See
Rowtime for details. |
| org.apache.flink.table.descriptors.SchemaValidator
See
Schema for details. |
| Method and Description |
|---|
| org.apache.flink.table.connector.sink.DataStreamSinkProvider.consumeDataStream(DataStream<RowData>)
Use
DataStreamSinkProvider.consumeDataStream(ProviderContext, DataStream)
and correctly set a unique identifier for each data stream transformation. |
org.apache.flink.table.factories.StreamTableSinkFactory.createStreamTableSink(Map<String, String>)
Context contains more information, and already contains table schema too.
Please use TableSinkFactory.createTableSink(Context) instead. |
org.apache.flink.table.factories.StreamTableSourceFactory.createStreamTableSource(Map<String, String>)
Context contains more information, and already contains table schema too.
Please use TableSourceFactory.createTableSource(Context) instead. |
| org.apache.flink.table.api.bridge.java.StreamTableEnvironment.createTemporaryView(String, DataStream<T>, Expression...)
Use
StreamTableEnvironment.createTemporaryView(String, DataStream, Schema) instead. In most
cases, StreamTableEnvironment.createTemporaryView(String, DataStream) should already be sufficient. It
integrates with the new type system and supports all kinds of DataTypes that the
table runtime can consume. The semantics might be slightly different for raw and
structured types. |
| org.apache.flink.table.descriptors.SchemaValidator.deriveTableSinkSchema(DescriptorProperties)
This method combines two separate concepts of table schema and field mapping.
This should be split into two methods once we have support for the corresponding
interfaces (see FLINK-9870).
|
| org.apache.flink.table.sources.CsvTableSource.Builder.field(String, TypeInformation<?>)
This method will be removed in future versions as it uses the old type
system. It is recommended to use
CsvTableSource.Builder.field(String, DataType) instead which uses
the new type system based on DataTypes. Please make sure to use either the
old or the new type system consistently to avoid unintended behavior. See the website
documentation for more information. |
| org.apache.flink.table.api.bridge.java.StreamTableEnvironment.fromDataStream(DataStream<T>, Expression...)
Use
StreamTableEnvironment.fromDataStream(DataStream, Schema) instead. In most cases, StreamTableEnvironment.fromDataStream(DataStream) should already be sufficient. It integrates with the new
type system and supports all kinds of DataTypes that the table runtime can
consume. The semantics might be slightly different for raw and structured types. |
| org.apache.flink.table.connector.source.DataStreamScanProvider.produceDataStream(StreamExecutionEnvironment) |
| org.apache.flink.table.api.bridge.java.StreamTableEnvironment.registerDataStream(String, DataStream<T>) |
| org.apache.flink.table.api.bridge.java.StreamTableEnvironment.registerFunction(String, AggregateFunction<T, ACC>)
Use
TableEnvironment.createTemporarySystemFunction(String, UserDefinedFunction) instead.
Please note that the new method also uses the new type system and reflective extraction
logic. It might be necessary to update the function implementation as well. See the
documentation of AggregateFunction for more information on the new function
design. |
| org.apache.flink.table.api.bridge.java.StreamTableEnvironment.registerFunction(String, TableAggregateFunction<T, ACC>)
Use
TableEnvironment.createTemporarySystemFunction(String, UserDefinedFunction) instead.
Please note that the new method also uses the new type system and reflective extraction
logic. It might be necessary to update the function implementation as well. See the
documentation of TableAggregateFunction for more information on the new function
design. |
| org.apache.flink.table.api.bridge.java.StreamTableEnvironment.registerFunction(String, TableFunction<T>)
Use
TableEnvironment.createTemporarySystemFunction(String, UserDefinedFunction) instead.
Please note that the new method also uses the new type system and reflective extraction
logic. It might be necessary to update the function implementation as well. See the
documentation of TableFunction for more information on the new function design. |
| org.apache.flink.table.api.bridge.java.StreamTableEnvironment.toAppendStream(Table, Class<T>)
Use
StreamTableEnvironment.toDataStream(Table, Class) instead. It integrates with the new type
system and supports all kinds of DataTypes that the table runtime can produce.
The semantics might be slightly different for raw and structured types. Use toDataStream(DataTypes.of(TypeInformation.of(Class))) if TypeInformation should
be used as source of truth. |
| org.apache.flink.table.api.bridge.java.StreamTableEnvironment.toAppendStream(Table, TypeInformation<T>)
Use
StreamTableEnvironment.toDataStream(Table, Class) instead. It integrates with the new type
system and supports all kinds of DataTypes that the table runtime can produce.
The semantics might be slightly different for raw and structured types. Use toDataStream(DataTypes.of(TypeInformation.of(Class))) if TypeInformation should
be used as source of truth. |
| org.apache.flink.table.api.bridge.java.StreamTableEnvironment.toRetractStream(Table, Class<T>)
Use
StreamTableEnvironment.toChangelogStream(Table, Schema) instead. It integrates with the new
type system and supports all kinds of DataTypes and every ChangelogMode
that the table runtime can produce. |
| org.apache.flink.table.api.bridge.java.StreamTableEnvironment.toRetractStream(Table, TypeInformation<T>)
Use
StreamTableEnvironment.toChangelogStream(Table, Schema) instead. It integrates with the new
type system and supports all kinds of DataTypes and every ChangelogMode
that the table runtime can produce. |
Copyright © 2014–2024 The Apache Software Foundation. All rights reserved.