@Internal public class KafkaDynamicTableFactory extends Object implements org.apache.flink.table.factories.DynamicTableSourceFactory, org.apache.flink.table.factories.DynamicTableSinkFactory
KafkaDynamicSource and KafkaDynamicSink.| 限定符和类型 | 字段和说明 |
|---|---|
static String |
IDENTIFIER |
| 构造器和说明 |
|---|
KafkaDynamicTableFactory() |
| 限定符和类型 | 方法和说明 |
|---|---|
org.apache.flink.table.connector.sink.DynamicTableSink |
createDynamicTableSink(org.apache.flink.table.factories.DynamicTableFactory.Context context) |
org.apache.flink.table.connector.source.DynamicTableSource |
createDynamicTableSource(org.apache.flink.table.factories.DynamicTableFactory.Context context) |
protected KafkaDynamicSink |
createKafkaTableSink(org.apache.flink.table.types.DataType physicalDataType,
org.apache.flink.table.connector.format.EncodingFormat<org.apache.flink.api.common.serialization.SerializationSchema<org.apache.flink.table.data.RowData>> keyEncodingFormat,
org.apache.flink.table.connector.format.EncodingFormat<org.apache.flink.api.common.serialization.SerializationSchema<org.apache.flink.table.data.RowData>> valueEncodingFormat,
int[] keyProjection,
int[] valueProjection,
String keyPrefix,
String topic,
Properties properties,
FlinkKafkaPartitioner<org.apache.flink.table.data.RowData> partitioner,
org.apache.flink.connector.base.DeliveryGuarantee deliveryGuarantee,
Integer parallelism,
String transactionalIdPrefix) |
protected KafkaDynamicSource |
createKafkaTableSource(org.apache.flink.table.types.DataType physicalDataType,
org.apache.flink.table.connector.format.DecodingFormat<org.apache.flink.api.common.serialization.DeserializationSchema<org.apache.flink.table.data.RowData>> keyDecodingFormat,
org.apache.flink.table.connector.format.DecodingFormat<org.apache.flink.api.common.serialization.DeserializationSchema<org.apache.flink.table.data.RowData>> valueDecodingFormat,
int[] keyProjection,
int[] valueProjection,
String keyPrefix,
List<String> topics,
Pattern topicPattern,
Properties properties,
StartupMode startupMode,
Map<KafkaTopicPartition,Long> specificStartupOffsets,
long startupTimestampMillis,
BoundedMode boundedMode,
Map<KafkaTopicPartition,Long> specificEndOffsets,
long endTimestampMillis,
String tableIdentifier) |
String |
factoryIdentifier() |
Set<org.apache.flink.configuration.ConfigOption<?>> |
forwardOptions() |
Set<org.apache.flink.configuration.ConfigOption<?>> |
optionalOptions() |
Set<org.apache.flink.configuration.ConfigOption<?>> |
requiredOptions() |
public String factoryIdentifier()
factoryIdentifier 在接口中 org.apache.flink.table.factories.Factorypublic Set<org.apache.flink.configuration.ConfigOption<?>> requiredOptions()
requiredOptions 在接口中 org.apache.flink.table.factories.Factorypublic Set<org.apache.flink.configuration.ConfigOption<?>> optionalOptions()
optionalOptions 在接口中 org.apache.flink.table.factories.Factorypublic Set<org.apache.flink.configuration.ConfigOption<?>> forwardOptions()
forwardOptions 在接口中 org.apache.flink.table.factories.DynamicTableFactorypublic org.apache.flink.table.connector.source.DynamicTableSource createDynamicTableSource(org.apache.flink.table.factories.DynamicTableFactory.Context context)
createDynamicTableSource 在接口中 org.apache.flink.table.factories.DynamicTableSourceFactorypublic org.apache.flink.table.connector.sink.DynamicTableSink createDynamicTableSink(org.apache.flink.table.factories.DynamicTableFactory.Context context)
createDynamicTableSink 在接口中 org.apache.flink.table.factories.DynamicTableSinkFactoryprotected KafkaDynamicSource createKafkaTableSource(org.apache.flink.table.types.DataType physicalDataType, @Nullable org.apache.flink.table.connector.format.DecodingFormat<org.apache.flink.api.common.serialization.DeserializationSchema<org.apache.flink.table.data.RowData>> keyDecodingFormat, org.apache.flink.table.connector.format.DecodingFormat<org.apache.flink.api.common.serialization.DeserializationSchema<org.apache.flink.table.data.RowData>> valueDecodingFormat, int[] keyProjection, int[] valueProjection, @Nullable String keyPrefix, @Nullable List<String> topics, @Nullable Pattern topicPattern, Properties properties, StartupMode startupMode, Map<KafkaTopicPartition,Long> specificStartupOffsets, long startupTimestampMillis, BoundedMode boundedMode, Map<KafkaTopicPartition,Long> specificEndOffsets, long endTimestampMillis, String tableIdentifier)
protected KafkaDynamicSink createKafkaTableSink(org.apache.flink.table.types.DataType physicalDataType, @Nullable org.apache.flink.table.connector.format.EncodingFormat<org.apache.flink.api.common.serialization.SerializationSchema<org.apache.flink.table.data.RowData>> keyEncodingFormat, org.apache.flink.table.connector.format.EncodingFormat<org.apache.flink.api.common.serialization.SerializationSchema<org.apache.flink.table.data.RowData>> valueEncodingFormat, int[] keyProjection, int[] valueProjection, @Nullable String keyPrefix, String topic, Properties properties, FlinkKafkaPartitioner<org.apache.flink.table.data.RowData> partitioner, org.apache.flink.connector.base.DeliveryGuarantee deliveryGuarantee, Integer parallelism, @Nullable String transactionalIdPrefix)
Copyright © 2014–2023 The Apache Software Foundation. All rights reserved.