dianfu commented on code in PR #19919:
URL: https://github.com/apache/flink/pull/19919#discussion_r895511261
##########
flink-python/pyflink/datastream/connectors/__init__.py:
##########
@@ -16,58 +16,26 @@
# limitations under the License.
################################################################################
from pyflink.datastream.connectors.base import Sink, Source, DeliveryGuarantee
+from pyflink.datastream.connectors.number_seq import NumberSequenceSource
from pyflink.datastream.connectors.file_system import (FileEnumeratorProvider,
FileSink, FileSource,
BucketAssigner,
FileSourceBuilder,
FileSplitAssignerProvider, OutputFileConfig,
RollingPolicy,
StreamFormat,
StreamingFileSink)
-from pyflink.datastream.connectors.jdbc import JdbcSink,
JdbcConnectionOptions, JdbcExecutionOptions
-from pyflink.datastream.connectors.kafka import FlinkKafkaConsumer,
FlinkKafkaProducer, Semantic
-from pyflink.datastream.connectors.number_seq import NumberSequenceSource
-from pyflink.datastream.connectors.pulsar import PulsarDeserializationSchema,
PulsarSource, \
- PulsarSourceBuilder, SubscriptionType, StartCursor, StopCursor,
PulsarSerializationSchema, \
- PulsarSink, PulsarSinkBuilder, MessageDelayer, TopicRoutingMode
-from pyflink.datastream.connectors.rabbitmq import RMQConnectionConfig,
RMQSource, RMQSink
-from pyflink.datastream.connectors.kinesis import (FlinkKinesisConsumer,
KinesisStreamsSink,
- KinesisFirehoseSink)
-
__all__ = [
'Sink',
'Source',
'DeliveryGuarantee',
+ 'NumberSequenceSource',
+ 'BucketAssigner',
'FileEnumeratorProvider',
'FileSink',
'FileSource',
- 'BucketAssigner',
'FileSourceBuilder',
'FileSplitAssignerProvider',
- 'FlinkKafkaConsumer',
- 'FlinkKafkaProducer',
- 'Semantic',
- 'JdbcSink',
- 'JdbcConnectionOptions',
- 'JdbcExecutionOptions',
- 'NumberSequenceSource',
'OutputFileConfig',
- 'PulsarDeserializationSchema',
- 'PulsarSource',
- 'PulsarSourceBuilder',
- 'SubscriptionType',
- 'PulsarSerializationSchema',
- 'PulsarSink',
- 'PulsarSinkBuilder',
- 'MessageDelayer',
- 'TopicRoutingMode',
- 'RMQConnectionConfig',
- 'RMQSource',
- 'RMQSink',
'RollingPolicy',
Review Comment:
@deadwind4 Yes, of course. BTW, we should also considering the backward
compatibility. One way that may work is adding the following code in the
connector/__init__.py:
```
def _install():
from pyflink.datastream.connectors import jdbc
setattr(pyflink.datastream.connectors, 'JdbcSink', jdbc.JdbcSink)
...
_install()
del _install
```
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]