cs3163077 opened a new issue, #5003:
URL: https://github.com/apache/seatunnel/issues/5003

   ### Search before asking
   
   - [X] I had searched in the 
[issues](https://github.com/apache/seatunnel/issues?q=is%3Aissue+label%3A%22bug%22)
 and found no similar issues.
   
   
   ### What happened
   
   SqlServer-CDC 没有数据,也没报错
   
   ### SeaTunnel Version
   
   2.3.2
   
   ### SeaTunnel Config
   
   ```conf
   env {
     job.mode = "STREAMING"
   }
   source {
     SqlServer-CDC {
       result_table_name = "area"    
        username = "**"
        password = "****"
        database-names = ["BI"]
        table-names = ["BI.dbo.area"]
       base-url="jdbc:sqlserver://******:**;databaseName=BI"
     }
   }
   transform {
   }
   sink {
        Console {
       }
   }
   ```
   
   
   ### Running Command
   
   ```shell
   ./start-seatunnel-spark-2-connector-v2.sh --config 
/home/apache-seatunnel-incubating-2.3.1/config/cdc.conf
   ```
   
   
   ### Error Exception
   
   ```log
   23/06/30 15:24:25 INFO spark.SparkContext: Added JAR 
file:/home/apache-seatunnel-incubating-2.3.1/starter/seatunnel-spark-2-starter.jar
 at spark://hadoop3:37437/jars/seatunnel-spark-2-starter.jar with timestamp 
1688109865948
   23/06/30 15:24:26 INFO executor.Executor: Starting executor ID driver on 
host localhost
   23/06/30 15:24:26 INFO util.Utils: Successfully started service 
'org.apache.spark.network.netty.NettyBlockTransferService' on port 41040.
   23/06/30 15:24:26 INFO netty.NettyBlockTransferService: Server created on 
hadoop3:41040
   23/06/30 15:24:26 INFO storage.BlockManager: Using 
org.apache.spark.storage.RandomBlockReplicationPolicy for block replication 
policy
   23/06/30 15:24:26 INFO storage.BlockManagerMaster: Registering BlockManager 
BlockManagerId(driver, hadoop3, 41040, None)
   23/06/30 15:24:26 INFO storage.BlockManagerMasterEndpoint: Registering block 
manager hadoop3:41040 with 366.3 MB RAM, BlockManagerId(driver, hadoop3, 41040, 
None)
   23/06/30 15:24:26 INFO storage.BlockManagerMaster: Registered BlockManager 
BlockManagerId(driver, hadoop3, 41040, None)
   23/06/30 15:24:26 INFO storage.BlockManager: external shuffle service port = 
7337
   23/06/30 15:24:26 INFO storage.BlockManager: Initialized BlockManager: 
BlockManagerId(driver, hadoop3, 41040, None)
   23/06/30 15:24:26 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@41b1f51e{/metrics/json,null,AVAILABLE,@Spark}
   23/06/30 15:24:27 INFO scheduler.EventLoggingListener: Logging events to 
hdfs://hadoop1:8020/user/spark/applicationHistory/local-1688109866035
   23/06/30 15:24:27 WARN lineage.LineageWriter: Lineage directory 
/var/log/spark/lineage doesn't exist or is not writable. Lineage for this 
application will be disabled.
   23/06/30 15:24:27 INFO util.Utils: Extension 
com.cloudera.spark.lineage.NavigatorAppListener not being initialized.
   23/06/30 15:24:27 INFO logging.DriverLogger$DfsAsyncWriter: Started driver 
log file sync to: /user/spark/driverLogs/local-1688109866035_driver.log
   23/06/30 15:24:28 INFO discovery.AbstractPluginDiscovery: Load 
SeaTunnelSource Plugin from 
/home/apache-seatunnel-incubating-2.3.1/connectors/seatunnel
   23/06/30 15:24:28 INFO discovery.AbstractPluginDiscovery: Discovery plugin 
jar: SqlServer-CDC at: 
file:/home/apache-seatunnel-incubating-2.3.1/connectors/seatunnel/connector-cdc-sqlserver-2.3.2.jar
   23/06/30 15:24:28 INFO discovery.AbstractPluginDiscovery: Load plugin: 
PluginIdentifier{engineType='seatunnel', pluginType='source', 
pluginName='SqlServer-CDC'} from classpath
   23/06/30 15:24:28 WARN sqlserver.SqlServerConnection: The 'server.timezone' 
option should be specified to avoid incorrect timestamp values in case of 
different timezones between the database server and this connector's JVM.
   23/06/30 15:24:28 INFO utils.TableDiscoveryUtils: Read list of available 
databases
   23/06/30 15:24:29 INFO utils.TableDiscoveryUtils:     list of available 
databases is: [master, tempdb, model, msdb, BI, BIDEMO, SYSD_BI_DEMO, 
SYSD_BI_DEV_JAVA, *****, SY_BI_COST, SY_BI_1141, TEST1, SYSD_BI_TEST_20211026, 
SYSD_BI_DEV_JAVA1, test123, testdatabig, test12134, SYBI_Test_TW, zhuhm_BIDW, 
bigdatatest, zhuhm_BIDW_E, zhuhm_BIDW1, *****, test_yx, S*****Test, 
SY_BI_CNCEC, BIDW5, *****, S*****, SY_BI_LSC]
   23/06/30 15:24:29 INFO utils.TableDiscoveryUtils: Read list of available 
tables in each database
   23/06/30 15:24:29 INFO utils.TableDiscoveryUtils:     including 
'BI.dbo.area' for further processing
   23/06/30 15:24:29 WARN utils.TableDiscoveryUtils:     skipping database 
'*****' due to error reading tables: 由于数据库 '*****' 离线,无法打开该数据库。
   23/06/30 15:24:29 WARN utils.TableDiscoveryUtils:     skipping database 
'*****' due to error reading tables: 由于数据库 '*****' 离线,无法打开该数据库。
   23/06/30 15:24:29 WARN utils.TableDiscoveryUtils:     skipping database 
'*****' due to error reading tables: 由于数据库 '*****' 离线,无法打开该数据库。
   23/06/30 15:24:29 WARN utils.TableDiscoveryUtils:     skipping database 
'S*****' due to error reading tables: 由于数据库 'S*****' 离线,无法打开该数据库。
   23/06/30 15:24:29 INFO jdbc.JdbcConnection: Connection gracefully closed
   23/06/30 15:24:29 WARN sqlserver.SqlServerConnection: The 'server.timezone' 
option should be specified to avoid incorrect timestamp values in case of 
different timezones between the database server and this connector's JVM.
   23/06/30 15:24:29 WARN sqlserver.SqlServerDefaultValueConverter: Cannot 
parse column default value '(getdate())' to type 'datetime'. Expression 
evaluation is not supported.
   23/06/30 15:24:29 WARN sqlserver.SqlServerDefaultValueConverter: Cannot 
parse column default value '(getdate())' to type 'datetime'. Expression 
evaluation is not supported.
   23/06/30 15:24:29 WARN sqlserver.SqlServerDefaultValueConverter: Cannot 
parse column default value '(getdate())' to type 'datetime'. Expression 
evaluation is not supported.
   23/06/30 15:24:30 INFO jdbc.JdbcConnection: Connection gracefully closed
   23/06/30 15:24:30 INFO execution.SparkRuntimeEnvironment: register plugins 
:[file:/home/apache-seatunnel-incubating-2.3.1/connectors/seatunnel/connector-cdc-sqlserver-2.3.2.jar]
   23/06/30 15:24:30 INFO discovery.AbstractPluginDiscovery: Load 
SeaTunnelTransform Plugin from /home/apache-seatunnel-incubating-2.3.1/lib
   23/06/30 15:24:30 INFO execution.SparkRuntimeEnvironment: register plugins 
:[]
   23/06/30 15:24:30 INFO discovery.AbstractPluginDiscovery: Load SeaTunnelSink 
Plugin from /home/apache-seatunnel-incubating-2.3.1/connectors/seatunnel
   23/06/30 15:24:30 INFO discovery.AbstractPluginDiscovery: Discovery plugin 
jar: Console at: 
file:/home/apache-seatunnel-incubating-2.3.1/connectors/seatunnel/connector-console-2.3.1.jar
   23/06/30 15:24:30 INFO discovery.AbstractPluginDiscovery: Load plugin: 
PluginIdentifier{engineType='seatunnel', pluginType='sink', 
pluginName='Console'} from classpath
   23/06/30 15:24:30 INFO execution.SparkRuntimeEnvironment: register plugins 
:[file:/home/apache-seatunnel-incubating-2.3.1/connectors/seatunnel/connector-console-2.3.1.jar]
   23/06/30 15:24:30 INFO internal.SharedState: loading hive config file: 
file:/etc/hive/conf.cloudera.hive/hive-site.xml
   23/06/30 15:24:30 INFO internal.SharedState: spark.sql.warehouse.dir is not 
set, but hive.metastore.warehouse.dir is set. Setting spark.sql.warehouse.dir 
to the value of hive.metastore.warehouse.dir ('/user/hive/warehouse').
   23/06/30 15:24:30 INFO internal.SharedState: Warehouse path is 
'/user/hive/warehouse'.
   23/06/30 15:24:30 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@4966bab1{/SQL,null,AVAILABLE,@Spark}
   23/06/30 15:24:30 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@7f0f84d4{/SQL/json,null,AVAILABLE,@Spark}
   23/06/30 15:24:30 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@590d6c76{/SQL/execution,null,AVAILABLE,@Spark}
   23/06/30 15:24:30 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@25791d40{/SQL/execution/json,null,AVAILABLE,@Spark}
   23/06/30 15:24:30 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@5a05dd30{/static/sql,null,AVAILABLE,@Spark}
   23/06/30 15:24:31 INFO state.StateStoreCoordinatorRef: Registered 
StateStoreCoordinator endpoint
   23/06/30 15:24:31 WARN lineage.LineageWriter: Lineage directory 
/var/log/spark/lineage doesn't exist or is not writable. Lineage for this 
application will be disabled.
   23/06/30 15:24:31 INFO util.Utils: Extension 
com.cloudera.spark.lineage.NavigatorQueryListener not being initialized.
   23/06/30 15:24:34 INFO v2.DataSourceV2Strategy: 
   Pushing operators to class 
org.apache.seatunnel.translation.spark.source.SeaTunnelSourceSupport
   Pushed Filters: 
   Post-Scan Filters: 
   Output: Areaid#0, Areaname#1
            
   23/06/30 15:24:34 INFO codegen.CodeGenerator: Code generated in 199.1582 ms
   23/06/30 15:24:34 INFO v2.WriteToDataSourceV2Exec: Start processing data 
source writer: 
org.apache.seatunnel.translation.spark.sink.writer.SparkDataSourceWriter@7b8d6c66.
 The input RDD has 1 partitions.
   23/06/30 15:24:34 INFO spark.SparkContext: Starting job: save at 
SinkExecuteProcessor.java:123
   23/06/30 15:24:34 INFO scheduler.DAGScheduler: Got job 0 (save at 
SinkExecuteProcessor.java:123) with 1 output partitions
   23/06/30 15:24:34 INFO scheduler.DAGScheduler: Final stage: ResultStage 0 
(save at SinkExecuteProcessor.java:123)
   23/06/30 15:24:34 INFO scheduler.DAGScheduler: Parents of final stage: List()
   23/06/30 15:24:34 INFO scheduler.DAGScheduler: Missing parents: List()
   23/06/30 15:24:34 INFO scheduler.DAGScheduler: Submitting ResultStage 0 
(MapPartitionsRDD[1] at save at SinkExecuteProcessor.java:123), which has no 
missing parents
   23/06/30 15:24:34 INFO memory.MemoryStore: Block broadcast_0 stored as 
values in memory (estimated size 5.9 KB, free 366.3 MB)
   23/06/30 15:24:35 INFO memory.MemoryStore: Block broadcast_0_piece0 stored 
as bytes in memory (estimated size 3.1 KB, free 366.3 MB)
   23/06/30 15:24:35 INFO storage.BlockManagerInfo: Added broadcast_0_piece0 in 
memory on hadoop3:41040 (size: 3.1 KB, free: 366.3 MB)
   23/06/30 15:24:35 INFO spark.SparkContext: Created broadcast 0 from 
broadcast at DAGScheduler.scala:1164
   23/06/30 15:24:35 INFO scheduler.DAGScheduler: Submitting 1 missing tasks 
from ResultStage 0 (MapPartitionsRDD[1] at save at 
SinkExecuteProcessor.java:123) (first 15 tasks are for partitions Vector(0))
   23/06/30 15:24:35 INFO scheduler.TaskSchedulerImpl: Adding task set 0.0 with 
1 tasks
   23/06/30 15:24:35 INFO scheduler.TaskSetManager: Starting task 0.0 in stage 
0.0 (TID 0, localhost, executor driver, partition 0, PROCESS_LOCAL, 16026 bytes)
   23/06/30 15:24:35 INFO executor.Executor: Running task 0.0 in stage 0.0 (TID 
0)
   23/06/30 15:24:35 INFO executor.Executor: Fetching 
spark://hadoop3:37437/jars/connector-cdc-sqlserver-2.3.2.jar with timestamp 
1688109865947
   23/06/30 15:24:35 INFO client.TransportClientFactory: Successfully created 
connection to hadoop3/192.168.0.213:37437 after 75 ms (0 ms spent in bootstraps)
   23/06/30 15:24:35 INFO util.Utils: Fetching 
spark://hadoop3:37437/jars/connector-cdc-sqlserver-2.3.2.jar to 
/tmp/spark-3382b920-11e6-4daf-ad0c-a523df4ead04/userFiles-a4ca48a0-10f3-468f-b5e3-b953bfa97755/fetchFileTemp8739054726675673564.tmp
   23/06/30 15:24:35 INFO executor.Executor: Adding 
file:/tmp/spark-3382b920-11e6-4daf-ad0c-a523df4ead04/userFiles-a4ca48a0-10f3-468f-b5e3-b953bfa97755/connector-cdc-sqlserver-2.3.2.jar
 to class loader
   23/06/30 15:24:35 INFO executor.Executor: Fetching 
spark://hadoop3:37437/jars/seatunnel-spark-2-starter.jar with timestamp 
1688109865948
   23/06/30 15:24:35 INFO util.Utils: Fetching 
spark://hadoop3:37437/jars/seatunnel-spark-2-starter.jar to 
/tmp/spark-3382b920-11e6-4daf-ad0c-a523df4ead04/userFiles-a4ca48a0-10f3-468f-b5e3-b953bfa97755/fetchFileTemp4417905759660334726.tmp
   23/06/30 15:24:36 INFO executor.Executor: Adding 
file:/tmp/spark-3382b920-11e6-4daf-ad0c-a523df4ead04/userFiles-a4ca48a0-10f3-468f-b5e3-b953bfa97755/seatunnel-spark-2-starter.jar
 to class loader
   23/06/30 15:24:36 INFO executor.Executor: Fetching 
spark://hadoop3:37437/jars/connector-console-2.3.1.jar with timestamp 
1688109865947
   23/06/30 15:24:36 INFO util.Utils: Fetching 
spark://hadoop3:37437/jars/connector-console-2.3.1.jar to 
/tmp/spark-3382b920-11e6-4daf-ad0c-a523df4ead04/userFiles-a4ca48a0-10f3-468f-b5e3-b953bfa97755/fetchFileTemp8783962742653812302.tmp
   23/06/30 15:24:36 INFO executor.Executor: Adding 
file:/tmp/spark-3382b920-11e6-4daf-ad0c-a523df4ead04/userFiles-a4ca48a0-10f3-468f-b5e3-b953bfa97755/connector-console-2.3.1.jar
 to class loader
   23/06/30 15:24:36 INFO executor.Executor: Fetching 
spark://hadoop3:37437/jars/seatunnel-transforms-v2.jar with timestamp 
1688109865945
   23/06/30 15:24:36 INFO util.Utils: Fetching 
spark://hadoop3:37437/jars/seatunnel-transforms-v2.jar to 
/tmp/spark-3382b920-11e6-4daf-ad0c-a523df4ead04/userFiles-a4ca48a0-10f3-468f-b5e3-b953bfa97755/fetchFileTemp6075022225571179099.tmp
   23/06/30 15:24:36 INFO executor.Executor: Adding 
file:/tmp/spark-3382b920-11e6-4daf-ad0c-a523df4ead04/userFiles-a4ca48a0-10f3-468f-b5e3-b953bfa97755/seatunnel-transforms-v2.jar
 to class loader
   23/06/30 15:24:36 INFO sink.ConsoleSinkWriter: output rowType: 
Areaid<STRING>, Areaname<STRING>
   23/06/30 15:24:36 WARN sqlserver.SqlServerConnection: The 'server.timezone' 
option should be specified to avoid incorrect timestamp values in case of 
different timezones between the database server and this connector's JVM.
   23/06/30 15:24:36 INFO utils.TableDiscoveryUtils: Read list of available 
databases
   23/06/30 15:24:36 INFO utils.TableDiscoveryUtils:     list of available 
databases is: [BI]
   23/06/30 15:24:36 INFO utils.TableDiscoveryUtils: Read list of available 
tables in each database
   23/06/30 15:24:36 INFO utils.TableDiscoveryUtils:     including 
'BI.dbo.area' for further processing
   23/06/30 15:24:36 WARN utils.TableDiscoveryUtils:     skipping database 
'*****' due to error reading tables: 由于数据库 '*****' 离线,无法打开该数据库。
   23/06/30 15:24:36 WARN utils.TableDiscoveryUtils:     skipping database 
'*****' due to error reading tables: 由于数据库 '*****' 离线,无法打开该数据库。
   23/06/30 15:24:36 WARN utils.TableDiscoveryUtils:     skipping database 
'*****' due to error reading tables: 由于数据库 '*****' 离线,无法打开该数据库。
   23/06/30 15:24:36 WARN utils.TableDiscoveryUtils:     skipping database 
'S*****' due to error reading tables: 由于数据库 'S*****' 离线,无法打开该数据库。
   23/06/30 15:24:36 INFO jdbc.JdbcConnection: Connection gracefully closed
   23/06/30 15:24:36 INFO reader.SourceReaderBase: Open Source Reader.
   23/06/30 15:24:36 WARN sqlserver.SqlServerConnection: The 'server.timezone' 
option should be specified to avoid incorrect timestamp values in case of 
different timezones between the database server and this connector's JVM.
   23/06/30 15:24:36 INFO eumerator.SqlServerChunkSplitter: Start splitting 
table BI.dbo.area into chunks...
   23/06/30 15:24:37 WARN sqlserver.SqlServerDefaultValueConverter: Cannot 
parse column default value '(getdate())' to type 'datetime'. Expression 
evaluation is not supported.
   23/06/30 15:24:37 WARN sqlserver.SqlServerDefaultValueConverter: Cannot 
parse column default value '(getdate())' to type 'datetime'. Expression 
evaluation is not supported.
   23/06/30 15:24:37 WARN sqlserver.SqlServerDefaultValueConverter: Cannot 
parse column default value '(getdate())' to type 'datetime'. Expression 
evaluation is not supported.
   23/06/30 15:24:37 INFO eumerator.SqlServerChunkSplitter: Use unevenly-sized 
chunks for table BI.dbo.area, the chunk size is 8096
   23/06/30 15:24:37 INFO eumerator.SqlServerChunkSplitter: Split table 
BI.dbo.area into 1 chunks, time cost: 971ms.
   23/06/30 15:24:37 INFO jdbc.JdbcConnection: Connection gracefully closed
   23/06/30 15:24:37 INFO fetcher.SplitFetcher: Starting split fetcher 0
   23/06/30 15:24:37 WARN sqlserver.SqlServerConnection: The 'server.timezone' 
option should be specified to avoid incorrect timestamp values in case of 
different timezones between the database server and this connector's JVM.
   23/06/30 15:24:37 WARN sqlserver.SqlServerConnection: The 'server.timezone' 
option should be specified to avoid incorrect timestamp values in case of 
different timezones between the database server and this connector's JVM.
   23/06/30 15:24:37 INFO history.DatabaseHistoryMetrics: Finished database 
history recovery of 0 change(s) in 1 ms
   23/06/30 15:24:38 INFO fetcher.SplitFetcher: Finished reading from splits 
[BI.dbo.area:0]
   ```
   
   
   ### Flink or Spark Version
   
   spark:2.4.0
   
   ### Java or Scala Version
   
   java:8
   
   ### Screenshots
   
   _No response_
   
   ### Are you willing to submit PR?
   
   - [X] Yes I am willing to submit a PR!
   
   ### Code of Conduct
   
   - [X] I agree to follow this project's [Code of 
Conduct](https://www.apache.org/foundation/policies/conduct)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to