Bingz2 opened a new issue, #2084: URL: https://github.com/apache/incubator-seatunnel/issues/2084
### Search before asking - [X] I had searched in the [issues](https://github.com/apache/incubator-seatunnel/issues?q=is%3Aissue+label%3A%22bug%22) and found no similar issues. ### What happened Reading a hive using Spark JDBC Source fails to return the correct data, which is actually column names rather than actual data ### SeaTunnel Version 2.1.2 ### SeaTunnel Config ```conf source { # This is a example input plugin **only for test and demonstrate the feature input plugin** jdbc { driver = "org.apache.hive.jdbc.HiveDriver" url = "jdbc:hive2://ip:10000/dws" table = "(select mac,user_id,province_code from dws.dws_gvp_user_mac_state_stat_dd where day='2022-06-01') tmp" result_table_name = "tmp" user = "hive" password = "hive" jdbc.fetchsize = 10000 } } transform { # you can also use other filter plugins, such as sql # sql { # sql = "select * from accesslog where request_time > 1000" # } } sink { console { limit = 10, serializer = "json" } } ``` ### Running Command ```shell ./bin/start-seatunnel-spark.sh -m yarn -e client -c ./config/spark.batch.hive.jdbc.conf ``` ### Error Exception ```log 22/06/28 17:39:12 INFO scheduler.DAGScheduler: Missing parents: List() 22/06/28 17:39:12 INFO scheduler.DAGScheduler: Submitting ResultStage 0 (MapPartitionsRDD[6] at take at Console.scala:47), which has no missing parents 22/06/28 17:39:13 INFO memory.MemoryStore: Block broadcast_0 stored as values in memory (estimated size 18.3 KB, free 366.3 MB) 22/06/28 17:39:13 INFO memory.MemoryStore: Block broadcast_0_piece0 stored as bytes in memory (estimated size 8.4 KB, free 366.3 MB) 22/06/28 17:39:13 INFO storage.BlockManagerInfo: Added broadcast_0_piece0 in memory on slave1.test.gitv.we:16907 (size: 8.4 KB, free: 366.3 MB) 22/06/28 17:39:13 INFO spark.SparkContext: Created broadcast 0 from broadcast at DAGScheduler.scala:1161 22/06/28 17:39:13 INFO scheduler.DAGScheduler: Submitting 1 missing tasks from ResultStage 0 (MapPartitionsRDD[6] at take at Console.scala:47) (first 15 tasks are for partitions Vector(0)) 22/06/28 17:39:13 INFO cluster.YarnScheduler: Adding task set 0.0 with 1 tasks 22/06/28 17:39:13 INFO yarn.SparkRackResolver: Got an error when resolving hostNames. Falling back to /default-rack for all 22/06/28 17:39:13 INFO scheduler.TaskSetManager: Starting task 0.0 in stage 0.0 (TID 0, slave4.test.gitv.we, executor 1, partition 0, PROCESS_LOCAL, 7701 bytes) 22/06/28 17:39:14 INFO storage.BlockManagerInfo: Added broadcast_0_piece0 in memory on slave4.test.gitv.we:39237 (size: 8.4 KB, free: 912.3 MB) 22/06/28 17:39:17 INFO scheduler.TaskSetManager: Finished task 0.0 in stage 0.0 (TID 0) in 4138 ms on slave4.test.gitv.we (executor 1) (1/1) 22/06/28 17:39:17 INFO cluster.YarnScheduler: Removed TaskSet 0.0, whose tasks have all completed, from pool 22/06/28 17:39:17 INFO scheduler.DAGScheduler: ResultStage 0 (take at Console.scala:47) finished in 4.616 s 22/06/28 17:39:17 INFO scheduler.DAGScheduler: Job 0 finished: take at Console.scala:47, took 4.674037 s {"mac":"mac","user_id":"user_id","province_code":"province_code"} {"mac":"mac","user_id":"user_id","province_code":"province_code"} {"mac":"mac","user_id":"user_id","province_code":"province_code"} {"mac":"mac","user_id":"user_id","province_code":"province_code"} {"mac":"mac","user_id":"user_id","province_code":"province_code"} {"mac":"mac","user_id":"user_id","province_code":"province_code"} {"mac":"mac","user_id":"user_id","province_code":"province_code"} {"mac":"mac","user_id":"user_id","province_code":"province_code"} {"mac":"mac","user_id":"user_id","province_code":"province_code"} {"mac":"mac","user_id":"user_id","province_code":"province_code"} 22/06/28 17:39:17 INFO spark.SparkContext: Invoking stop() from shutdown hook 22/06/28 17:39:17 INFO server.AbstractConnector: Stopped Spark@4e628b52{HTTP/1.1,[http/1.1]}{0.0.0.0:4040} 22/06/28 17:39:17 INFO ui.SparkUI: Stopped Spark web UI at http://slave1.test.gitv.we:4040 22/06/28 17:39:17 INFO cluster.YarnClientSchedulerBackend: Interrupting monitor thread 22/06/28 17:39:17 INFO cluster.YarnClientSchedulerBackend: Shutting down all executors 22/06/28 17:39:17 INFO cluster.YarnSchedulerBackend$YarnDriverEndpoint: Asking each executor to shut down 22/06/28 17:39:17 INFO cluster.SchedulerExtensionServices: Stopping SchedulerExtensionServices (serviceOption=None, services=List(), started=false) 22/06/28 17:39:17 INFO cluster.YarnClientSchedulerBackend: Stopped 22/06/28 17:39:17 INFO spark.MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped! 22/06/28 17:39:17 INFO memory.MemoryStore: MemoryStore cleared 22/06/28 17:39:17 INFO storage.BlockManager: BlockManager stopped 22/06/28 17:39:17 INFO storage.BlockManagerMaster: BlockManagerMaster stopped 22/06/28 17:39:17 INFO scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: OutputCommitCoordinator stopped! 22/06/28 17:39:17 INFO spark.SparkContext: Successfully stopped SparkContext 22/06/28 17:39:17 INFO util.ShutdownHookManager: Shutdown hook called ``` ### Flink or Spark Version Spark version 2.4.0.cloudera2 ### Java or Scala Version Scala version 2.11.12 java version 1.8.0_112 ### Screenshots  ### Are you willing to submit PR? - [X] Yes I am willing to submit a PR! ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct) -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: [email protected] For queries about this service, please contact Infrastructure at: [email protected]
