您好:
非常感谢大家的回复帮忙。我看到TM的日志了(如下),而且TM根本就没有跑起来:
2019-10-21 15:39:33,192 INFO
org.apache.flink.runtime.taskexecutor.TaskManagerServices - Temporary file
directory
'/opt/appdata/disk24/yarn/nm/usercache/flink/appcache/application_1562833757356_1956':
total 3723 GB, usable 3722 GB (99.97% usable)
2019-10-21 15:39:33,389 INFO
org.apache.flink.runtime.io.network.buffer.NetworkBufferPool - Allocated 294
MB for network buffer pool (number of memory segments: 9425, bytes per segment:
32768).
2019-10-21 15:39:33,445 INFO
org.apache.flink.runtime.query.QueryableStateUtils - Could not load
Queryable State Client Proxy. Probable reason: flink-queryable-state-runtime is
not in the classpath. To enable Queryable State, please move the
flink-queryable-state-runtime jar from the opt to the lib folder.
2019-10-21 15:39:33,446 INFO
org.apache.flink.runtime.query.QueryableStateUtils - Could not load
Queryable State Server. Probable reason: flink-queryable-state-runtime is not
in the classpath. To enable Queryable State, please move the
flink-queryable-state-runtime jar from the opt to the lib folder.
2019-10-21 15:39:33,447 INFO
org.apache.flink.runtime.io.network.NetworkEnvironment - Starting the
network environment and its components.
2019-10-21 15:39:33,484 INFO
org.apache.flink.runtime.io.network.netty.NettyClient - Successful
initialization (took 35 ms).
2019-10-21 15:39:33,526 INFO
org.apache.flink.runtime.io.network.netty.NettyServer - Successful
initialization (took 41 ms). Listening on SocketAddress /10.221.124.46:43869.
2019-10-21 15:39:33,527 INFO
org.apache.flink.runtime.taskexecutor.TaskManagerServices - Limiting
managed memory to 0.7 of the currently free heap space (1839 MB), memory will
be allocated lazily.
2019-10-21 15:39:33,531 INFO
org.apache.flink.runtime.io.disk.iomanager.IOManager - I/O manager
uses directory
/opt/appdata/disk01/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-io-1eed2f21-58ad-4f92-8e33-c87d2b9cc519
for spill files.
2019-10-21 15:39:33,531 INFO
org.apache.flink.runtime.io.disk.iomanager.IOManager - I/O manager
uses directory
/opt/appdata/disk02/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-io-e917619a-2c43-403b-b2ad-c51098d3ac55
for spill files.
2019-10-21 15:39:33,531 INFO
org.apache.flink.runtime.io.disk.iomanager.IOManager - I/O manager
uses directory
/opt/appdata/disk03/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-io-8f7ca801-647b-421e-a28a-00da5c62c213
for spill files.
2019-10-21 15:39:33,531 INFO
org.apache.flink.runtime.io.disk.iomanager.IOManager - I/O manager
uses directory
/opt/appdata/disk04/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-io-d6b4a21c-5a8b-4f5f-856d-4dae41e0ad30
for spill files.
2019-10-21 15:39:33,532 INFO
org.apache.flink.runtime.io.disk.iomanager.IOManager - I/O manager
uses directory
/opt/appdata/disk05/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-io-71e74e70-a582-47cb-8481-15b821ab51f8
for spill files.
2019-10-21 15:39:33,532 INFO
org.apache.flink.runtime.io.disk.iomanager.IOManager - I/O manager
uses directory
/opt/appdata/disk06/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-io-78be3561-3d4a-46ce-ae71-41efde4c08d1
for spill files.
2019-10-21 15:39:33,532 INFO
org.apache.flink.runtime.io.disk.iomanager.IOManager - I/O manager
uses directory
/opt/appdata/disk07/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-io-54d336b9-d697-4234-a5ba-a3ecf2236a5b
for spill files.
2019-10-21 15:39:33,532 INFO
org.apache.flink.runtime.io.disk.iomanager.IOManager - I/O manager
uses directory
/opt/appdata/disk08/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-io-0ee42053-dd58-431a-981e-2199ffe363bc
for spill files.
2019-10-21 15:39:33,532 INFO
org.apache.flink.runtime.io.disk.iomanager.IOManager - I/O manager
uses directory
/opt/appdata/disk09/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-io-1b18f7b3-059f-4c9a-92bc-8c55ac972da8
for spill files.
2019-10-21 15:39:33,532 INFO
org.apache.flink.runtime.io.disk.iomanager.IOManager - I/O manager
uses directory
/opt/appdata/disk10/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-io-6d3b5a27-6733-4001-b5f5-3556b24a8929
for spill files.
2019-10-21 15:39:33,533 INFO
org.apache.flink.runtime.io.disk.iomanager.IOManager - I/O manager
uses directory
/opt/appdata/disk11/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-io-2beee747-7161-44c3-aecf-fcf17fd85e97
for spill files.
2019-10-21 15:39:33,533 INFO
org.apache.flink.runtime.io.disk.iomanager.IOManager - I/O manager
uses directory
/opt/appdata/disk12/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-io-6b769ce1-a9ab-47cc-a92f-d4e3602db491
for spill files.
2019-10-21 15:39:33,533 INFO
org.apache.flink.runtime.io.disk.iomanager.IOManager - I/O manager
uses directory
/opt/appdata/disk13/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-io-48302e87-6887-4b92-9bbd-0e813878fe2e
for spill files.
2019-10-21 15:39:33,533 INFO
org.apache.flink.runtime.io.disk.iomanager.IOManager - I/O manager
uses directory
/opt/appdata/disk14/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-io-5fd30eb2-0854-462c-bff0-42fad5928ab7
for spill files.
2019-10-21 15:39:33,533 INFO
org.apache.flink.runtime.io.disk.iomanager.IOManager - I/O manager
uses directory
/opt/appdata/disk15/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-io-ae3d47e4-9155-4d07-8f0e-a4798ca2281d
for spill files.
2019-10-21 15:39:33,534 INFO
org.apache.flink.runtime.io.disk.iomanager.IOManager - I/O manager
uses directory
/opt/appdata/disk16/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-io-a8bf30d0-c617-4c8a-8672-ab5636341fb3
for spill files.
2019-10-21 15:39:33,534 INFO
org.apache.flink.runtime.io.disk.iomanager.IOManager - I/O manager
uses directory
/opt/appdata/disk17/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-io-1271499d-63c2-45fc-bbc4-c3f092b39ccd
for spill files.
2019-10-21 15:39:33,534 INFO
org.apache.flink.runtime.io.disk.iomanager.IOManager - I/O manager
uses directory
/opt/appdata/disk18/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-io-035c1b6c-933f-4b9a-a4d2-c884798b4df7
for spill files.
2019-10-21 15:39:33,534 INFO
org.apache.flink.runtime.io.disk.iomanager.IOManager - I/O manager
uses directory
/opt/appdata/disk19/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-io-0627e795-a79c-4a5d-9a8f-6f3189323490
for spill files.
2019-10-21 15:39:33,534 INFO
org.apache.flink.runtime.io.disk.iomanager.IOManager - I/O manager
uses directory
/opt/appdata/disk20/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-io-f5f71ab0-1362-4f06-b2b3-a27f9095cf7f
for spill files.
2019-10-21 15:39:33,535 INFO
org.apache.flink.runtime.io.disk.iomanager.IOManager - I/O manager
uses directory
/opt/appdata/disk21/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-io-7b2a6d98-bce5-4fab-b482-a238e64bd897
for spill files.
2019-10-21 15:39:33,535 INFO
org.apache.flink.runtime.io.disk.iomanager.IOManager - I/O manager
uses directory
/opt/appdata/disk22/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-io-6a075ee4-a5e6-4950-8a57-90f8ca1f063f
for spill files.
2019-10-21 15:39:33,535 INFO
org.apache.flink.runtime.io.disk.iomanager.IOManager - I/O manager
uses directory
/opt/appdata/disk23/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-io-d323bf60-4129-475c-ae73-494e44570e1c
for spill files.
2019-10-21 15:39:33,535 INFO
org.apache.flink.runtime.io.disk.iomanager.IOManager - I/O manager
uses directory
/opt/appdata/disk24/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-io-e19de3fb-ae65-47d9-a4dd-4316151cb67a
for spill files.
2019-10-21 15:39:33,598 INFO
org.apache.flink.runtime.taskexecutor.TaskManagerConfiguration - Messages have
a max timeout of 10000 ms
2019-10-21 15:39:33,609 INFO org.apache.flink.runtime.rpc.akka.AkkaRpcService
- Starting RPC endpoint for
org.apache.flink.runtime.taskexecutor.TaskExecutor at
akka://flink/user/taskmanager_0 .
2019-10-21 15:39:33,623 INFO
org.apache.flink.runtime.taskexecutor.JobLeaderService - Start job
leader service.
2019-10-21 15:39:33,623 INFO
org.apache.flink.runtime.taskexecutor.TaskExecutor - Connecting to
ResourceManager
akka.tcp://[email protected]:38079/user/resourcemanager(00000000000000000000000000000000).
2019-10-21 15:39:33,624 INFO org.apache.flink.runtime.filecache.FileCache
- User file cache uses directory
/opt/appdata/disk01/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-dist-cache-5498311b-ec06-4806-99e3-ab4fc1a80b22
2019-10-21 15:39:33,624 INFO org.apache.flink.runtime.filecache.FileCache
- User file cache uses directory
/opt/appdata/disk02/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-dist-cache-d4dd8e64-94f1-4994-8255-59c47f6e9f81
2019-10-21 15:39:33,624 INFO org.apache.flink.runtime.filecache.FileCache
- User file cache uses directory
/opt/appdata/disk03/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-dist-cache-e82c4569-039b-4e9f-af99-676626233ef7
2019-10-21 15:39:33,624 INFO org.apache.flink.runtime.filecache.FileCache
- User file cache uses directory
/opt/appdata/disk04/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-dist-cache-29bdc137-d509-4935-b02d-166c418f4389
2019-10-21 15:39:33,624 INFO org.apache.flink.runtime.filecache.FileCache
- User file cache uses directory
/opt/appdata/disk05/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-dist-cache-b0f12bab-1241-422c-9927-21e3d31dc6e1
2019-10-21 15:39:33,624 INFO org.apache.flink.runtime.filecache.FileCache
- User file cache uses directory
/opt/appdata/disk06/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-dist-cache-c9be4b9b-91be-4138-8117-d72370283804
2019-10-21 15:39:33,625 INFO org.apache.flink.runtime.filecache.FileCache
- User file cache uses directory
/opt/appdata/disk07/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-dist-cache-bd836961-21b3-4f79-84a7-be392a9de752
2019-10-21 15:39:33,625 INFO org.apache.flink.runtime.filecache.FileCache
- User file cache uses directory
/opt/appdata/disk08/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-dist-cache-a0c776dd-68da-4b25-9cf4-0530e464c9f3
2019-10-21 15:39:33,625 INFO org.apache.flink.runtime.filecache.FileCache
- User file cache uses directory
/opt/appdata/disk09/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-dist-cache-de986724-8bcf-458c-aec5-012abd0e0b6f
2019-10-21 15:39:33,625 INFO org.apache.flink.runtime.filecache.FileCache
- User file cache uses directory
/opt/appdata/disk10/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-dist-cache-9bcddf05-d8d8-4709-bef4-7483f62c9d03
2019-10-21 15:39:33,625 INFO org.apache.flink.runtime.filecache.FileCache
- User file cache uses directory
/opt/appdata/disk11/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-dist-cache-7e03fafc-edaa-4ccc-b936-a3188642dc08
2019-10-21 15:39:33,625 INFO org.apache.flink.runtime.filecache.FileCache
- User file cache uses directory
/opt/appdata/disk12/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-dist-cache-33ae11b3-3259-4e18-8db4-42dbedf3c8c6
2019-10-21 15:39:33,626 INFO org.apache.flink.runtime.filecache.FileCache
- User file cache uses directory
/opt/appdata/disk13/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-dist-cache-57dce04a-402e-4840-bdde-164701995aec
2019-10-21 15:39:33,626 INFO org.apache.flink.runtime.filecache.FileCache
- User file cache uses directory
/opt/appdata/disk14/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-dist-cache-05211842-51b3-447a-b586-5d1c4d03644a
2019-10-21 15:39:33,626 INFO org.apache.flink.runtime.filecache.FileCache
- User file cache uses directory
/opt/appdata/disk15/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-dist-cache-7b8b2314-f2f4-4f0e-a49e-7eff9276886a
2019-10-21 15:39:33,626 INFO org.apache.flink.runtime.filecache.FileCache
- User file cache uses directory
/opt/appdata/disk16/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-dist-cache-185841ad-5b1b-4601-a83f-55664cc27cbe
2019-10-21 15:39:33,626 INFO org.apache.flink.runtime.filecache.FileCache
- User file cache uses directory
/opt/appdata/disk17/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-dist-cache-e71950e1-bf91-4d7e-96d7-7590547e3ec4
2019-10-21 15:39:33,626 INFO org.apache.flink.runtime.filecache.FileCache
- User file cache uses directory
/opt/appdata/disk18/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-dist-cache-99f4927a-4285-4962-b919-041673b99995
2019-10-21 15:39:33,627 INFO org.apache.flink.runtime.filecache.FileCache
- User file cache uses directory
/opt/appdata/disk19/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-dist-cache-25e3472a-b60a-4def-82d9-b75667ceb6f1
2019-10-21 15:39:33,627 INFO org.apache.flink.runtime.filecache.FileCache
- User file cache uses directory
/opt/appdata/disk20/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-dist-cache-e586a3c3-0fe7-4e64-b94e-826bcf7ddff1
2019-10-21 15:39:33,627 INFO org.apache.flink.runtime.filecache.FileCache
- User file cache uses directory
/opt/appdata/disk21/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-dist-cache-4957010f-4771-404c-be63-7ffb70ad020c
2019-10-21 15:39:33,627 INFO org.apache.flink.runtime.filecache.FileCache
- User file cache uses directory
/opt/appdata/disk22/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-dist-cache-9137dfa2-c3fb-43ae-93a6-d4c51440495b
2019-10-21 15:39:33,627 INFO org.apache.flink.runtime.filecache.FileCache
- User file cache uses directory
/opt/appdata/disk23/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-dist-cache-487e0ed7-77e5-49a0-a1cf-2900fa4f066a
2019-10-21 15:39:33,627 INFO org.apache.flink.runtime.filecache.FileCache
- User file cache uses directory
/opt/appdata/disk24/yarn/nm/usercache/flink/appcache/application_1562833757356_1956/flink-dist-cache-49d2866b-6b8b-48b3-bf40-20d28ccab1a2
2019-10-21 15:39:33,812 INFO
org.apache.flink.runtime.taskexecutor.TaskExecutor - Resolved
ResourceManager address, beginning registration
2019-10-21 15:39:33,812 INFO
org.apache.flink.runtime.taskexecutor.TaskExecutor - Registration at
ResourceManager attempt 1 (timeout=100ms)
2019-10-21 15:39:33,859 INFO
org.apache.flink.runtime.taskexecutor.TaskExecutor - Successful
registration at resource manager
akka.tcp://[email protected]:38079/user/resourcemanager under registration id
2f7bb48440ad95aeec8a05357b618733.
2019-10-21 15:39:33,876 INFO
org.apache.flink.runtime.taskexecutor.TaskExecutor - Receive slot
request AllocationID{6203d66760d158ff8a22d86a4e37d309} for job
968350be49f51fee2e660184572ac618 from resource manager with leader id
00000000000000000000000000000000.
2019-10-21 15:39:33,877 INFO
org.apache.flink.runtime.taskexecutor.TaskExecutor - Allocated slot
for AllocationID{6203d66760d158ff8a22d86a4e37d309}.
2019-10-21 15:39:33,877 INFO
org.apache.flink.runtime.taskexecutor.JobLeaderService - Add job
968350be49f51fee2e660184572ac618 for job leader monitoring.
2019-10-21 15:39:33,878 INFO
org.apache.flink.runtime.taskexecutor.JobLeaderService - Try to register
at job manager akka.tcp://[email protected]:38079/user/jobmanager_0 with leader id
00000000-0000-0000-0000-000000000000.
2019-10-21 15:39:33,891 INFO
org.apache.flink.runtime.taskexecutor.JobLeaderService - Resolved
JobManager address, beginning registration
2019-10-21 15:39:33,892 INFO
org.apache.flink.runtime.taskexecutor.JobLeaderService - Registration at
JobManager attempt 1 (timeout=100ms)
2019-10-21 15:39:33,906 INFO
org.apache.flink.runtime.taskexecutor.JobLeaderService - Successful
registration at job manager akka.tcp://[email protected]:38079/user/jobmanager_0 for
job 968350be49f51fee2e660184572ac618.
2019-10-21 15:39:33,907 INFO
org.apache.flink.runtime.taskexecutor.TaskExecutor - Establish
JobManager connection for job 968350be49f51fee2e660184572ac618.
2019-10-21 15:39:33,910 INFO
org.apache.flink.runtime.taskexecutor.TaskExecutor - Offer reserved
slots to the leader of job 968350be49f51fee2e660184572ac618.
2019-10-21 15:39:33,921 INFO
org.apache.flink.runtime.taskexecutor.slot.TaskSlotTable - Activate slot
AllocationID{6203d66760d158ff8a22d86a4e37d309}.
2019-10-21 15:39:33,947 INFO
org.apache.flink.runtime.taskexecutor.TaskExecutor - Received task
Source: Custom Source -> Flat Map -> Map -> Map -> Sink: flink-conncetors-kafka
(1/1).
2019-10-21 15:39:33,947 INFO org.apache.flink.runtime.taskmanager.Task
- Source: Custom Source -> Flat Map -> Map -> Map -> Sink:
flink-conncetors-kafka (1/1) (6ada18dcebd7025ec5013aeecef2c3c3) switched from
CREATED to DEPLOYING.
2019-10-21 15:39:33,947 INFO org.apache.flink.runtime.taskmanager.Task
- Creating FileSystem stream leak safety net for task Source:
Custom Source -> Flat Map -> Map -> Map -> Sink: flink-conncetors-kafka (1/1)
(6ada18dcebd7025ec5013aeecef2c3c3) [DEPLOYING]
2019-10-21 15:39:33,950 INFO org.apache.flink.runtime.taskmanager.Task
- Loading JAR files for task Source: Custom Source -> Flat Map ->
Map -> Map -> Sink: flink-conncetors-kafka (1/1)
(6ada18dcebd7025ec5013aeecef2c3c3) [DEPLOYING].
2019-10-21 15:39:33,953 INFO org.apache.flink.runtime.blob.BlobClient
- Downloading
968350be49f51fee2e660184572ac618/p-e063684ba81f692adfcda81a3e9c67dbfbd9a154-145a4657337e8272759a7eed0332c1e8
from p.com/10.221.124.46:33338
2019-10-21 15:39:35,256 INFO org.apache.flink.runtime.taskmanager.Task
- Registering task at network: Source: Custom Source -> Flat Map ->
Map -> Map -> Sink: flink-conncetors-kafka (1/1)
(6ada18dcebd7025ec5013aeecef2c3c3) [DEPLOYING].
2019-10-21 15:39:35,268 INFO org.apache.flink.runtime.taskmanager.Task
- Source: Custom Source -> Flat Map -> Map -> Map -> Sink:
flink-conncetors-kafka (1/1) (6ada18dcebd7025ec5013aeecef2c3c3) switched from
DEPLOYING to RUNNING.
2019-10-21 15:39:35,276 INFO
org.apache.flink.streaming.runtime.tasks.StreamTask - No state
backend has been configured, using default (Memory / JobManager)
MemoryStateBackend (data in heap memory / checkpoints to JobManager)
(checkpoints: 'null', savepoints: 'null', asynchronous: TRUE, maxStateSize:
5242880)
2019-10-21 15:39:35,441 INFO org.apache.flink.api.java.typeutils.TypeExtractor
- class
org.apache.flink.streaming.connectors.kafka.internals.KafkaTopicPartition does
not contain a setter for field topic
2019-10-21 15:39:35,441 INFO org.apache.flink.api.java.typeutils.TypeExtractor
- Class class
org.apache.flink.streaming.connectors.kafka.internals.KafkaTopicPartition
cannot be used as a POJO type because not all fields are valid POJO fields, and
must be processed as GenericType. Please read the Flink documentation on "Data
Types & Serialization" for details of the effect on performance.
2019-10-21 15:39:35,444 INFO
org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumerBase - No
restore state for FlinkKafkaConsumer.
2019-10-21 15:39:35,456 INFO org.apache.kafka.clients.producer.ProducerConfig
- ProducerConfig values:
===================
其中,我认为是“.KafkaTopicPartition cannot be used as a POJO type because not
all fields are valid POJO fields”这个原因,但是还是在google没有找到具体的解决办法,在钉钉群里尝试问一次也没有回信。
同时可能是序列化的问题,所以将serialVersionUID 加上【注:本地跑依旧正常,但是提交到Yarn集群上就报错!】,:
public class KafkaFlinkKafkaTestFour implements Serializable{
private static final long serialVersionUID = 5624696288943005947L;
public static Log log = LogFactory.getLog(KafkaFlinkKafkaTestFour.class);
public static void main(String[] args) throws Exception {
System.out.println("11111111111111111111111");
// readRedis redis = new readRedis();
final StreamExecutionEnvironment env =
StreamExecutionEnvironment.getExecutionEnvironment();
env.enableCheckpointing(1000);
Properties properties1 = new Properties();
properties1.setProperty("bootstrap.servers", "xxx:9092, xxx:9092,
xxx:9092");
properties1.setProperty("group.id", "test_bigdata");
System.out.println("333333333333333333333333");
final FlinkKafkaConsumer010<String> myConsumer = new
FlinkKafkaConsumer010<String>("bigeyeservertopictest", new
SimpleStringSchema(), properties1);
// FlinkKafkaProducer010<String> SinkProducer = new
FlinkKafkaProducer010<String>("test", new SimpleStringSchema(), properties2);
DataStream<String> stream = env.addSource(myConsumer);
// stream.print();
DataStream<String> student = stream.flatMap(new
StringToJsonObject()).map(value -> value.toJSONArraytoString())
.map(new MapFunction<JSONArray, String>() {
public String map(JSONArray jsonArray) throws Exception{
String source= "bigeye";
JSONObject jsonObject1 = new JSONObject();
String result = null;
// readRedis redis = new readRedis();
for(int i=0;i<jsonArray.size();i++){
// log.info("1111111111111111");
String jsonObjectStr = jsonArray.getString(i);
String timeValue =
JSONObject.parseObject(jsonObjectStr).getString("time");
String valueValue =
JSONObject.parseObject(jsonObjectStr).getString("value");
String agentHostnameValue =
JSONObject.parseObject(jsonObjectStr).getString("agentHostname");
String streamIDValue =
JSONObject.parseObject(jsonObjectStr).getString("streamID");
// Map<String, String> redisMap =
redis.getInfo(agentHostnameValue,streamIDValue,source);
// 创建opentsdb数据对象
Map<String, Object> opentsdbValue = new HashMap<>();
opentsdbValue.put("metric", streamIDValue);
opentsdbValue.put("timestamp", timeValue);
opentsdbValue.put("value", valueValue);
Gson gson = new Gson();
// opentsdbValue.put("tags", gson.fromJson((String)
redisMap.get("tags"), Map.class));
// jsonObject1.put(flag,timeValue+"+"+streamIDValue);
jsonObject1.put("BigeyeValue",opentsdbValue);
// redis.close();
}
result = jsonObject1.toString();
return result;
}
})
;
student.print();
student.addSink(new FlinkKafkaProducer010<String>("xxxx:9092","test",
new SimpleStringSchema())).name("flink-conncetors-kafka");
// .addSink(SinkProducer);
// .print();
System.out.println("4444444444444444444444444");
env.execute("WordCount from Kafka data");
=================
感谢!
[email protected]
发件人: Xintong Song
发送时间: 2019-10-21 12:14
收件人: user-zh
主题: Re: Flink提jar包部署到Yarn上报错
看报错是TM挂了,具体原因需要分析TM日志,有可能是上面答复中相同的问题,也有可能是其他原因造成的。
Thank you~
Xintong Song
On Mon, Oct 21, 2019 at 11:36 AM [email protected] <[email protected]> wrote:
> 参考:
> http://mail-archives.apache.org/mod_mbox/flink-user-zh/201905.mbox/%[email protected]%3E
>
>
>
> [email protected]
>
> 发件人: [email protected]
> 发送时间: 2019-10-21 11:05
> 收件人: user-zh
> 抄送: zhangjunjie1130
> 主题: Flink提jar包部署到Yarn上报错
> 您好:
> 我的程序是从kafka取数,然后Flink处理后有写入kafka;本地运行正常,上传到Yarn集群上报错,
> Flink版本是:1.7.2
>
> 错误是:
> 2019-10-21 09:52:30,054 INFO
> org.apache.flink.runtime.checkpoint.CheckpointCoordinator - Checkpoint
> triggering task Source: Custom Source -> Flat Map -> Map -> Map -> Sink:
> flink-conncetors-kafka (1/1) of job 7d5dfa42776d679eb240fa833444bc22 is not
> in state RUNNING but DEPLOYING instead. Aborting checkpoint.
> 2019-10-21 09:52:30,389 INFO
> org.apache.flink.runtime.executiongraph.ExecutionGraph - Source:
> Custom Source -> Flat Map -> Map -> Map -> Sink: flink-conncetors-kafka
> (1/1) (05b67b88bf084a9c9884201d224768b4) switched from DEPLOYING to RUNNING.
> 2019-10-21 09:52:31,054 INFO
> org.apache.flink.runtime.checkpoint.CheckpointCoordinator - Triggering
> checkpoint 1 @ 1571622751054 for job 7d5dfa42776d679eb240fa833444bc22.
> 2019-10-21 09:52:31,488 ERROR
> org.apache.flink.runtime.rest.handler.taskmanager.TaskManagerDetailsHandler
> - Implementation error: Unhandled exception.
> org.apache.flink.runtime.resourcemanager.exceptions.UnknownTaskExecutorException:
> No TaskExecutor registered under container_e12_1562833757356_1953_01_000002.
> at
> org.apache.flink.runtime.resourcemanager.ResourceManager.requestTaskManagerInfo(ResourceManager.java:563)
> at sun.reflect.GeneratedMethodAccessor43.invoke(Unknown Source)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498)
> at
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcInvocation(AkkaRpcActor.java:247)
> at
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:162)
> at
> org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:70)
> at
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.onReceive(AkkaRpcActor.java:142)
> at
> org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.onReceive(FencedAkkaRpcActor.java:40)
> at
> akka.actor.UntypedActor$$anonfun$receive$1.applyOrElse(UntypedActor.scala:165)
> at akka.actor.Actor$class.aroundReceive(Actor.scala:502)
> at akka.actor.UntypedActor.aroundReceive(UntypedActor.scala:95)
> at akka.actor.ActorCell.receiveMessage(ActorCell.scala:526)
> at akka.actor.ActorCell.invoke(ActorCell.scala:495)
> at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:257)
> at akka.dispatch.Mailbox.run(Mailbox.scala:224)
> at akka.dispatch.Mailbox.exec(Mailbox.scala:234)
> at
> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
> at
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
> at
> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
> at
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> 2019-10-21 09:52:34,467 ERROR
> org.apache.flink.runtime.rest.handler.taskmanager.TaskManagerDetailsHandler
> - Implementation error: Unhandled exception.
> org.apache.flink.runtime.resourcemanager.exceptions.UnknownTaskExecutorException:
> No TaskExecutor registered under container_e12_1562833757356_1953_01_000002.
> at
> org.apache.flink.runtime.resourcemanager.ResourceManager.requestTaskManagerInfo(ResourceManager.java:563)
> at sun.reflect.GeneratedMethodAccessor43.invoke(Unknown Source)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498)
> at
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcInvocation(AkkaRpcActor.java:247)
> at
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:162)
> at
> org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:70)
> at
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.onReceive(AkkaRpcActor.java:142)
> at
> org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.onReceive(FencedAkkaRpcActor.java:40)
> at
> akka.actor.UntypedActor$$anonfun$receive$1.applyOrElse(UntypedActor.scala:165)
> at akka.actor.Actor$class.aroundReceive(Actor.scala:502)
> at akka.actor.UntypedActor.aroundReceive(UntypedActor.scala:95)
> at akka.actor.ActorCell.receiveMessage(ActorCell.scala:526)
> at akka.actor.ActorCell.invoke(ActorCell.scala:495)
> at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:257)
> at akka.dispatch.Mailbox.run(Mailbox.scala:224)
> at akka.dispatch.Mailbox.exec(Mailbox.scala:234)
> at
> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
> at
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
> at
> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
> at
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
>
> ========================================
> 相关代码:
> final StreamExecutionEnvironment env =
> StreamExecutionEnvironment.getExecutionEnvironment();
> env.enableCheckpointing(1000);
> Properties properties1 = new Properties();
> properties1.setProperty("bootstrap.servers", "xxx:9092");
> properties1.setProperty("group.id", "test");
>
> System.out.println("333333333333333333333333");
>
> FlinkKafkaConsumer010<String> myConsumer = new
> FlinkKafkaConsumer010<String>("bigeyeservertopictest", new
> SimpleStringSchema(), properties1);
>
> DataStream<String> stream = env.addSource(myConsumer);
> // stream.print();
> DataStream<String> student = stream.flatMap(new
> StringToJsonObject()).map(value -> value.toJSONArraytoString())
> .map(new MapFunction<JSONArray, String>() {
> public String map(JSONArray jsonArray) throws
> Exception{
> String source= "bigeye";
> JSONObject jsonObject1 = new JSONObject();
> String result = null;
> // readRedis redis = new readRedis();
> for(int i=0;i<jsonArray.size();i++){
> // log.info("1111111111111111");
> String jsonObjectStr = jsonArray.getString(i);
> String timeValue =
> JSONObject.parseObject(jsonObjectStr).getString("time");
> String valueValue =
> JSONObject.parseObject(jsonObjectStr).getString("value");
> String agentHostnameValue =
> JSONObject.parseObject(jsonObjectStr).getString("agentHostname");
> String streamIDValue =
> JSONObject.parseObject(jsonObjectStr).getString("streamID");
> // Map<String, String> redisMap =
> redis.getInfo(agentHostnameValue,streamIDValue,source);
> // 创建opentsdb数据对象
> Map<String, Object> opentsdbValue = new
> HashMap<>();
> opentsdbValue.put("metric", streamIDValue);
> opentsdbValue.put("timestamp", timeValue);
> opentsdbValue.put("value", valueValue);
> Gson gson = new Gson();
> // opentsdbValue.put("tags",
> gson.fromJson((String) redisMap.get("tags"), Map.class));
>
> //
> jsonObject1.put(flag,timeValue+"+"+streamIDValue);
> jsonObject1.put("BigeyeValue",opentsdbValue);
> // redis.close();
>
> }
>
> result = jsonObject1.toString();
> return result;
> }
> })
> ;
> student.addSink(new
> FlinkKafkaProducer010<String>("xxxx:9092","test", new
> SimpleStringSchema())).name("flink-kafka");
>
> 万分感谢!
>
>
>
> [email protected]
>