See
<https://ci-beam.apache.org/job/beam_PostCommit_PortableJar_Flink/3040/display/redirect?page=changes>
Changes:
[Kyle Weaver] [BEAM-10925] SQL AggregateFn: Separate mutable accumulator from
[noreply] [BEAM-11329]:HDFS not deduplicating identical configuration paths
------------------------------------------
[...truncated 532.57 KB...]
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.SingleThreadEventExecutor$6.run(SingleThreadEventExecutor.java:1081)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.lang.Thread.run(Thread.java:748)
Exception in thread "grpc-nio-worker-ELG-3-3" Exception in thread
"grpc-nio-worker-ELG-3-6" java.lang.IllegalAccessError: tried to access class
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolArena$1 from class
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolArena
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolArena.freeChunk(PoolArena.java:296)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache$MemoryRegionCache.freeEntry(PoolThreadCache.java:464)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache$MemoryRegionCache.free(PoolThreadCache.java:430)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache$MemoryRegionCache.free(PoolThreadCache.java:422)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache.free(PoolThreadCache.java:279)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache.free(PoolThreadCache.java:270)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache.free(PoolThreadCache.java:241)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PooledByteBufAllocator$PoolThreadLocalCache.onRemoval(PooledByteBufAllocator.java:477)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PooledByteBufAllocator$PoolThreadLocalCache.onRemoval(PooledByteBufAllocator.java:444)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.FastThreadLocal.remove(FastThreadLocal.java:256)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.FastThreadLocal.removeAll(FastThreadLocal.java:67)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.SingleThreadEventExecutor$6.run(SingleThreadEventExecutor.java:1081)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.lang.Thread.run(Thread.java:748)
Exception in thread "grpc-nio-worker-ELG-3-4" java.lang.IllegalAccessError:
tried to access class
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolArena$1 from class
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolArena
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolArena.freeChunk(PoolArena.java:296)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache$MemoryRegionCache.freeEntry(PoolThreadCache.java:464)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache$MemoryRegionCache.free(PoolThreadCache.java:430)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache$MemoryRegionCache.free(PoolThreadCache.java:422)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache.free(PoolThreadCache.java:279)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache.free(PoolThreadCache.java:270)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache.free(PoolThreadCache.java:241)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PooledByteBufAllocator$PoolThreadLocalCache.onRemoval(PooledByteBufAllocator.java:477)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PooledByteBufAllocator$PoolThreadLocalCache.onRemoval(PooledByteBufAllocator.java:444)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.FastThreadLocal.remove(FastThreadLocal.java:256)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.FastThreadLocal.removeAll(FastThreadLocal.java:67)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.SingleThreadEventExecutor$6.run(SingleThreadEventExecutor.java:1081)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.lang.Thread.run(Thread.java:748)
java.lang.IllegalAccessError: tried to access class
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolArena$1 from class
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolArena
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolArena.freeChunk(PoolArena.java:296)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache$MemoryRegionCache.freeEntry(PoolThreadCache.java:464)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache$MemoryRegionCache.free(PoolThreadCache.java:430)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache$MemoryRegionCache.free(PoolThreadCache.java:422)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache.free(PoolThreadCache.java:279)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache.free(PoolThreadCache.java:270)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache.free(PoolThreadCache.java:241)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PooledByteBufAllocator$PoolThreadLocalCache.onRemoval(PooledByteBufAllocator.java:477)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PooledByteBufAllocator$PoolThreadLocalCache.onRemoval(PooledByteBufAllocator.java:444)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.FastThreadLocal.remove(FastThreadLocal.java:256)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.FastThreadLocal.removeAll(FastThreadLocal.java:67)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.SingleThreadEventExecutor$6.run(SingleThreadEventExecutor.java:1081)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.lang.Thread.run(Thread.java:748)
java.lang.IllegalAccessError: tried to access class
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolArena$1 from class
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolArena
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolArena.freeChunk(PoolArena.java:296)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache$MemoryRegionCache.freeEntry(PoolThreadCache.java:464)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache$MemoryRegionCache.free(PoolThreadCache.java:430)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache$MemoryRegionCache.free(PoolThreadCache.java:422)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache.free(PoolThreadCache.java:279)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache.free(PoolThreadCache.java:270)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache.free(PoolThreadCache.java:241)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PooledByteBufAllocator$PoolThreadLocalCache.onRemoval(PooledByteBufAllocator.java:477)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PooledByteBufAllocator$PoolThreadLocalCache.onRemoval(PooledByteBufAllocator.java:444)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.FastThreadLocal.remove(FastThreadLocal.java:256)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.FastThreadLocal.removeAll(FastThreadLocal.java:67)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.SingleThreadEventExecutor$6.run(SingleThreadEventExecutor.java:1081)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.lang.Thread.run(Thread.java:748)
Dec 19, 2020 12:14:39 AM
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.DefaultPromise
safeExecute
SEVERE: Failed to submit a listener notification task. Event loop shut down?
java.lang.IllegalAccessError: tried to access class
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.GlobalEventExecutor$2
from class
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.GlobalEventExecutor
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.GlobalEventExecutor.startThread(GlobalEventExecutor.java:227)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.GlobalEventExecutor.execute(GlobalEventExecutor.java:215)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.DefaultPromise.safeExecute(DefaultPromise.java:841)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:498)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:615)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.DefaultPromise.setSuccess0(DefaultPromise.java:604)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.DefaultPromise.setSuccess(DefaultPromise.java:96)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.SingleThreadEventExecutor$6.run(SingleThreadEventExecutor.java:1089)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.lang.Thread.run(Thread.java:748)
Exception in thread "grpc-nio-worker-ELG-3-1" java.lang.IllegalAccessError:
tried to access class
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolArena$1 from class
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolArena
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolArena.freeChunk(PoolArena.java:296)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache$MemoryRegionCache.freeEntry(PoolThreadCache.java:464)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache$MemoryRegionCache.free(PoolThreadCache.java:430)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache$MemoryRegionCache.free(PoolThreadCache.java:422)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache.free(PoolThreadCache.java:279)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache.free(PoolThreadCache.java:270)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PoolThreadCache.free(PoolThreadCache.java:241)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PooledByteBufAllocator$PoolThreadLocalCache.onRemoval(PooledByteBufAllocator.java:477)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.buffer.PooledByteBufAllocator$PoolThreadLocalCache.onRemoval(PooledByteBufAllocator.java:444)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.FastThreadLocal.remove(FastThreadLocal.java:256)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.FastThreadLocal.removeAll(FastThreadLocal.java:67)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.SingleThreadEventExecutor$6.run(SingleThreadEventExecutor.java:1081)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at
org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.lang.Thread.run(Thread.java:748)
Dec 19, 2020 12:14:40 AM org.apache.flink.runtime.taskmanager.Task
transitionState
INFO: DataSink (DiscardingOutput) (1/1) (e3ffbde7ba3ef51acd6c4342fac4bb3f)
switched from RUNNING to FINISHED.
Dec 19, 2020 12:14:40 AM org.apache.flink.runtime.taskmanager.Task doRun
INFO: Freeing task resources for DataSink (DiscardingOutput) (1/1)
(e3ffbde7ba3ef51acd6c4342fac4bb3f).
Dec 19, 2020 12:14:40 AM org.apache.flink.runtime.taskmanager.Task doRun
INFO: Ensuring all FileSystem streams are closed for task DataSink
(DiscardingOutput) (1/1) (e3ffbde7ba3ef51acd6c4342fac4bb3f) [FINISHED]
Dec 19, 2020 12:14:40 AM org.apache.flink.runtime.taskexecutor.TaskExecutor
unregisterTaskAndNotifyFinalState
INFO: Un-registering task and sending final execution state FINISHED to
JobManager for task DataSink (DiscardingOutput) (1/1)
e3ffbde7ba3ef51acd6c4342fac4bb3f.
Dec 19, 2020 12:14:40 AM org.apache.flink.runtime.executiongraph.Execution
transitionState
INFO: DataSink (DiscardingOutput) (1/1) (e3ffbde7ba3ef51acd6c4342fac4bb3f)
switched from RUNNING to FINISHED.
Dec 19, 2020 12:14:40 AM org.apache.flink.runtime.executiongraph.ExecutionGraph
transitionState
INFO: Job Flink Java Job at Sat Dec 19 00:14:13 UTC 2020
(8f45cafc47677ffc97e451a7b850e1b2) switched from state RUNNING to FINISHED.
Dec 19, 2020 12:14:40 AM org.apache.flink.runtime.dispatcher.Dispatcher
jobReachedGloballyTerminalState
INFO: Job 8f45cafc47677ffc97e451a7b850e1b2 reached globally terminal state
FINISHED.
Dec 19, 2020 12:14:40 AM org.apache.flink.runtime.jobmaster.JobMaster onStop
INFO: Stopping the JobMaster for job Flink Java Job at Sat Dec 19 00:14:13 UTC
2020(8f45cafc47677ffc97e451a7b850e1b2).
Dec 19, 2020 12:14:40 AM
org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl suspend
INFO: Suspending SlotPool.
Dec 19, 2020 12:14:40 AM org.apache.flink.runtime.jobmaster.JobMaster
dissolveResourceManagerConnection
INFO: Close ResourceManager connection c91efb5483f53735073c3837c129b69c:
JobManager is shutting down..
Dec 19, 2020 12:14:40 AM
org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl close
INFO: Stopping SlotPool.
Dec 19, 2020 12:14:40 AM
org.apache.flink.runtime.resourcemanager.ResourceManager
closeJobManagerConnection
INFO: Disconnect job manager
b3874573fbfb777898d50f71b7104442@akka://flink/user/jobmanager_1 for job
8f45cafc47677ffc97e451a7b850e1b2 from the resource manager.
Dec 19, 2020 12:14:40 AM
org.apache.flink.runtime.taskexecutor.slot.TaskSlotTableImpl freeSlotInternal
INFO: Free slot TaskSlot(index:0, state:ACTIVE, resource profile:
ResourceProfile{managedMemory=128.000mb (134217728 bytes),
networkMemory=64.000mb (67108864 bytes)}, allocationId:
fc8a35a7c50ee2b4a1faa87b938b70ee, jobId: 8f45cafc47677ffc97e451a7b850e1b2).
Dec 19, 2020 12:14:40 AM
org.apache.flink.runtime.taskexecutor.TaskExecutor$JobLeaderListenerImpl
jobManagerLostLeadership
INFO: JobManager for job 8f45cafc47677ffc97e451a7b850e1b2 with leader id
b3874573fbfb777898d50f71b7104442 lost leadership.
Dec 19, 2020 12:14:40 AM org.apache.flink.runtime.taskexecutor.JobLeaderService
removeJob
INFO: Remove job 8f45cafc47677ffc97e451a7b850e1b2 from job leader monitoring.
Dec 19, 2020 12:14:40 AM org.apache.flink.runtime.taskexecutor.TaskExecutor
closeJobManagerConnection
INFO: Close JobManager connection for job 8f45cafc47677ffc97e451a7b850e1b2.
Dec 19, 2020 12:14:40 AM org.apache.flink.runtime.taskexecutor.TaskExecutor
closeJobManagerConnection
INFO: Close JobManager connection for job 8f45cafc47677ffc97e451a7b850e1b2.
Dec 19, 2020 12:14:40 AM org.apache.flink.runtime.taskexecutor.JobLeaderService
reconnect
INFO: Cannot reconnect to job 8f45cafc47677ffc97e451a7b850e1b2 because it is
not registered.
Dec 19, 2020 12:14:40 AM org.apache.flink.runtime.taskexecutor.TaskExecutor
closeJobManagerConnection
INFO: Close JobManager connection for job 8f45cafc47677ffc97e451a7b850e1b2.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
kill %1 || echo "Failed to shut down Flink mini cluster"
rm -rf "$ENV_DIR"
Dec 19, 2020 12:14:41 AM
org.apache.flink.runtime.state.TaskExecutorLocalStateStoresManager shutdown
INFO: Shutting down TaskExecutorLocalStateStoresManager.
if [[ "$TEST_EXIT_CODE" -eq 0 ]]; then
echo ">>> SUCCESS"
else
echo ">>> FAILURE"
fi
exit $TEST_EXIT_CODE
> Task
> :sdks:python:test-suites:portable:py37:testUberJarFlinkRunnerSaveMainSession
Dec 19, 2020 12:14:42 AM org.apache.flink.runtime.taskmanager.Task
transitionState
INFO: DataSink (DiscardingOutput) (1/1) (1f753fd6eed7523cf5d51dba94420769)
switched from RUNNING to FINISHED.
Dec 19, 2020 12:14:42 AM org.apache.flink.runtime.taskmanager.Task doRun
INFO: Freeing task resources for DataSink (DiscardingOutput) (1/1)
(1f753fd6eed7523cf5d51dba94420769).
Dec 19, 2020 12:14:42 AM org.apache.flink.runtime.taskmanager.Task doRun
INFO: Ensuring all FileSystem streams are closed for task DataSink
(DiscardingOutput) (1/1) (1f753fd6eed7523cf5d51dba94420769) [FINISHED]
Dec 19, 2020 12:14:42 AM org.apache.flink.runtime.taskexecutor.TaskExecutor
unregisterTaskAndNotifyFinalState
INFO: Un-registering task and sending final execution state FINISHED to
JobManager for task DataSink (DiscardingOutput) (1/1)
1f753fd6eed7523cf5d51dba94420769.
Dec 19, 2020 12:14:42 AM org.apache.flink.runtime.executiongraph.Execution
transitionState
INFO: DataSink (DiscardingOutput) (1/1) (1f753fd6eed7523cf5d51dba94420769)
switched from RUNNING to FINISHED.
Dec 19, 2020 12:14:42 AM org.apache.flink.runtime.executiongraph.ExecutionGraph
transitionState
INFO: Job Flink Java Job at Sat Dec 19 00:14:12 UTC 2020
(888946415a4871db09fc747e2da4ae93) switched from state RUNNING to FINISHED.
Dec 19, 2020 12:14:42 AM org.apache.flink.runtime.dispatcher.Dispatcher
jobReachedGloballyTerminalState
INFO: Job 888946415a4871db09fc747e2da4ae93 reached globally terminal state
FINISHED.
Dec 19, 2020 12:14:42 AM org.apache.flink.runtime.jobmaster.JobMaster onStop
INFO: Stopping the JobMaster for job Flink Java Job at Sat Dec 19 00:14:12 UTC
2020(888946415a4871db09fc747e2da4ae93).
Dec 19, 2020 12:14:42 AM
org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl suspend
INFO: Suspending SlotPool.
Dec 19, 2020 12:14:42 AM org.apache.flink.runtime.jobmaster.JobMaster
dissolveResourceManagerConnection
INFO: Close ResourceManager connection 1a6ba21d1546b21c2fad41e362402e07:
JobManager is shutting down..
Dec 19, 2020 12:14:42 AM
org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl close
INFO: Stopping SlotPool.
Dec 19, 2020 12:14:42 AM
org.apache.flink.runtime.resourcemanager.ResourceManager
closeJobManagerConnection
INFO: Disconnect job manager
9e2888c1bccfe222dcf2696e5ab14ea1@akka://flink/user/jobmanager_1 for job
888946415a4871db09fc747e2da4ae93 from the resource manager.
Dec 19, 2020 12:14:42 AM
org.apache.flink.runtime.taskexecutor.slot.TaskSlotTableImpl freeSlotInternal
INFO: Free slot TaskSlot(index:0, state:ACTIVE, resource profile:
ResourceProfile{managedMemory=128.000mb (134217728 bytes),
networkMemory=64.000mb (67108864 bytes)}, allocationId:
5287ef9aba77cce1130421da7bd517bc, jobId: 888946415a4871db09fc747e2da4ae93).
Dec 19, 2020 12:14:42 AM
org.apache.flink.runtime.taskexecutor.TaskExecutor$JobLeaderListenerImpl
jobManagerLostLeadership
INFO: JobManager for job 888946415a4871db09fc747e2da4ae93 with leader id
9e2888c1bccfe222dcf2696e5ab14ea1 lost leadership.
Dec 19, 2020 12:14:42 AM org.apache.flink.runtime.taskexecutor.JobLeaderService
removeJob
INFO: Remove job 888946415a4871db09fc747e2da4ae93 from job leader monitoring.
Dec 19, 2020 12:14:42 AM org.apache.flink.runtime.taskexecutor.TaskExecutor
closeJobManagerConnection
INFO: Close JobManager connection for job 888946415a4871db09fc747e2da4ae93.
Dec 19, 2020 12:14:42 AM org.apache.flink.runtime.taskexecutor.TaskExecutor
closeJobManagerConnection
INFO: Close JobManager connection for job 888946415a4871db09fc747e2da4ae93.
Dec 19, 2020 12:14:42 AM org.apache.flink.runtime.taskexecutor.JobLeaderService
reconnect
INFO: Cannot reconnect to job 888946415a4871db09fc747e2da4ae93 because it is
not registered.
Dec 19, 2020 12:14:42 AM org.apache.flink.runtime.taskexecutor.TaskExecutor
closeJobManagerConnection
INFO: Close JobManager connection for job 888946415a4871db09fc747e2da4ae93.
INFO:apache_beam.runners.portability.portable_runner:Job state changed to DONE
kill %1 || echo "Failed to shut down Flink mini cluster"
rm -rf "$ENV_DIR"
Dec 19, 2020 12:14:46 AM
org.apache.flink.runtime.state.TaskExecutorLocalStateStoresManager shutdown
INFO: Shutting down TaskExecutorLocalStateStoresManager.
if [[ "$TEST_EXIT_CODE" -eq 0 ]]; then
echo ">>> SUCCESS"
else
echo ">>> FAILURE"
fi
exit $TEST_EXIT_CODE
FAILURE: Build completed with 2 failures.
1: Task failed with an exception.
-----------
* Where:
Script
'<https://ci-beam.apache.org/job/beam_PostCommit_PortableJar_Flink/ws/src/sdks/python/test-suites/portable/common.gradle'>
line: 202
* What went wrong:
Execution failed for task
':sdks:python:test-suites:portable:py37:testJavaJarCreatorFlinkRunner'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug
option to get more log output. Run with --scan to get full insights.
==============================================================================
2: Task failed with an exception.
-----------
* Where:
Script
'<https://ci-beam.apache.org/job/beam_PostCommit_PortableJar_Flink/ws/src/sdks/python/test-suites/portable/common.gradle'>
line: 202
* What went wrong:
Execution failed for task
':sdks:python:test-suites:portable:py36:testJavaJarCreatorFlinkRunner'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug
option to get more log output. Run with --scan to get full insights.
==============================================================================
* Get more help at https://help.gradle.org
Deprecated Gradle features were used in this build, making it incompatible with
Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See
https://docs.gradle.org/6.7/userguide/command_line_interface.html#sec:command_line_warnings
BUILD FAILED in 14m 24s
154 actionable tasks: 112 executed, 39 from cache, 3 up-to-date
Gradle was unable to watch the file system for changes. The inotify watches
limit is too low.
Publishing build scan...
https://gradle.com/s/ym4cg7at4zncu
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]