Jenkins build is back to normal : beam_PostCommit_Java11_ValidatesRunner_Direct #2375

2019-11-05 Thread Apache Jenkins Server
See 



-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org



Build failed in Jenkins: beam_PostCommit_Python35 #913

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[chambers] Fix expected step name in Python PTransformTest test_read_metrics

[chambers] generalized step name matching


--
[...truncated 130.28 KB...]
},
"output_name": "out",
"user_name": "ReadFromPubSub/Read.out"
  }
],
"pubsub_subscription": 
"projects/apache-beam-testing/subscriptions/exercise_streaming_metrics_subscription_input649a2f12-7079-4ee5-a32f-d6f6793fa787",
"user_name": "ReadFromPubSub/Read"
  }
},
{
  "kind": "ParallelDo",
  "name": "s2",
  "properties": {
"display_data": [
  {
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.ParDo",
"shortValue": "StreamingUserMetricsDoFn",
"type": "STRING",
"value": 
"apache_beam.runners.dataflow.dataflow_exercise_streaming_metrics_pipeline.StreamingUserMetricsDoFn"
  }
],
"non_parallel_inputs": {},
"output_info": [
  {
"encoding": {
  "@type": "kind:windowed_value",
  "component_encodings": [
{
  "@type": "kind:bytes"
},
{
  "@type": "kind:global_window"
}
  ],
  "is_wrapper": true
},
"output_name": "out",
"user_name": "generate_metrics.out"
  }
],
"parallel_input": {
  "@type": "OutputReference",
  "output_name": "out",
  "step_name": "s1"
},
"serialized_fn": "ref_AppliedPTransform_generate_metrics_4",
"user_name": "generate_metrics"
  }
},
{
  "kind": "ParallelWrite",
  "name": "s3",
  "properties": {
"display_data": [],
"encoding": {
  "@type": "kind:windowed_value",
  "component_encodings": [
{
  "@type": "kind:bytes"
},
{
  "@type": "kind:global_window"
}
  ],
  "is_wrapper": true
},
"format": "pubsub",
"parallel_input": {
  "@type": "OutputReference",
  "output_name": "out",
  "step_name": "s2"
},
"pubsub_topic": 
"projects/apache-beam-testing/topics/exercise_streaming_metrics_topic_output649a2f12-7079-4ee5-a32f-d6f6793fa787",
"user_name": "dump_to_pub/Write/NativeWrite"
  }
}
  ],
  "type": "JOB_TYPE_STREAMING"
}
root: INFO: Create job: 
root: INFO: Created job with id: [2019-11-05_12_41_36-17104288937130719037]
root: INFO: To access the Dataflow monitoring console, please navigate to 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_12_41_36-17104288937130719037?project=apache-beam-testing
root: INFO: Job 2019-11-05_12_41_36-17104288937130719037 is in state 
JOB_STATE_RUNNING
root: INFO: 2019-11-05T20:41:40.308Z: JOB_MESSAGE_DETAILED: Checking 
permissions granted to controller Service Account.
root: INFO: 2019-11-05T20:41:41.400Z: JOB_MESSAGE_BASIC: Worker configuration: 
n1-standard-4 in us-central1-f.
root: INFO: 2019-11-05T20:41:42.049Z: JOB_MESSAGE_DETAILED: Expanding 
SplittableParDo operations into optimizable parts.
root: INFO: 2019-11-05T20:41:42.051Z: JOB_MESSAGE_DETAILED: Expanding 
CollectionToSingleton operations into optimizable parts.
root: INFO: 2019-11-05T20:41:42.059Z: JOB_MESSAGE_DETAILED: Expanding 
CoGroupByKey operations into optimizable parts.
root: INFO: 2019-11-05T20:41:42.069Z: JOB_MESSAGE_DETAILED: Expanding 
SplittableProcessKeyed operations into optimizable parts.
root: INFO: 2019-11-05T20:41:42.071Z: JOB_MESSAGE_DETAILED: Expanding 
GroupByKey operations into streaming Read/Write steps
root: INFO: 2019-11-05T20:41:42.075Z: JOB_MESSAGE_DEBUG: Annotating graph with 
Autotuner information.
root: INFO: 2019-11-05T20:41:42.091Z: JOB_MESSAGE_DETAILED: Fusing adjacent 
ParDo, Read, Write, and Flatten operations
root: INFO: 2019-11-05T20:41:42.093Z: JOB_MESSAGE_DETAILED: Fusing consumer 
generate_metrics into ReadFromPubSub/Read
root: INFO: 2019-11-05T20:41:42.096Z: JOB_MESSAGE_DETAILED: Fusing consumer 
dump_to_pub/Write/NativeWrite into generate_metrics
root: INFO: 2019-11-05T20:41:42.104Z: JOB_MESSAGE_DEBUG: Adding StepResource 
setup and teardown to workflow graph.
root: INFO: 2019-11-05T20:41:42.119Z: JOB_MESSAGE_DEBUG: Adding workflow start 
and stop steps.
root: INFO: 2019-11-05T20:41:42.156Z: JOB_MESSAGE_DEBUG: Assigning stage ids.
root: INFO: 2019-11-05T20:41:42.370Z: JOB_MESSAGE_DEBUG: Executing wait step 
start2
root: INFO: 2019-11-05T20:41:42.386Z: JOB_MESSAGE_DEBUG: Starting worker pool 
setup.
root: INFO: 2019-11-05T20:41:42.392Z: JOB_MESSAGE_BASIC: Starting 1 workers...
root: INFO: 

Build failed in Jenkins: beam_PostCommit_Python37 #871

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[github] Removing some trailing whitespace.


--
[...truncated 231.30 KB...]
datanode_1  | 19/11/05 23:33:15 INFO ipc.CallQueueManager: Using callQueue: 
class java.util.concurrent.LinkedBlockingQueue queueCapacity: 1000 scheduler: 
class org.apache.hadoop.ipc.DefaultRpcScheduler
datanode_1  | 19/11/05 23:33:15 INFO ipc.Server: Starting Socket Reader #1 for 
port 50020
datanode_1  | 19/11/05 23:33:15 INFO datanode.DataNode: Opened IPC server at 
/0.0.0.0:50020
datanode_1  | 19/11/05 23:33:16 INFO datanode.DataNode: Refresh request 
received for nameservices: null
datanode_1  | 19/11/05 23:33:16 INFO datanode.DataNode: Starting 
BPOfferServices for nameservices: 
datanode_1  | 19/11/05 23:33:16 INFO datanode.DataNode: Block pool 
 (Datanode Uuid unassigned) service to namenode/192.168.208.2:8020 
starting to offer service
datanode_1  | 19/11/05 23:33:16 INFO ipc.Server: IPC Server Responder: starting
datanode_1  | 19/11/05 23:33:16 INFO ipc.Server: IPC Server listener on 50020: 
starting
namenode_1  | 19/11/05 23:33:16 INFO namenode.NameCache: initialized with 0 
entries 0 lookups
namenode_1  | 19/11/05 23:33:16 INFO namenode.FSNamesystem: Finished loading 
FSImage in 374 msecs
test_1  | Waiting for safe mode to end.
namenode_1  | 19/11/05 23:33:16 INFO namenode.NameNode: RPC server is binding 
to 0.0.0.0:8020
namenode_1  | 19/11/05 23:33:16 INFO ipc.CallQueueManager: Using callQueue: 
class java.util.concurrent.LinkedBlockingQueue queueCapacity: 1000 scheduler: 
class org.apache.hadoop.ipc.DefaultRpcScheduler
namenode_1  | 19/11/05 23:33:16 INFO ipc.Server: Starting Socket Reader #1 for 
port 8020
namenode_1  | 19/11/05 23:33:16 INFO namenode.FSNamesystem: Registered 
FSNamesystemState MBean
namenode_1  | 19/11/05 23:33:16 INFO namenode.LeaseManager: Number of blocks 
under construction: 0
namenode_1  | 19/11/05 23:33:16 INFO blockmanagement.BlockManager: initializing 
replication queues
namenode_1  | 19/11/05 23:33:16 INFO hdfs.StateChange: STATE* Leaving safe mode 
after 0 secs
namenode_1  | 19/11/05 23:33:16 INFO hdfs.StateChange: STATE* Network topology 
has 0 racks and 0 datanodes
namenode_1  | 19/11/05 23:33:16 INFO hdfs.StateChange: STATE* 
UnderReplicatedBlocks has 0 blocks
namenode_1  | 19/11/05 23:33:16 INFO blockmanagement.BlockManager: Total number 
of blocks= 0
namenode_1  | 19/11/05 23:33:16 INFO blockmanagement.BlockManager: Number of 
invalid blocks  = 0
namenode_1  | 19/11/05 23:33:16 INFO blockmanagement.BlockManager: Number of 
under-replicated blocks = 0
namenode_1  | 19/11/05 23:33:16 INFO blockmanagement.BlockManager: Number of  
over-replicated blocks = 0
namenode_1  | 19/11/05 23:33:16 INFO blockmanagement.BlockManager: Number of 
blocks being written= 0
namenode_1  | 19/11/05 23:33:16 INFO hdfs.StateChange: STATE* Replication Queue 
initialization scan for invalid, over- and under-replicated blocks completed in 
12 msec
namenode_1  | 19/11/05 23:33:16 INFO ipc.Server: IPC Server Responder: starting
namenode_1  | 19/11/05 23:33:16 INFO ipc.Server: IPC Server listener on 8020: 
starting
namenode_1  | 19/11/05 23:33:16 INFO namenode.NameNode: NameNode RPC up at: 
namenode/192.168.208.2:8020
namenode_1  | 19/11/05 23:33:16 INFO namenode.FSNamesystem: Starting services 
required for active state
namenode_1  | 19/11/05 23:33:16 INFO namenode.FSDirectory: Initializing quota 
with 4 thread(s)
namenode_1  | 19/11/05 23:33:16 INFO namenode.FSDirectory: Quota initialization 
completed in 4 milliseconds
namenode_1  | name space=1
namenode_1  | storage space=0
namenode_1  | storage types=RAM_DISK=0, SSD=0, DISK=0, ARCHIVE=0
namenode_1  | 19/11/05 23:33:16 INFO blockmanagement.CacheReplicationMonitor: 
Starting CacheReplicationMonitor with interval 3 milliseconds
datanode_1  | 19/11/05 23:33:17 INFO ipc.Client: Retrying connect to server: 
namenode/192.168.208.2:8020. Already tried 0 time(s); retry policy is 
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)
datanode_1  | 19/11/05 23:33:17 INFO datanode.DataNode: Acknowledging ACTIVE 
Namenode during handshakeBlock pool  (Datanode Uuid unassigned) 
service to namenode/192.168.208.2:8020
datanode_1  | 19/11/05 23:33:17 INFO common.Storage: Using 1 threads to upgrade 
data directories (dfs.datanode.parallel.volumes.load.threads.num=1, dataDirs=1)
datanode_1  | 19/11/05 23:33:17 INFO common.Storage: Lock on 
/hadoop/dfs/data/in_use.lock acquired by nodename 82@datanode
datanode_1  | 19/11/05 23:33:17 INFO common.Storage: Storage directory 
/hadoop/dfs/data is not formatted for namespace 1981295257. Formatting...
datanode_1  | 19/11/05 23:33:17 INFO common.Storage: Generated new storageID 
DS-b1eb30d9-0f2e-4c2c-9bbc-99d9e42f6346 for directory /hadoop/dfs/data
datanode_1  | 19/11/05 23:33:17 INFO common.Storage: Analyzing storage 
directories 

Build failed in Jenkins: beam_PostCommit_Py_VR_Dataflow #5006

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[github] Removing some trailing whitespace.


--
[...truncated 148.93 KB...]
test_as_singleton_without_unique_labels 
(apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_iterable_side_input 
(apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_flattened_side_input 
(apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_multi_valued_singleton_side_input 
(apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok

--
XML: nosetests-validatesRunnerBatchTests-df-py35.xml
--
XML: 

--
Ran 19 tests in 1574.573s

OK

> Task :sdks:python:test-suites:dataflow:py35:validatesRunnerStreamingTests
>>> RUNNING integration tests with pipeline options: 
>>> --runner=TestDataflowRunner --project=apache-beam-testing 
>>> --staging_location=gs://temp-storage-for-end-to-end-tests/staging-it 
>>> --temp_location=gs://temp-storage-for-end-to-end-tests/temp-it 
>>> --output=gs://temp-storage-for-end-to-end-tests/py-it-cloud/output 
>>> --sdk_location=
>>>  --requirements_file=postcommit_requirements.txt --num_workers=1 
>>> --sleep_secs=20 --streaming 
>>> --dataflow_worker_jar=
>>>  
>>> --kms_key_name=projects/apache-beam-testing/locations/global/keyRings/beam-it/cryptoKeys/test
>>>  
>>> --dataflow_kms_key=projects/apache-beam-testing/locations/global/keyRings/beam-it/cryptoKeys/test
>>>   test options: --nocapture --processes=8 --process-timeout=4500 
>>> --attr=ValidatesRunner,!sickbay-streaming
running nosetests
running egg_info
writing entry points to apache_beam.egg-info/entry_points.txt
writing dependency_links to apache_beam.egg-info/dependency_links.txt
writing requirements to apache_beam.egg-info/requires.txt
writing top-level names to apache_beam.egg-info/top_level.txt
writing apache_beam.egg-info/PKG-INFO
reading manifest file 'apache_beam.egg-info/SOURCES.txt'
reading manifest template 'MANIFEST.in'
:475:
 UserWarning: Normalizing '2.18.0.dev' to '2.18.0.dev0'
  normalized_version,
warning: no files found matching 'README.md'
warning: no files found matching 'NOTICE'
warning: no files found matching 'LICENSE'
writing manifest file 'apache_beam.egg-info/SOURCES.txt'
:59:
 UserWarning: Datastore IO will support Python 3 after replacing 
googledatastore by google-cloud-datastore, see: BEAM-4543.
  warnings.warn('Datastore IO will support Python 3 after replacing '
:47:
 UserWarning: VCF IO will support Python 3 after migration to Nucleus, see: 
BEAM-5628.
  warnings.warn("VCF IO will support Python 3 after migration to Nucleus, "
:511:
 YAMLLoadWarning: calling yaml.load_all() without Loader=... is deprecated, as 
the default Loader is unsafe. Please read https://msg.pyyaml.org/load for full 
details.
  for spec in yaml.load_all(open(transcript_filename)):
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)

> Task :sdks:python:test-suites:dataflow:py2:validatesRunnerStreamingTests
test_dofn_lifecycle 
(apache_beam.transforms.dofn_lifecycle_test.DoFnLifecycleTest) ... ok
(unset)
test_multiple_empty_outputs 
(apache_beam.transforms.ptransform_test.PTransformTest) ... ok
test_par_do_with_multiple_outputs_and_using_return 
(apache_beam.transforms.ptransform_test.PTransformTest) ... ok
(unset)
(unset)
Runs streaming Dataflow job and verifies that user metrics are reported ... ok
(unset)
test_impulse (apache_beam.transforms.ptransform_test.PTransformTest) ... ok
(unset)
test_flatten_multiple_pcollections_having_multiple_consumers 
(apache_beam.transforms.ptransform_test.PTransformTest) ... ok
test_par_do_with_multiple_outputs_and_using_yield 
(apache_beam.transforms.ptransform_test.PTransformTest) ... ok
(unset)
(unset)
test_undeclared_outputs (apache_beam.transforms.ptransform_test.PTransformTest) 
... ok
(unset)

> Task 

Build failed in Jenkins: beam_PostCommit_Java_Nexmark_Flink #4223

2019-11-05 Thread Apache Jenkins Server
See 


Changes:


--
[...truncated 9.42 KB...]
> Task :runners:core-java:processResources NO-SOURCE
> Task :runners:direct-java:processResources NO-SOURCE
> Task :sdks:java:core:generateAvroJava NO-SOURCE
> Task :runners:flink:1.9:copyResourcesOverrides NO-SOURCE
> Task :model:fn-execution:extractProto
> Task :model:job-management:extractProto
> Task :sdks:java:extensions:join-library:processResources NO-SOURCE
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :sdks:java:io:mongodb:processResources NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:extractProto
> Task :sdks:java:io:parquet:processResources NO-SOURCE
> Task :sdks:java:testing:nexmark:processResources
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :model:fn-execution:processResources
> Task :model:job-management:processResources
> Task :runners:flink:1.9:copySourceOverrides
> Task :runners:flink:1.9:copyTestResourcesOverrides NO-SOURCE
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :runners:flink:1.9:processResources
> Task :sdks:java:build-tools:compileJava FROM-CACHE
> Task :sdks:java:build-tools:processResources
> Task :sdks:java:build-tools:classes
> Task :sdks:java:core:processResources
> Task :sdks:java:build-tools:jar
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromCalciteCore
> Task :model:pipeline:generateProto
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromSrc
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :sdks:java:extensions:sql:generateFmppSources
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:generateProto
> Task :model:fn-execution:generateProto
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes

> Task :sdks:java:extensions:sql:compileJavacc
Java Compiler Compiler Version 4.0 (Parser Generator)
(type "javacc" with no arguments for help)
Reading from file 

 . . .
Note: UNICODE_INPUT option is specified. Please make sure you create the 
parser/lexer using a Reader with the correct character encoding.
Warning: Lookahead adequacy checking not being performed since option LOOKAHEAD 
is more than 1.  Set option FORCE_LA_CHECK to true to force checking.
File "TokenMgrError.java" does not exist.  Will create one.
File "ParseException.java" does not exist.  Will create one.
File "Token.java" does not exist.  Will create one.
File "SimpleCharStream.java" does not exist.  Will create one.
Parser generated with 0 errors and 1 warnings.

> Task :sdks:java:extensions:sql:processResources
> Task :model:pipeline:shadowJar
> Task :model:job-management:shadowJar
> Task :model:fn-execution:shadowJar
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :runners:local-java:jar
> Task :sdks:java:io:mongodb:compileJava FROM-CACHE
> Task :sdks:java:io:mongodb:classes UP-TO-DATE
> Task :sdks:java:extensions:join-library:compileJava FROM-CACHE
> Task :sdks:java:extensions:join-library:classes UP-TO-DATE
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :sdks:java:extensions:join-library:jar
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :sdks:java:io:mongodb:jar
> Task :vendor:sdks-java-extensions-protobuf:compileJava FROM-CACHE
> Task :vendor:sdks-java-extensions-protobuf:classes UP-TO-DATE
> Task :sdks:java:io:kafka:jar
> Task :vendor:sdks-java-extensions-protobuf:shadowJar
> Task :sdks:java:core:jar
> Task :runners:core-construction-java:jar
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:fn-execution:jar
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:jar
> Task :sdks:java:io:parquet:compileJava FROM-CACHE
> Task 

Build failed in Jenkins: beam_PerformanceTests_MongoDBIO_IT #2355

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[chambers] Fix expected step name in Python PTransformTest test_read_metrics

[chambers] generalized step name matching

[github] Removing some trailing whitespace.


--
[...truncated 524.34 KB...]
at 
com.mongodb.internal.connection.InternalStreamConnection.open(InternalStreamConnection.java:126)
at 
com.mongodb.internal.connection.DefaultServerMonitor$ServerMonitorRunnable.run(DefaultServerMonitor.java:117)
at java.lang.Thread.run(Thread.java:748)
Caused by: java.net.UnknownHostException: mongodb-host
at java.net.InetAddress.getAllByName0(InetAddress.java:1281)
at java.net.InetAddress.getAllByName(InetAddress.java:1193)
at java.net.InetAddress.getAllByName(InetAddress.java:1127)
at com.mongodb.ServerAddress.getSocketAddresses(ServerAddress.java:203)
... 5 more

Nov 06, 2019 1:09:29 AM com.mongodb.diagnostics.logging.SLF4JLogger info
INFO: Exception in monitor thread while connecting to server 
mongodb-host:27017
com.mongodb.MongoSocketException: mongodb-host: Name or service not known
at com.mongodb.ServerAddress.getSocketAddresses(ServerAddress.java:211)
at 
com.mongodb.internal.connection.SocketStream.initializeSocket(SocketStream.java:75)
at 
com.mongodb.internal.connection.SocketStream.open(SocketStream.java:65)
at 
com.mongodb.internal.connection.InternalStreamConnection.open(InternalStreamConnection.java:126)
at 
com.mongodb.internal.connection.DefaultServerMonitor$ServerMonitorRunnable.run(DefaultServerMonitor.java:117)
at java.lang.Thread.run(Thread.java:748)
Caused by: java.net.UnknownHostException: mongodb-host: Name or service not 
known
at java.net.Inet6AddressImpl.lookupAllHostAddr(Native Method)
at java.net.InetAddress$2.lookupAllHostAddr(InetAddress.java:929)
at 
java.net.InetAddress.getAddressesFromNameService(InetAddress.java:1324)
at java.net.InetAddress.getAllByName0(InetAddress.java:1277)
at java.net.InetAddress.getAllByName(InetAddress.java:1193)
at java.net.InetAddress.getAllByName(InetAddress.java:1127)
at com.mongodb.ServerAddress.getSocketAddresses(ServerAddress.java:203)
... 5 more

Nov 06, 2019 1:09:29 AM com.mongodb.diagnostics.logging.SLF4JLogger info
INFO: Exception in monitor thread while connecting to server 
mongodb-host:27017
com.mongodb.MongoSocketException: mongodb-host
at com.mongodb.ServerAddress.getSocketAddresses(ServerAddress.java:211)
at 
com.mongodb.internal.connection.SocketStream.initializeSocket(SocketStream.java:75)
at 
com.mongodb.internal.connection.SocketStream.open(SocketStream.java:65)
at 
com.mongodb.internal.connection.InternalStreamConnection.open(InternalStreamConnection.java:126)
at 
com.mongodb.internal.connection.DefaultServerMonitor$ServerMonitorRunnable.run(DefaultServerMonitor.java:117)
at java.lang.Thread.run(Thread.java:748)
Caused by: java.net.UnknownHostException: mongodb-host
at java.net.InetAddress.getAllByName0(InetAddress.java:1281)
at java.net.InetAddress.getAllByName(InetAddress.java:1193)
at java.net.InetAddress.getAllByName(InetAddress.java:1127)
at com.mongodb.ServerAddress.getSocketAddresses(ServerAddress.java:203)
... 5 more

Nov 06, 2019 1:09:39 AM com.mongodb.diagnostics.logging.SLF4JLogger info
INFO: Exception in monitor thread while connecting to server 
mongodb-host:27017
com.mongodb.MongoSocketException: mongodb-host: Name or service not known
at com.mongodb.ServerAddress.getSocketAddresses(ServerAddress.java:211)
at 
com.mongodb.internal.connection.SocketStream.initializeSocket(SocketStream.java:75)
at 
com.mongodb.internal.connection.SocketStream.open(SocketStream.java:65)
at 
com.mongodb.internal.connection.InternalStreamConnection.open(InternalStreamConnection.java:126)
at 
com.mongodb.internal.connection.DefaultServerMonitor$ServerMonitorRunnable.run(DefaultServerMonitor.java:117)
at java.lang.Thread.run(Thread.java:748)
Caused by: java.net.UnknownHostException: mongodb-host: Name or service not 
known
at java.net.Inet6AddressImpl.lookupAllHostAddr(Native Method)
at java.net.InetAddress$2.lookupAllHostAddr(InetAddress.java:929)
at 
java.net.InetAddress.getAddressesFromNameService(InetAddress.java:1324)
at java.net.InetAddress.getAllByName0(InetAddress.java:1277)
at java.net.InetAddress.getAllByName(InetAddress.java:1193)
at java.net.InetAddress.getAllByName(InetAddress.java:1127)
at com.mongodb.ServerAddress.getSocketAddresses(ServerAddress.java:203)
... 5 more

Nov 06, 2019 1:09:39 AM com.mongodb.diagnostics.logging.SLF4JLogger info

Build failed in Jenkins: beam_PostCommit_XVR_Flink #857

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[lostluck] [Go SDK] Correctly return EOFs from boolDecoder


--
[...truncated 4.28 MB...]
[flink-akka.actor.default-dispatcher-5] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (12/16) (144c5d75fef67d2f83fae3e5c354bd0d) switched from 
RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - MapPartition (MapPartition at 
[3]assert_that/{Group, Unkey, Match}) (2/16) (a6e5b29ec26585aa26bd098c7093d648) 
switched from RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - Freeing task resources for 
MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/16) 
(a6e5b29ec26585aa26bd098c7093d648).
[flink-akka.actor.default-dispatcher-5] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (2/16) (9c80110244b8f1a5854b959a5b923ebd) switched from 
CREATED to SCHEDULED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem 
streams are closed for task MapPartition (MapPartition at 
[3]assert_that/{Group, Unkey, Match}) (2/16) (a6e5b29ec26585aa26bd098c7093d648) 
[FINISHED]
[flink-akka.actor.default-dispatcher-4] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task MapPartition 
(MapPartition at [3]assert_that/{Group, Unkey, Match}) 
a6e5b29ec26585aa26bd098c7093d648.
[flink-akka.actor.default-dispatcher-5] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (2/16) (9c80110244b8f1a5854b959a5b923ebd) switched from 
SCHEDULED to DEPLOYING.
[flink-akka.actor.default-dispatcher-5] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - Deploying DataSink 
(DiscardingOutput) (2/16) (attempt #0) to 07319706-1e65-4eb9-87b3-053a0291781d 
@ localhost (dataPort=-1)
[flink-akka.actor.default-dispatcher-4] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Received task DataSink 
(DiscardingOutput) (2/16).
[flink-akka.actor.default-dispatcher-5] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - MapPartition 
(MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/16) 
(a6e5b29ec26585aa26bd098c7093d648) switched from RUNNING to FINISHED.
[DataSink (DiscardingOutput) (2/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (2/16) 
(9c80110244b8f1a5854b959a5b923ebd) switched from CREATED to DEPLOYING.
[DataSink (DiscardingOutput) (2/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Creating FileSystem stream leak 
safety net for task DataSink (DiscardingOutput) (2/16) 
(9c80110244b8f1a5854b959a5b923ebd) [DEPLOYING]
[DataSink (DiscardingOutput) (2/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Loading JAR files for task DataSink 
(DiscardingOutput) (2/16) (9c80110244b8f1a5854b959a5b923ebd) [DEPLOYING].
[DataSink (DiscardingOutput) (2/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Registering task at network: 
DataSink (DiscardingOutput) (2/16) (9c80110244b8f1a5854b959a5b923ebd) 
[DEPLOYING].
[DataSink (DiscardingOutput) (2/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (2/16) 
(9c80110244b8f1a5854b959a5b923ebd) switched from DEPLOYING to RUNNING.
[flink-akka.actor.default-dispatcher-7] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (2/16) (9c80110244b8f1a5854b959a5b923ebd) switched from 
DEPLOYING to RUNNING.
[DataSink (DiscardingOutput) (2/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (2/16) 
(9c80110244b8f1a5854b959a5b923ebd) switched from RUNNING to FINISHED.
[DataSink (DiscardingOutput) (2/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Freeing task resources for DataSink 
(DiscardingOutput) (2/16) (9c80110244b8f1a5854b959a5b923ebd).
[DataSink (DiscardingOutput) (2/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem streams are 
closed for task DataSink (DiscardingOutput) (2/16) 
(9c80110244b8f1a5854b959a5b923ebd) [FINISHED]
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task DataSink 
(DiscardingOutput) 9c80110244b8f1a5854b959a5b923ebd.
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (2/16) (9c80110244b8f1a5854b959a5b923ebd) switched from 
RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (1/16)] 

Jenkins build is back to normal : beam_PostCommit_Python35 #916

2019-11-05 Thread Apache Jenkins Server
See 



-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org



beam_PostCommit_Python36 - Build # 919 - Aborted

2019-11-05 Thread Apache Jenkins Server
The Apache Jenkins build system has built beam_PostCommit_Python36 (build #919)

Status: Aborted

Check console output at 
https://builds.apache.org/job/beam_PostCommit_Python36/919/ to view the results.

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org

Build failed in Jenkins: beam_PostCommit_XVR_Flink #856

2019-11-05 Thread Apache Jenkins Server
See 


Changes:


--
[...truncated 4.26 MB...]
[DataSink (DiscardingOutput) (8/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Loading JAR files for task DataSink 
(DiscardingOutput) (8/16) (e1a73c5303c734766fcf56e479140096) [DEPLOYING].
[DataSink (DiscardingOutput) (8/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Registering task at network: 
DataSink (DiscardingOutput) (8/16) (e1a73c5303c734766fcf56e479140096) 
[DEPLOYING].
[DataSink (DiscardingOutput) (8/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (8/16) 
(e1a73c5303c734766fcf56e479140096) switched from DEPLOYING to RUNNING.
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (8/16) (e1a73c5303c734766fcf56e479140096) switched from 
DEPLOYING to RUNNING.
[DataSink (DiscardingOutput) (8/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (8/16) 
(e1a73c5303c734766fcf56e479140096) switched from RUNNING to FINISHED.
[DataSink (DiscardingOutput) (8/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Freeing task resources for DataSink 
(DiscardingOutput) (8/16) (e1a73c5303c734766fcf56e479140096).
[DataSink (DiscardingOutput) (8/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem streams are 
closed for task DataSink (DiscardingOutput) (8/16) 
(e1a73c5303c734766fcf56e479140096) [FINISHED]
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task DataSink 
(DiscardingOutput) e1a73c5303c734766fcf56e479140096.
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (8/16) (e1a73c5303c734766fcf56e479140096) switched from 
RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (14/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - MapPartition (MapPartition at 
[3]assert_that/{Group, Unkey, Match}) (14/16) 
(fe9068de05b1c714a3da28f8bf671635) switched from RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (14/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - Freeing task resources for 
MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (14/16) 
(fe9068de05b1c714a3da28f8bf671635).
[flink-akka.actor.default-dispatcher-7] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (14/16) (8072e5c5ab322846a0275b9b6b3b8f19) switched from 
CREATED to SCHEDULED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (14/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem 
streams are closed for task MapPartition (MapPartition at 
[3]assert_that/{Group, Unkey, Match}) (14/16) 
(fe9068de05b1c714a3da28f8bf671635) [FINISHED]
[flink-akka.actor.default-dispatcher-7] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (14/16) (8072e5c5ab322846a0275b9b6b3b8f19) switched from 
SCHEDULED to DEPLOYING.
[flink-akka.actor.default-dispatcher-7] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - Deploying DataSink 
(DiscardingOutput) (14/16) (attempt #0) to ab16ecfc-4719-4481-acf2-e982df9cff22 
@ localhost (dataPort=-1)
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task MapPartition 
(MapPartition at [3]assert_that/{Group, Unkey, Match}) 
fe9068de05b1c714a3da28f8bf671635.
[flink-akka.actor.default-dispatcher-7] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - MapPartition 
(MapPartition at [3]assert_that/{Group, Unkey, Match}) (14/16) 
(fe9068de05b1c714a3da28f8bf671635) switched from RUNNING to FINISHED.
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Received task DataSink 
(DiscardingOutput) (14/16).
[DataSink (DiscardingOutput) (14/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (14/16) 
(8072e5c5ab322846a0275b9b6b3b8f19) switched from CREATED to DEPLOYING.
[DataSink (DiscardingOutput) (14/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Creating FileSystem stream leak 
safety net for task DataSink (DiscardingOutput) (14/16) 
(8072e5c5ab322846a0275b9b6b3b8f19) [DEPLOYING]
[DataSink (DiscardingOutput) (14/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Loading JAR files for task DataSink 
(DiscardingOutput) (14/16) (8072e5c5ab322846a0275b9b6b3b8f19) [DEPLOYING].
[DataSink (DiscardingOutput) (14/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Registering task at network: 
DataSink (DiscardingOutput) 

Build failed in Jenkins: beam_PreCommit_Java_Cron #2001

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[kirillkozlov] Created a MongoDbTable and a provider for it

[chambers] Fix expected step name in Python PTransformTest test_read_metrics

[github] Updated assertion format string to use [] vs ?

[github] [BEAM-8539] Provide an initial definition of all job states and the

[chambers] generalized step name matching

[github] Removing some trailing whitespace.


--
[...truncated 503.18 KB...]
> Task :sdks:java:extensions:sql:datacatalog:spotbugsMain
SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder".
SLF4J: Defaulting to no-operation (NOP) logger implementation
SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further 
details.

> Task :sdks:java:io:cassandra:test
> Task :sdks:java:io:cassandra:check
> Task :sdks:java:io:cassandra:build
> Task :sdks:java:io:cassandra:buildDependents

> Task :sdks:java:testing:nexmark:compileJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.

> Task :sdks:java:testing:nexmark:classes
> Task :sdks:java:testing:nexmark:jar
> Task :sdks:java:testing:nexmark:assemble
> Task :sdks:java:testing:nexmark:analyzeClassesDependencies SKIPPED
> Task :sdks:java:extensions:sql:zetasql:compileTestJava
> Task :sdks:java:io:amqp:test

> Task :sdks:java:extensions:sql:zetasql:compileTestJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.

> Task :sdks:java:extensions:sql:zetasql:testClasses
> Task :sdks:java:extensions:sql:zetasql:analyzeTestClassesDependencies SKIPPED
> Task :sdks:java:extensions:sql:zetasql:analyzeDependencies SKIPPED
> Task :sdks:java:extensions:sql:datacatalog:test
> Task :sdks:java:extensions:sql:datacatalog:check
> Task :sdks:java:extensions:sql:datacatalog:build
> Task :sdks:java:extensions:sql:datacatalog:buildDependents
> Task :runners:google-cloud-dataflow-java:buildDependents
> Task :sdks:java:extensions:sql:zetasql:checkstyleMain
> Task :sdks:java:extensions:sql:zetasql:checkstyleTest
> Task :sdks:java:io:amazon-web-services:check
> Task :sdks:java:io:amazon-web-services:build
> Task :sdks:java:io:amazon-web-services:buildDependents
> Task :runners:direct-java:test
> Task :sdks:java:extensions:sorter:check
> Task :sdks:java:extensions:sorter:build
> Task :sdks:java:extensions:sorter:buildDependents
> Task :sdks:java:extensions:sql:jdbc:shadowJar
> Task :sdks:java:io:clickhouse:test

> Task :sdks:java:testing:nexmark:compileTestJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.

> Task :sdks:java:testing:nexmark:testClasses
> Task :sdks:java:testing:nexmark:analyzeTestClassesDependencies SKIPPED
> Task :sdks:java:testing:nexmark:analyzeDependencies SKIPPED
> Task :sdks:java:extensions:euphoria:check
> Task :sdks:java:extensions:euphoria:build
> Task :sdks:java:extensions:euphoria:buildDependents
> Task :sdks:java:extensions:kryo:buildDependents
> Task :sdks:java:testing:nexmark:checkstyleMain
> Task :sdks:java:testing:nexmark:checkstyleTest
> Task :sdks:java:extensions:sql:zetasql:javadoc
> Task :sdks:java:io:amqp:check
> Task :sdks:java:io:amqp:build
> Task :sdks:java:io:amqp:buildDependents
[main] INFO org.gradle.internal.nativeintegration.services.NativeServices - 
Initialized native services in: /home/jenkins/.gradle/native

> Task :sdks:java:extensions:sql:zetasql:spotbugsMain
SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder".
SLF4J: Defaulting to no-operation (NOP) logger implementation
SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further 
details.

> Task :sdks:java:testing:nexmark:javadoc
> Task :sdks:java:io:hbase:test

[main] INFO org.gradle.internal.nativeintegration.services.NativeServices - 
Initialized native services in: /home/jenkins/.gradle/native

> Task :sdks:java:testing:nexmark:spotbugsMain
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in 
[jar:file:/home/jenkins/.gradle/caches/modules-2/files-2.1/org.slf4j/slf4j-simple/1.7.25/8dacf9514f0c707cbbcdd6fd699e8940d42fb54e/slf4j-simple-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in 
[jar:file:/home/jenkins/.gradle/wrapper/dists/gradle-5.2.1-all/bviwmvmbexq6idcscbicws5me/gradle-5.2.1/lib/gradle-logging-5.2.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.SimpleLoggerFactory]

> Task :sdks:java:io:hcatalog:test
> Task :sdks:java:io:jdbc:test
> Task :sdks:java:io:jms:test
> Task :sdks:java:io:kafka:test
> Task 

Build failed in Jenkins: beam_PostCommit_Java11_ValidatesRunner_Direct #2379

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[lostluck] [Go SDK] Correctly return EOFs from boolDecoder


--
Started by GitHub push by lostluck
Started by GitHub push by lostluck
Started by GitHub push by lostluck
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-11 (beam) in workspace 

No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init 
 > 
 >  # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git fetch --tags --progress https://github.com/apache/beam.git 
 > +refs/heads/*:refs/remotes/origin/*
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # 
 > timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git 
 > +refs/heads/*:refs/remotes/origin/* 
 > +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/*
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 22bd9b16a7f66f91f772fb1e619327c9484cef8f (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 22bd9b16a7f66f91f772fb1e619327c9484cef8f
Commit message: "[Go SDK] Correctly return EOFs from boolDecoder"
 > git rev-list --no-walk 1bff5ae7fab341a06af02ed0c62395f1cf0195ca # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the 
SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ 

 -Dorg.gradle.java.home=/usr/lib/jvm/java-8-openjdk-amd64 
:runners:direct-java:shadowJar :runners:direct-java:shadowTestJar
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for 
details
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy
> Task :buildSrc:spotlessGroovyCheck
> Task :buildSrc:spotlessGroovyGradle
> Task :buildSrc:spotlessGroovyGradleCheck
> Task :buildSrc:spotlessCheck
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validateTaskProperties FROM-CACHE
> Task :buildSrc:check
> Task :buildSrc:build
Configuration on demand is an incubating feature.
> Task :runners:core-construction-java:processResources NO-SOURCE
> Task :runners:local-java:processResources NO-SOURCE
> Task :sdks:java:core:generateAvroProtocol NO-SOURCE
> Task :runners:direct-java:processResources NO-SOURCE
> Task :sdks:java:harness:processResources NO-SOURCE
> Task :sdks:java:extensions:google-cloud-platform-core:processResources 
> NO-SOURCE
> Task :vendor:sdks-java-extensions-protobuf:processResources NO-SOURCE
> Task :runners:core-java:processResources NO-SOURCE
> Task :runners:java-fn-execution:processResources NO-SOURCE
> Task :sdks:java:fn-execution:processResources NO-SOURCE
> Task :runners:core-java:processTestResources NO-SOURCE
> Task :runners:direct-java:processTestResources NO-SOURCE
> Task :sdks:java:core:generateAvroJava NO-SOURCE
> Task :model:fn-execution:extractProto
> Task :model:job-management:extractProto
> Task :model:job-management:processResources
> Task :model:fn-execution:processResources
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :sdks:java:core:processResources
> Task :sdks:java:core:generateTestAvroProtocol NO-SOURCE
> Task :sdks:java:core:generateTestAvroJava
> Task :sdks:java:core:generateTestGrammarSource NO-SOURCE
> Task :sdks:java:core:processTestResources
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :model:pipeline:generateProto
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:extractIncludeProto
> Task :model:job-management:generateProto
> Task :model:fn-execution:generateProto
> Task 

beam_PostCommit_Python37 - Build # 872 - Aborted

2019-11-05 Thread Apache Jenkins Server
The Apache Jenkins build system has built beam_PostCommit_Python37 (build #872)

Status: Aborted

Check console output at 
https://builds.apache.org/job/beam_PostCommit_Python37/872/ to view the results.

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org

beam_PostCommit_Java_PVR_Spark_Batch - Build # 1214 - Aborted

2019-11-05 Thread Apache Jenkins Server
The Apache Jenkins build system has built beam_PostCommit_Java_PVR_Spark_Batch 
(build #1214)

Status: Aborted

Check console output at 
https://builds.apache.org/job/beam_PostCommit_Java_PVR_Spark_Batch/1214/ to 
view the results.

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org

Build failed in Jenkins: beam_PostCommit_Python37 #873

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[lostluck] [Go SDK] Correctly return EOFs from boolDecoder


--
[...truncated 303.94 KB...]
  }
},
{
  "kind": "GroupByKey",
  "name": "s4",
  "properties": {
"display_data": [],
"output_info": [
  {
"encoding": {
  "@type": "kind:windowed_value",
  "component_encodings": [
{
  "@type": "kind:pair",
  "component_encodings": [
{
  "@type": 
"StrUtf8Coder$eNprYE5OLEhMzkiNT0pNzNVLzk9JLSqGUlzBJUWhJWkWziAeVyGDZmMhY20hU5IeAAbXEkc=",
  "component_encodings": []
},
{
  "@type": "kind:stream",
  "component_encodings": [
{
  "@type": "kind:varint"
}
  ],
  "is_stream_like": true
}
  ],
  "is_pair_like": true
},
{
  "@type": "kind:global_window"
}
  ],
  "is_wrapper": true
},
"output_name": "out",
"user_name": "GroupByKey.out"
  }
],
"parallel_input": {
  "@type": "OutputReference",
  "output_name": "out",
  "step_name": "s3"
},
"serialized_fn": 
"%0AB%22%40%0A%1Dref_Coder_GlobalWindowCoder_1%12%1F%0A%1D%0A%1Bbeam%3Acoder%3Aglobal_window%3Av1jT%0A%25%0A%23%0A%21beam%3Awindowfn%3Aglobal_windows%3Av0.1%10%01%1A%1Dref_Coder_GlobalWindowCoder_1%22%02%3A%00%28%010%018%01H%01",
"user_name": "GroupByKey"
  }
},
{
  "kind": "ParallelDo",
  "name": "s5",
  "properties": {
"display_data": [
  {
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.ParDo",
"shortValue": "CallableWrapperDoFn",
"type": "STRING",
"value": "apache_beam.transforms.core.CallableWrapperDoFn"
  },
  {
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.CallableWrapperDoFn",
"type": "STRING",
"value": ""
  }
],
"non_parallel_inputs": {},
"output_info": [
  {
"encoding": {
  "@type": "kind:windowed_value",
  "component_encodings": [
{
  "@type": 
"FastPrimitivesCoder$eNprYE5OLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqYIfgYGhvi0xJycpMTk7HiwlkJ8pgVkJmfnpEJNYQGawlpbyJZUnKQHACYlLgM=",
  "component_encodings": [
{
  "@type": 
"FastPrimitivesCoder$eNprYE5OLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqYIfgYGhvi0xJycpMTk7HiwlkJ8pgVkJmfnpEJNYQGawlpbyJZUnKQHACYlLgM=",
  "component_encodings": []
},
{
  "@type": 
"FastPrimitivesCoder$eNprYE5OLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqYIfgYGhvi0xJycpMTk7HiwlkJ8pgVkJmfnpEJNYQGawlpbyJZUnKQHACYlLgM=",
  "component_encodings": []
}
  ],
  "is_pair_like": true
},
{
  "@type": "kind:global_window"
}
  ],
  "is_wrapper": true
},
"output_name": "out",
"user_name": "m_out.out"
  }
],
"parallel_input": {
  "@type": "OutputReference",
  "output_name": "out",
  "step_name": "s4"
},
"serialized_fn": 
"eNqFU1lv1DAQzu72TMt93zdkgSbQct+wFYdWWqptJfKCLCf2btw6ccZ2WCpRCR5aeOUfwB/gNzJJF0RBgKIk4/k839gz37xveDHNaZxwEnGa+lbTzPSUTo0fK83dFpWSRpK/0jTPuZ5XTzMXnOYHqK1B3QsbjuOQXgaNmAkpfVJ+XRJrTi0nvSKLrVAYMOJtwaWijNjVnLswGk4gRUsxvoRrGFuH8S5MeO1a28F3qr3YmmU1VmcNNsJGrbNc+1JnY++cr/WhPf673Z9YdGDS62B0vd1oj7RHwxHMoLKYgxuOomkHAu0pC9MVUl4btoVjaOZvqCw4bA+ncbFE+33OXhY2Lyzs+Ag7wxp638KuddgdfkYzSFTKg2WerYjM/PjPGEnf8GCg9IrBsvKgpCcLytiWSlNhycKqTVQ2dzMwOg4MWzFBXnmCX7oQ6CLLuDYBo5b2pBr8NAh/y3UsDCcpt1rEhuQi51Jk3M9XYU9VzHuSphGjD2Bv+1trwqm5tWl8YF+zaWF/Fw5saUWfW0Kt1S4crIKjQkiL94BD4TguES5ROLwBR7pwdEuoSHOlLUkVKyR28lj4EAN+1dLwFv6Pw/v/vwUcX4cTXThZnYVgotgSAqc24HQXziR7O501OBtOlVgpH5KIzBo4t1XBCFR+n3GUMLVKG/fFy1Jez0u3C+dRvheQyfMqKpFhhys+A81O1XtVdX3ou9gpNuBSZCxc7sJMhaO0cSx03+Q8Br9TlSrXKubGQJDM/HHCK0mZ8yrmnPWSZiepKOeiInoN19bg+mu48c8hfCUypgYi67twE3lurcFtr6rQoAJw/u78LX5zh/tMqojKTR6swF1kuRdOlsOgBcpcI8X9v1EMt7jzvEcLaZeGS3iAJA/DXWXT47hIC0nLYS/1wOFRuxbuLulFyo2laU5ilUbYYA2PEarqIwxhm5Tw5FMRWWj53wEJUH/z",
"user_name": 

Build failed in Jenkins: beam_PostCommit_SQL #3128

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[github] Removing some trailing whitespace.


--
[...truncated 4.31 KB...]
> Task :buildSrc:spotlessGroovyGradleCheck
> Task :buildSrc:spotlessCheck
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validateTaskProperties FROM-CACHE
> Task :buildSrc:check
> Task :buildSrc:build
Configuration on demand is an incubating feature.
> Task :sdks:java:core:generateAvroProtocol NO-SOURCE
> Task :sdks:java:harness:processResources NO-SOURCE
> Task :runners:core-java:processResources NO-SOURCE
> Task :runners:local-java:processResources NO-SOURCE
> Task :sdks:java:fn-execution:processResources NO-SOURCE
> Task :runners:core-construction-java:processResources NO-SOURCE
> Task :vendor:sdks-java-extensions-protobuf:processResources NO-SOURCE
> Task :runners:java-fn-execution:processResources NO-SOURCE
> Task :sdks:java:extensions:google-cloud-platform-core:processResources 
> NO-SOURCE
> Task :sdks:java:core:generateAvroJava NO-SOURCE
> Task :runners:direct-java:processResources NO-SOURCE
> Task :sdks:java:extensions:join-library:processResources NO-SOURCE
> Task :model:job-management:extractProto
> Task :model:fn-execution:extractProto
> Task :sdks:java:io:parquet:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :sdks:java:extensions:sql:datacatalog:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processTestResources NO-SOURCE
> Task :sdks:java:extensions:sql:zetasql:processResources NO-SOURCE
> Task :sdks:java:extensions:sql:datacatalog:processTestResources NO-SOURCE
> Task :sdks:java:io:common:compileJava NO-SOURCE
> Task :sdks:java:extensions:google-cloud-platform-core:processTestResources 
> NO-SOURCE
> Task :sdks:java:io:common:processResources NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :sdks:java:io:common:classes UP-TO-DATE
> Task :sdks:java:io:mongodb:processResources NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processTestResources NO-SOURCE
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :sdks:java:extensions:protobuf:extractProto
> Task :sdks:java:io:mongodb:processTestResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :model:job-management:processResources
> Task :model:fn-execution:processResources
> Task :sdks:java:io:common:jar
> Task :sdks:java:io:common:processTestResources NO-SOURCE
> Task :sdks:java:core:processResources
> Task :sdks:java:core:generateTestAvroProtocol NO-SOURCE
> Task :sdks:java:extensions:sql:jdbc:processResources
> Task :sdks:java:extensions:sql:jdbc:processTestResources NO-SOURCE
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :sdks:java:core:generateTestAvroJava
> Task :sdks:java:core:generateTestGrammarSource NO-SOURCE
> Task :sdks:java:core:processTestResources
> Task :model:pipeline:generateProto
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromCalciteCore
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromSrc
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:generateProto
> Task :model:job-management:generateProto
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes
> Task :model:pipeline:shadowJar
> Task :sdks:java:extensions:sql:generateFmppSources
> Task :model:job-management:shadowJar
> Task :model:fn-execution:shadowJar
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes

> Task :sdks:java:extensions:sql:compileJavacc
Java Compiler Compiler Version 4.0 (Parser Generator)
(type "javacc" with no arguments for help)
Reading from file 

 . . .
Note: UNICODE_INPUT option is specified. Please make sure you create the 
parser/lexer using a Reader with the correct character encoding.
Warning: Lookahead adequacy checking not being performed since option LOOKAHEAD 
is more than 1.  Set option FORCE_LA_CHECK to true to force checking.
File "TokenMgrError.java" does not exist.  Will create one.
File "ParseException.java" does not exist.  Will create one.
File "Token.java" does not exist.  Will create one.
File "SimpleCharStream.java" does not exist.  Will create one.
Parser generated 

Build failed in Jenkins: beam_PostCommit_Java_Nexmark_Flink #4222

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[github] Removing some trailing whitespace.


--
[...truncated 9.43 KB...]
> Task :runners:java-fn-execution:processResources NO-SOURCE
> Task :runners:flink:1.9:copyResourcesOverrides NO-SOURCE
> Task :runners:direct-java:processResources NO-SOURCE
> Task :sdks:java:core:generateAvroJava NO-SOURCE
> Task :model:fn-execution:extractProto
> Task :model:job-management:extractProto
> Task :sdks:java:extensions:join-library:processResources NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:extractProto
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :sdks:java:io:mongodb:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :sdks:java:io:parquet:processResources NO-SOURCE
> Task :sdks:java:testing:nexmark:processResources
> Task :model:job-management:processResources
> Task :model:fn-execution:processResources
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :runners:flink:1.9:copySourceOverrides
> Task :runners:flink:1.9:copyTestResourcesOverrides NO-SOURCE
> Task :runners:flink:1.9:processResources
> Task :sdks:java:build-tools:compileJava FROM-CACHE
> Task :sdks:java:build-tools:processResources
> Task :sdks:java:build-tools:classes
> Task :sdks:java:core:processResources
> Task :sdks:java:build-tools:jar
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromCalciteCore
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromSrc
> Task :model:pipeline:generateProto
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :sdks:java:extensions:sql:generateFmppSources
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:generateProto
> Task :model:fn-execution:generateProto
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes

> Task :sdks:java:extensions:sql:compileJavacc
Java Compiler Compiler Version 4.0 (Parser Generator)
(type "javacc" with no arguments for help)
Reading from file 

 . . .
Warning: Lookahead adequacy checking not being performed since option LOOKAHEAD 
is more than 1.  Set option FORCE_LA_CHECK to true to force checking.
Note: UNICODE_INPUT option is specified. Please make sure you create the 
parser/lexer using a Reader with the correct character encoding.
File "TokenMgrError.java" does not exist.  Will create one.
File "ParseException.java" does not exist.  Will create one.
File "Token.java" does not exist.  Will create one.
File "SimpleCharStream.java" does not exist.  Will create one.
Parser generated with 0 errors and 1 warnings.

> Task :sdks:java:extensions:sql:processResources
> Task :model:pipeline:shadowJar
> Task :model:job-management:shadowJar
> Task :model:fn-execution:shadowJar
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :sdks:java:extensions:join-library:compileJava FROM-CACHE
> Task :sdks:java:extensions:join-library:classes UP-TO-DATE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :sdks:java:extensions:join-library:jar
> Task :sdks:java:io:mongodb:compileJava FROM-CACHE
> Task :runners:local-java:jar
> Task :sdks:java:io:mongodb:classes UP-TO-DATE
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :vendor:sdks-java-extensions-protobuf:compileJava FROM-CACHE
> Task :vendor:sdks-java-extensions-protobuf:classes UP-TO-DATE
> Task :sdks:java:io:mongodb:jar
> Task :sdks:java:io:kafka:jar
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :vendor:sdks-java-extensions-protobuf:shadowJar
> Task :sdks:java:core:jar
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :runners:core-construction-java:jar
> Task :sdks:java:fn-execution:jar
> Task :sdks:java:io:parquet:compileJava FROM-CACHE
> Task :sdks:java:io:parquet:classes UP-TO-DATE
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :sdks:java:io:parquet:jar
> Task 

Build failed in Jenkins: beam_PostCommit_Java11_ValidatesRunner_Direct #2378

2019-11-05 Thread Apache Jenkins Server
See 


Changes:


--
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-13 (beam) in workspace 

No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init 
 > 
 >  # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git fetch --tags --progress https://github.com/apache/beam.git 
 > +refs/heads/*:refs/remotes/origin/*
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # 
 > timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git 
 > +refs/heads/*:refs/remotes/origin/* 
 > +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/*
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 1bff5ae7fab341a06af02ed0c62395f1cf0195ca (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 1bff5ae7fab341a06af02ed0c62395f1cf0195ca
Commit message: "Merge pull request #10001 from youngoli/patch-9"
 > git rev-list --no-walk 1bff5ae7fab341a06af02ed0c62395f1cf0195ca # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the 
SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ 

 -Dorg.gradle.java.home=/usr/lib/jvm/java-8-openjdk-amd64 
:runners:direct-java:shadowJar :runners:direct-java:shadowTestJar
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for 
details
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy
> Task :buildSrc:spotlessGroovyCheck
> Task :buildSrc:spotlessGroovyGradle
> Task :buildSrc:spotlessGroovyGradleCheck
> Task :buildSrc:spotlessCheck
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validateTaskProperties FROM-CACHE
> Task :buildSrc:check
> Task :buildSrc:build
Configuration on demand is an incubating feature.
> Task :sdks:java:core:generateAvroProtocol NO-SOURCE
> Task :sdks:java:harness:processResources NO-SOURCE
> Task :runners:java-fn-execution:processResources NO-SOURCE
> Task :sdks:java:fn-execution:processResources NO-SOURCE
> Task :sdks:java:extensions:google-cloud-platform-core:processResources 
> NO-SOURCE
> Task :runners:core-construction-java:processResources NO-SOURCE
> Task :runners:core-java:processResources NO-SOURCE
> Task :vendor:sdks-java-extensions-protobuf:processResources NO-SOURCE
> Task :runners:local-java:processResources NO-SOURCE
> Task :runners:direct-java:processResources NO-SOURCE
> Task :sdks:java:core:generateAvroJava NO-SOURCE
> Task :runners:direct-java:processTestResources NO-SOURCE
> Task :runners:core-java:processTestResources NO-SOURCE
> Task :model:job-management:extractProto
> Task :model:fn-execution:extractProto
> Task :model:job-management:processResources
> Task :model:fn-execution:processResources
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :sdks:java:core:processResources
> Task :sdks:java:core:generateTestAvroProtocol NO-SOURCE
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :sdks:java:core:generateTestAvroJava
> Task :sdks:java:core:generateTestGrammarSource NO-SOURCE
> Task :sdks:java:core:processTestResources
> Task :model:pipeline:generateProto
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:extractIncludeProto
> Task :model:job-management:generateProto
> Task :model:fn-execution:generateProto
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task 

Jenkins build is back to normal : beam_PostCommit_Python35 #914

2019-11-05 Thread Apache Jenkins Server
See 



-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org



Build failed in Jenkins: beam_PostCommit_Py_VR_Dataflow #5007

2019-11-05 Thread Apache Jenkins Server
See 


Changes:


--
[...truncated 269.35 KB...]
  }
},
"workerHarnessContainerImage": 
"gcr.io/cloud-dataflow/v1beta3/python3-fnapi:beam-master-20191029"
  }
]
  },
  "name": "beamapp-jenkins-1106005844-139481",
  "steps": [
{
  "kind": "ParallelRead",
  "name": "s1",
  "properties": {
"display_data": [
  {
"key": "subscription",
"label": "Pubsub Subscription",
"namespace": "apache_beam.io.gcp.pubsub._PubSubSource",
"type": "STRING",
"value": 
"projects/apache-beam-testing/subscriptions/exercise_streaming_metrics_subscription_input2e2072ef-e4a0-431b-a596-2ea9003b2508"
  },
  {
"key": "with_attributes",
"label": "With Attributes",
"namespace": "apache_beam.io.gcp.pubsub._PubSubSource",
"type": "BOOLEAN",
"value": false
  },
  {
"key": "source",
"label": "Read Source",
"namespace": "apache_beam.io.iobase.Read",
"shortValue": "_PubSubSource",
"type": "STRING",
"value": "apache_beam.io.gcp.pubsub._PubSubSource"
  }
],
"format": "pubsub",
"output_info": [
  {
"encoding": {
  "@type": "kind:windowed_value",
  "component_encodings": [
{
  "@type": "kind:bytes"
},
{
  "@type": "kind:global_window"
}
  ],
  "is_wrapper": true
},
"output_name": "out",
"user_name": "ReadFromPubSub/Read.out"
  }
],
"pubsub_subscription": 
"projects/apache-beam-testing/subscriptions/exercise_streaming_metrics_subscription_input2e2072ef-e4a0-431b-a596-2ea9003b2508",
"user_name": "ReadFromPubSub/Read"
  }
},
{
  "kind": "ParallelDo",
  "name": "s2",
  "properties": {
"display_data": [
  {
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.ParDo",
"shortValue": "StreamingUserMetricsDoFn",
"type": "STRING",
"value": 
"apache_beam.runners.dataflow.dataflow_exercise_streaming_metrics_pipeline.StreamingUserMetricsDoFn"
  }
],
"non_parallel_inputs": {},
"output_info": [
  {
"encoding": {
  "@type": "kind:windowed_value",
  "component_encodings": [
{
  "@type": "kind:bytes"
},
{
  "@type": "kind:global_window"
}
  ],
  "is_wrapper": true
},
"output_name": "out",
"user_name": "generate_metrics.out"
  }
],
"parallel_input": {
  "@type": "OutputReference",
  "output_name": "out",
  "step_name": "s1"
},
"serialized_fn": "ref_AppliedPTransform_generate_metrics_4",
"user_name": "generate_metrics"
  }
},
{
  "kind": "ParallelWrite",
  "name": "s3",
  "properties": {
"display_data": [],
"encoding": {
  "@type": "kind:windowed_value",
  "component_encodings": [
{
  "@type": "kind:bytes"
},
{
  "@type": "kind:global_window"
}
  ],
  "is_wrapper": true
},
"format": "pubsub",
"parallel_input": {
  "@type": "OutputReference",
  "output_name": "out",
  "step_name": "s2"
},
"pubsub_topic": 
"projects/apache-beam-testing/topics/exercise_streaming_metrics_topic_output2e2072ef-e4a0-431b-a596-2ea9003b2508",
"user_name": "dump_to_pub/Write/NativeWrite"
  }
}
  ],
  "type": "JOB_TYPE_STREAMING"
}
root: INFO: Create job: 
root: INFO: Created job with id: [2019-11-05_16_59_14-6333562570795186832]
root: INFO: To access the Dataflow monitoring console, please navigate to 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_16_59_14-6333562570795186832?project=apache-beam-testing
root: INFO: Job 2019-11-05_16_59_14-6333562570795186832 is in state 
JOB_STATE_RUNNING
root: INFO: 2019-11-06T00:59:18.762Z: JOB_MESSAGE_DETAILED: Checking 
permissions granted to controller Service Account.
root: INFO: 2019-11-06T00:59:19.516Z: JOB_MESSAGE_BASIC: Worker configuration: 
n1-standard-4 in us-central1-f.
root: INFO: 2019-11-06T00:59:21.170Z: JOB_MESSAGE_DETAILED: Expanding 
SplittableParDo operations into optimizable parts.
root: INFO: 2019-11-06T00:59:21.172Z: JOB_MESSAGE_DETAILED: Expanding 
CollectionToSingleton 

beam_PostCommit_Java_PVR_Spark_Batch - Build # 1213 - Aborted

2019-11-05 Thread Apache Jenkins Server
The Apache Jenkins build system has built beam_PostCommit_Java_PVR_Spark_Batch 
(build #1213)

Status: Aborted

Check console output at 
https://builds.apache.org/job/beam_PostCommit_Java_PVR_Spark_Batch/1213/ to 
view the results.

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org

Build failed in Jenkins: beam_PostCommit_Python35 #915

2019-11-05 Thread Apache Jenkins Server
See 


Changes:


--
[...truncated 129.41 KB...]
},
"output_name": "out",
"user_name": "ReadFromPubSub/Read.out"
  }
],
"pubsub_subscription": 
"projects/apache-beam-testing/subscriptions/exercise_streaming_metrics_subscription_input5cba2496-2d0f-4491-8585-9df1f8736107",
"user_name": "ReadFromPubSub/Read"
  }
},
{
  "kind": "ParallelDo",
  "name": "s2",
  "properties": {
"display_data": [
  {
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.ParDo",
"shortValue": "StreamingUserMetricsDoFn",
"type": "STRING",
"value": 
"apache_beam.runners.dataflow.dataflow_exercise_streaming_metrics_pipeline.StreamingUserMetricsDoFn"
  }
],
"non_parallel_inputs": {},
"output_info": [
  {
"encoding": {
  "@type": "kind:windowed_value",
  "component_encodings": [
{
  "@type": "kind:bytes"
},
{
  "@type": "kind:global_window"
}
  ],
  "is_wrapper": true
},
"output_name": "out",
"user_name": "generate_metrics.out"
  }
],
"parallel_input": {
  "@type": "OutputReference",
  "output_name": "out",
  "step_name": "s1"
},
"serialized_fn": "ref_AppliedPTransform_generate_metrics_4",
"user_name": "generate_metrics"
  }
},
{
  "kind": "ParallelWrite",
  "name": "s3",
  "properties": {
"display_data": [],
"encoding": {
  "@type": "kind:windowed_value",
  "component_encodings": [
{
  "@type": "kind:bytes"
},
{
  "@type": "kind:global_window"
}
  ],
  "is_wrapper": true
},
"format": "pubsub",
"parallel_input": {
  "@type": "OutputReference",
  "output_name": "out",
  "step_name": "s2"
},
"pubsub_topic": 
"projects/apache-beam-testing/topics/exercise_streaming_metrics_topic_output5cba2496-2d0f-4491-8585-9df1f8736107",
"user_name": "dump_to_pub/Write/NativeWrite"
  }
}
  ],
  "type": "JOB_TYPE_STREAMING"
}
root: INFO: Create job: 
root: INFO: Created job with id: [2019-11-05_17_20_33-10087238313192822627]
root: INFO: To access the Dataflow monitoring console, please navigate to 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_17_20_33-10087238313192822627?project=apache-beam-testing
root: INFO: Job 2019-11-05_17_20_33-10087238313192822627 is in state 
JOB_STATE_RUNNING
root: INFO: 2019-11-06T01:20:38.332Z: JOB_MESSAGE_DETAILED: Checking 
permissions granted to controller Service Account.
root: INFO: 2019-11-06T01:20:39.314Z: JOB_MESSAGE_BASIC: Worker configuration: 
n1-standard-4 in us-central1-f.
root: INFO: 2019-11-06T01:20:40.022Z: JOB_MESSAGE_DETAILED: Expanding 
SplittableParDo operations into optimizable parts.
root: INFO: 2019-11-06T01:20:40.025Z: JOB_MESSAGE_DETAILED: Expanding 
CollectionToSingleton operations into optimizable parts.
root: INFO: 2019-11-06T01:20:40.031Z: JOB_MESSAGE_DETAILED: Expanding 
CoGroupByKey operations into optimizable parts.
root: INFO: 2019-11-06T01:20:40.039Z: JOB_MESSAGE_DETAILED: Expanding 
SplittableProcessKeyed operations into optimizable parts.
root: INFO: 2019-11-06T01:20:40.041Z: JOB_MESSAGE_DETAILED: Expanding 
GroupByKey operations into streaming Read/Write steps
root: INFO: 2019-11-06T01:20:40.044Z: JOB_MESSAGE_DEBUG: Annotating graph with 
Autotuner information.
root: INFO: 2019-11-06T01:20:40.057Z: JOB_MESSAGE_DETAILED: Fusing adjacent 
ParDo, Read, Write, and Flatten operations
root: INFO: 2019-11-06T01:20:40.060Z: JOB_MESSAGE_DETAILED: Fusing consumer 
generate_metrics into ReadFromPubSub/Read
root: INFO: 2019-11-06T01:20:40.062Z: JOB_MESSAGE_DETAILED: Fusing consumer 
dump_to_pub/Write/NativeWrite into generate_metrics
root: INFO: 2019-11-06T01:20:40.069Z: JOB_MESSAGE_DEBUG: Adding StepResource 
setup and teardown to workflow graph.
root: INFO: 2019-11-06T01:20:40.079Z: JOB_MESSAGE_DEBUG: Adding workflow start 
and stop steps.
root: INFO: 2019-11-06T01:20:40.130Z: JOB_MESSAGE_DEBUG: Assigning stage ids.
root: INFO: 2019-11-06T01:20:40.237Z: JOB_MESSAGE_DEBUG: Executing wait step 
start2
root: INFO: 2019-11-06T01:20:40.249Z: JOB_MESSAGE_DEBUG: Starting worker pool 
setup.
root: INFO: 2019-11-06T01:20:40.253Z: JOB_MESSAGE_BASIC: Starting 1 workers...
root: INFO: 2019-11-06T01:20:44.060Z: JOB_MESSAGE_BASIC: Executing operation 
ReadFromPubSub/Read+generate_metrics+dump_to_pub/Write/NativeWrite
root: INFO: 

Build failed in Jenkins: beam_PostCommit_Py_VR_Dataflow #5008

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[lostluck] [Go SDK] Correctly return EOFs from boolDecoder


--
[...truncated 307.15 KB...]
"parallel_input": {
  "@type": "OutputReference",
  "output_name": "out",
  "step_name": "s34"
},
"serialized_fn": "ref_AppliedPTransform_assert:even/Unkey_47",
"user_name": "assert:even/Unkey"
  }
},
{
  "kind": "ParallelDo",
  "name": "s36",
  "properties": {
"display_data": [
  {
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.ParDo",
"shortValue": "CallableWrapperDoFn",
"type": "STRING",
"value": "apache_beam.transforms.core.CallableWrapperDoFn"
  },
  {
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.CallableWrapperDoFn",
"type": "STRING",
"value": "_equal"
  }
],
"non_parallel_inputs": {},
"output_info": [
  {
"encoding": {
  "@type": "kind:windowed_value",
  "component_encodings": [
{
  "@type": 
"FastPrimitivesCoder$eNprYE5OLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqYIfgYGhvi0xJycpMTk7HiwlkJ8pgVkJmfnpEJNYQGawlpbyJZUnKQHACYlLgM=",
  "component_encodings": [
{
  "@type": 
"FastPrimitivesCoder$eNprYE5OLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqYIfgYGhvi0xJycpMTk7HiwlkJ8pgVkJmfnpEJNYQGawlpbyJZUnKQHACYlLgM=",
  "component_encodings": [],
  "pipeline_proto_coder_id": 
"ref_Coder_FastPrimitivesCoder_3"
},
{
  "@type": 
"FastPrimitivesCoder$eNprYE5OLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqYIfgYGhvi0xJycpMTk7HiwlkJ8pgVkJmfnpEJNYQGawlpbyJZUnKQHACYlLgM=",
  "component_encodings": [],
  "pipeline_proto_coder_id": 
"ref_Coder_FastPrimitivesCoder_3"
}
  ],
  "is_pair_like": true,
  "pipeline_proto_coder_id": "ref_Coder_FastPrimitivesCoder_3"
},
{
  "@type": "kind:global_window"
}
  ],
  "is_wrapper": true
},
"output_name": "out",
"user_name": "assert:even/Match.out"
  }
],
"parallel_input": {
  "@type": "OutputReference",
  "output_name": "out",
  "step_name": "s35"
},
"serialized_fn": "ref_AppliedPTransform_assert:even/Match_48",
"user_name": "assert:even/Match"
  }
}
  ],
  "type": "JOB_TYPE_STREAMING"
}
- >> end captured logging << -
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_17_50_25-4151374984955458477?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_17_57_34-1853134947885522616?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_17_50_22-3106095049146219546?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_17_58_10-3617371906697416729?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_17_50_23-815927072816230046?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_17_57_27-2344521185772272676?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_17_50_22-787999664565167636?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_17_58_16-14039457137518362215?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_17_50_23-3562321632129814751?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_17_57_26-14938224553431338860?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_17_50_23-462015695272351993?project=apache-beam-testing
Worker logs: 

beam_PostCommit_Java_PVR_Spark_Batch - Build # 1215 - Aborted

2019-11-05 Thread Apache Jenkins Server
The Apache Jenkins build system has built beam_PostCommit_Java_PVR_Spark_Batch 
(build #1215)

Status: Aborted

Check console output at 
https://builds.apache.org/job/beam_PostCommit_Java_PVR_Spark_Batch/1215/ to 
view the results.

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org

Jenkins build is back to normal : beam_PostCommit_Java11_ValidatesRunner_Direct #2380

2019-11-05 Thread Apache Jenkins Server
See 



-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org



Build failed in Jenkins: beam_PostCommit_XVR_Flink #855

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[github] Removing some trailing whitespace.


--
[...truncated 4.27 MB...]
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (9/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - MapPartition (MapPartition at 
[3]assert_that/{Group, Unkey, Match}) (9/16) (30c03e4328a33a7e6f0899c1e4fcf198) 
switched from RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (9/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - Freeing task resources for 
MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (9/16) 
(30c03e4328a33a7e6f0899c1e4fcf198).
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (9/16) (6f1f38f9bce7b3fde8b5862c656094a1) switched from 
CREATED to SCHEDULED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (9/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem 
streams are closed for task MapPartition (MapPartition at 
[3]assert_that/{Group, Unkey, Match}) (9/16) (30c03e4328a33a7e6f0899c1e4fcf198) 
[FINISHED]
[flink-akka.actor.default-dispatcher-5] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task MapPartition 
(MapPartition at [3]assert_that/{Group, Unkey, Match}) 
30c03e4328a33a7e6f0899c1e4fcf198.
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (9/16) (6f1f38f9bce7b3fde8b5862c656094a1) switched from 
SCHEDULED to DEPLOYING.
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - Deploying DataSink 
(DiscardingOutput) (9/16) (attempt #0) to 0ea596a3-3c7b-4d1e-b390-ca0308b1b331 
@ localhost (dataPort=-1)
[flink-akka.actor.default-dispatcher-5] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Received task DataSink 
(DiscardingOutput) (9/16).
[flink-akka.actor.default-dispatcher-4] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - MapPartition 
(MapPartition at [3]assert_that/{Group, Unkey, Match}) (9/16) 
(30c03e4328a33a7e6f0899c1e4fcf198) switched from RUNNING to FINISHED.
[DataSink (DiscardingOutput) (9/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (9/16) 
(6f1f38f9bce7b3fde8b5862c656094a1) switched from CREATED to DEPLOYING.
[DataSink (DiscardingOutput) (9/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Creating FileSystem stream leak 
safety net for task DataSink (DiscardingOutput) (9/16) 
(6f1f38f9bce7b3fde8b5862c656094a1) [DEPLOYING]
[DataSink (DiscardingOutput) (9/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Loading JAR files for task DataSink 
(DiscardingOutput) (9/16) (6f1f38f9bce7b3fde8b5862c656094a1) [DEPLOYING].
[DataSink (DiscardingOutput) (9/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Registering task at network: 
DataSink (DiscardingOutput) (9/16) (6f1f38f9bce7b3fde8b5862c656094a1) 
[DEPLOYING].
[DataSink (DiscardingOutput) (9/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (9/16) 
(6f1f38f9bce7b3fde8b5862c656094a1) switched from DEPLOYING to RUNNING.
[flink-akka.actor.default-dispatcher-8] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (9/16) (6f1f38f9bce7b3fde8b5862c656094a1) switched from 
DEPLOYING to RUNNING.
[DataSink (DiscardingOutput) (9/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (9/16) 
(6f1f38f9bce7b3fde8b5862c656094a1) switched from RUNNING to FINISHED.
[DataSink (DiscardingOutput) (9/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Freeing task resources for DataSink 
(DiscardingOutput) (9/16) (6f1f38f9bce7b3fde8b5862c656094a1).
[DataSink (DiscardingOutput) (9/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem streams are 
closed for task DataSink (DiscardingOutput) (9/16) 
(6f1f38f9bce7b3fde8b5862c656094a1) [FINISHED]
[flink-akka.actor.default-dispatcher-8] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task DataSink 
(DiscardingOutput) 6f1f38f9bce7b3fde8b5862c656094a1.
[flink-akka.actor.default-dispatcher-8] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (9/16) (6f1f38f9bce7b3fde8b5862c656094a1) switched from 
RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (13/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - MapPartition (MapPartition at 
[3]assert_that/{Group, Unkey, Match}) (13/16) 
(8cbbf7364163e17bc7ed23f2151e0e0a) switched from RUNNING to FINISHED.
[MapPartition (MapPartition at 

Build failed in Jenkins: beam_PostCommit_SQL #3129

2019-11-05 Thread Apache Jenkins Server
See 

Changes:


--
[...truncated 6.60 KB...]
> Task :buildSrc:build
Configuration on demand is an incubating feature.
> Task :vendor:sdks-java-extensions-protobuf:processResources NO-SOURCE
> Task :runners:java-fn-execution:processResources NO-SOURCE
> Task :sdks:java:fn-execution:processResources NO-SOURCE
> Task :sdks:java:extensions:google-cloud-platform-core:processResources 
> NO-SOURCE
> Task :runners:local-java:processResources NO-SOURCE
> Task :sdks:java:core:generateAvroProtocol NO-SOURCE
> Task :runners:core-construction-java:processResources NO-SOURCE
> Task :sdks:java:harness:processResources NO-SOURCE
> Task :runners:core-java:processResources NO-SOURCE
> Task :sdks:java:extensions:join-library:processResources NO-SOURCE
> Task :runners:direct-java:processResources NO-SOURCE
> Task :sdks:java:core:generateAvroJava NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :sdks:java:io:mongodb:processResources NO-SOURCE
> Task :sdks:java:io:parquet:processResources NO-SOURCE
> Task :sdks:java:io:common:compileJava NO-SOURCE
> Task :sdks:java:io:mongodb:processTestResources NO-SOURCE
> Task :sdks:java:extensions:sql:datacatalog:processResources NO-SOURCE
> Task :sdks:java:io:common:processResources NO-SOURCE
> Task :sdks:java:io:common:classes UP-TO-DATE
> Task :sdks:java:extensions:sql:zetasql:processResources NO-SOURCE
> Task :sdks:java:extensions:google-cloud-platform-core:processTestResources 
> NO-SOURCE
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processTestResources NO-SOURCE
> Task :model:fn-execution:extractProto
> Task :model:job-management:extractProto
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :sdks:java:extensions:sql:datacatalog:processTestResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processTestResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:extractProto
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :sdks:java:io:common:jar
> Task :model:job-management:processResources
> Task :sdks:java:io:common:processTestResources NO-SOURCE
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :model:fn-execution:processResources
> Task :sdks:java:core:processResources
> Task :sdks:java:core:generateTestAvroProtocol NO-SOURCE
> Task :sdks:java:extensions:sql:jdbc:processResources
> Task :sdks:java:extensions:sql:jdbc:processTestResources NO-SOURCE
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :sdks:java:core:generateTestAvroJava
> Task :sdks:java:core:generateTestGrammarSource NO-SOURCE
> Task :sdks:java:core:processTestResources
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromCalciteCore
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromSrc
> Task :model:pipeline:generateProto
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :sdks:java:extensions:sql:generateFmppSources
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:generateProto
> Task :model:fn-execution:generateProto
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes

> Task :sdks:java:extensions:sql:compileJavacc
Java Compiler Compiler Version 4.0 (Parser Generator)
(type "javacc" with no arguments for help)
Reading from file 

 . . .
Warning: Lookahead adequacy checking not being performed since option LOOKAHEAD 
is more than 1.  Set option FORCE_LA_CHECK to true to force checking.
Note: UNICODE_INPUT option is specified. Please make sure you create the 
parser/lexer using a Reader with the correct character encoding.
File "TokenMgrError.java" does not exist.  Will create one.
File "ParseException.java" does not exist.  Will create one.
File "Token.java" does not exist.  Will create one.
File "SimpleCharStream.java" does not exist.  Will create one.
Parser generated with 0 errors and 1 warnings.

> Task :sdks:java:extensions:sql:processResources
> Task :sdks:java:extensions:sql:processTestResources
> Task :model:pipeline:shadowJar
> Task :model:job-management:shadowJar
> Task :model:fn-execution:shadowJar
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :runners:local-java:jar
> Task :sdks:java:io:common:compileTestJava 

Build failed in Jenkins: beam_PreCommit_Python_pytest_Cron #8

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[kirillkozlov] Created a MongoDbTable and a provider for it

[chambers] Fix expected step name in Python PTransformTest test_read_metrics

[github] Updated assertion format string to use [] vs ?

[github] [BEAM-8539] Provide an initial definition of all job states and the

[chambers] generalized step name matching

[github] Removing some trailing whitespace.


--
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-15 (beam) in workspace 

No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init 
 >  # 
 > timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git fetch --tags --progress https://github.com/apache/beam.git 
 > +refs/heads/*:refs/remotes/origin/*
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # 
 > timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git 
 > +refs/heads/*:refs/remotes/origin/* 
 > +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/*
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 1bff5ae7fab341a06af02ed0c62395f1cf0195ca (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 1bff5ae7fab341a06af02ed0c62395f1cf0195ca
Commit message: "Merge pull request #10001 from youngoli/patch-9"
 > git rev-list --no-walk 16fbe89104fd9ef71179627f4cebff8b0ea9cd90 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the 
SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ 

 --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g 
-Dorg.gradle.jvmargs=-Xmx4g :pythonPreCommitPytest
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for 
details
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy
> Task :buildSrc:spotlessGroovyCheck
> Task :buildSrc:spotlessGroovyGradle
> Task :buildSrc:spotlessGroovyGradleCheck
> Task :buildSrc:spotlessCheck
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validateTaskProperties FROM-CACHE
> Task :buildSrc:check
> Task :buildSrc:build
Configuration on demand is an incubating feature.

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task 
':sdks:python:test-suites:tox:py35:preCommitPy35Pytest'.
> Task with path 'lint' not found in project 
> ':sdks:python:test-suites:tox:py35'.

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug 
option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with 
Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See 
https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 12s

Publishing build scan...
https://gradle.com/s/puilvttjubg7w

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org



Build failed in Jenkins: beam_PostCommit_Java_Nexmark_Flink #4224

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[lostluck] [Go SDK] Correctly return EOFs from boolDecoder


--
[...truncated 9.96 KB...]
> Task :runners:direct-java:processResources NO-SOURCE
> Task :runners:flink:1.9:copyResourcesOverrides NO-SOURCE
> Task :model:job-management:extractProto
> Task :model:fn-execution:extractProto
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:extractProto
> Task :sdks:java:io:parquet:processResources NO-SOURCE
> Task :sdks:java:extensions:join-library:processResources NO-SOURCE
> Task :sdks:java:io:mongodb:processResources NO-SOURCE
> Task :model:job-management:processResources
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :sdks:java:testing:nexmark:processResources
> Task :model:fn-execution:processResources
> Task :runners:flink:1.9:copySourceOverrides
> Task :runners:flink:1.9:copyTestResourcesOverrides NO-SOURCE
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :runners:flink:1.9:processResources
> Task :sdks:java:build-tools:compileJava FROM-CACHE
> Task :sdks:java:build-tools:processResources
> Task :sdks:java:build-tools:classes
> Task :sdks:java:core:processResources
> Task :sdks:java:build-tools:jar
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromCalciteCore
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromSrc
> Task :model:pipeline:generateProto
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :sdks:java:extensions:sql:generateFmppSources
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:extractIncludeProto
> Task :model:job-management:generateProto
> Task :model:fn-execution:generateProto
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes

> Task :sdks:java:extensions:sql:compileJavacc
Java Compiler Compiler Version 4.0 (Parser Generator)
(type "javacc" with no arguments for help)
Reading from file 

 . . .
Note: UNICODE_INPUT option is specified. Please make sure you create the 
parser/lexer using a Reader with the correct character encoding.
Warning: Lookahead adequacy checking not being performed since option LOOKAHEAD 
is more than 1.  Set option FORCE_LA_CHECK to true to force checking.
File "TokenMgrError.java" does not exist.  Will create one.
File "ParseException.java" does not exist.  Will create one.
File "Token.java" does not exist.  Will create one.
File "SimpleCharStream.java" does not exist.  Will create one.
Parser generated with 0 errors and 1 warnings.

> Task :sdks:java:extensions:sql:processResources
> Task :model:pipeline:shadowJar
> Task :model:job-management:shadowJar
> Task :model:fn-execution:shadowJar
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :runners:local-java:jar
> Task :sdks:java:extensions:join-library:compileJava FROM-CACHE
> Task :sdks:java:extensions:join-library:classes UP-TO-DATE
> Task :sdks:java:extensions:join-library:jar
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :vendor:sdks-java-extensions-protobuf:compileJava FROM-CACHE
> Task :vendor:sdks-java-extensions-protobuf:classes UP-TO-DATE
> Task :sdks:java:io:kafka:jar
> Task :sdks:java:core:jar
> Task :sdks:java:io:mongodb:compileJava FROM-CACHE
> Task :sdks:java:io:mongodb:classes UP-TO-DATE
> Task :vendor:sdks-java-extensions-protobuf:shadowJar
> Task :sdks:java:io:mongodb:jar
> Task :runners:core-construction-java:jar
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:fn-execution:jar
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:io:parquet:compileJava FROM-CACHE
> Task :sdks:java:io:parquet:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :sdks:java:io:parquet:jar
> Task 

beam_PostCommit_Python36 - Build # 917 - Aborted

2019-11-05 Thread Apache Jenkins Server
The Apache Jenkins build system has built beam_PostCommit_Python36 (build #917)

Status: Aborted

Check console output at 
https://builds.apache.org/job/beam_PostCommit_Python36/917/ to view the results.

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org

Build failed in Jenkins: beam_PostCommit_SQL #3130

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[lostluck] [Go SDK] Correctly return EOFs from boolDecoder


--
[...truncated 1.07 KB...]
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git 
 > +refs/heads/*:refs/remotes/origin/* 
 > +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/*
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 22bd9b16a7f66f91f772fb1e619327c9484cef8f (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 22bd9b16a7f66f91f772fb1e619327c9484cef8f
Commit message: "[Go SDK] Correctly return EOFs from boolDecoder"
 > git rev-list --no-walk 1bff5ae7fab341a06af02ed0c62395f1cf0195ca # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the 
SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $  
--continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g 
-Dorg.gradle.jvmargs=-Xmx4g :sqlPostCommit
Starting a Gradle Daemon, 1 busy and 1 stopped Daemons could not be reused, use 
--status for details
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy
> Task :buildSrc:spotlessGroovyCheck
> Task :buildSrc:spotlessGroovyGradle
> Task :buildSrc:spotlessGroovyGradleCheck
> Task :buildSrc:spotlessCheck
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validateTaskProperties FROM-CACHE
> Task :buildSrc:check
> Task :buildSrc:build
Configuration on demand is an incubating feature.
> Task :runners:core-construction-java:processResources NO-SOURCE
> Task :sdks:java:extensions:google-cloud-platform-core:processResources 
> NO-SOURCE
> Task :sdks:java:harness:processResources NO-SOURCE
> Task :sdks:java:core:generateAvroProtocol NO-SOURCE
> Task :sdks:java:fn-execution:processResources NO-SOURCE
> Task :runners:core-java:processResources NO-SOURCE
> Task :runners:local-java:processResources NO-SOURCE
> Task :vendor:sdks-java-extensions-protobuf:processResources NO-SOURCE
> Task :runners:java-fn-execution:processResources NO-SOURCE
> Task :runners:direct-java:processResources NO-SOURCE
> Task :sdks:java:core:generateAvroJava NO-SOURCE
> Task :model:fn-execution:extractProto
> Task :model:job-management:extractProto
> Task :sdks:java:io:mongodb:processResources NO-SOURCE
> Task :sdks:java:extensions:join-library:processResources NO-SOURCE
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :sdks:java:io:parquet:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:extractProto
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :sdks:java:io:common:compileJava NO-SOURCE
> Task :sdks:java:io:mongodb:processTestResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :sdks:java:extensions:sql:datacatalog:processResources NO-SOURCE
> Task :sdks:java:extensions:sql:zetasql:processResources NO-SOURCE
> Task :sdks:java:io:common:processResources NO-SOURCE
> Task :sdks:java:extensions:google-cloud-platform-core:processTestResources 
> NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processTestResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processTestResources NO-SOURCE
> Task :sdks:java:io:common:classes UP-TO-DATE
> Task :sdks:java:extensions:sql:datacatalog:processTestResources NO-SOURCE
> Task :model:fn-execution:processResources
> Task :model:job-management:processResources
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :sdks:java:io:common:jar
> Task :sdks:java:io:common:processTestResources NO-SOURCE
> Task :sdks:java:core:processResources
> Task :sdks:java:core:generateTestAvroProtocol NO-SOURCE
> Task :sdks:java:extensions:sql:jdbc:processResources
> Task :sdks:java:extensions:sql:jdbc:processTestResources NO-SOURCE
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :sdks:java:core:generateTestAvroJava
> Task :sdks:java:core:generateTestGrammarSource NO-SOURCE
> Task :sdks:java:core:processTestResources
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromCalciteCore
> Task 

beam_sonarqube_report - Build # 999 - Aborted

2019-11-05 Thread Apache Jenkins Server
The Apache Jenkins build system has built beam_sonarqube_report (build #999)

Status: Aborted

Check console output at 
https://builds.apache.org/job/beam_sonarqube_report/999/ to view the results.

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org

Build failed in Jenkins: beam_PostCommit_Py_VR_Dataflow #5009

2019-11-05 Thread Apache Jenkins Server
See 


Changes:


--
[...truncated 220.25 KB...]
"output_info": [
  {
"encoding": {
  "@type": "kind:windowed_value",
  "component_encodings": [
{
  "@type": "kind:bytes"
},
{
  "@type": "kind:global_window"
}
  ],
  "is_wrapper": true
},
"output_name": "out",
"user_name": "generate_metrics.out"
  }
],
"parallel_input": {
  "@type": "OutputReference",
  "output_name": "out",
  "step_name": "s1"
},
"serialized_fn": "ref_AppliedPTransform_generate_metrics_4",
"user_name": "generate_metrics"
  }
},
{
  "kind": "ParallelWrite",
  "name": "s3",
  "properties": {
"display_data": [],
"encoding": {
  "@type": "kind:windowed_value",
  "component_encodings": [
{
  "@type": "kind:bytes"
},
{
  "@type": "kind:global_window"
}
  ],
  "is_wrapper": true
},
"format": "pubsub",
"parallel_input": {
  "@type": "OutputReference",
  "output_name": "out",
  "step_name": "s2"
},
"pubsub_topic": 
"projects/apache-beam-testing/topics/exercise_streaming_metrics_topic_outputfb9e31c5-b1c9-4b7c-b734-c6cd5b256e89",
"user_name": "dump_to_pub/Write/NativeWrite"
  }
}
  ],
  "type": "JOB_TYPE_STREAMING"
}
root: INFO: Create job: 
root: INFO: Created job with id: [2019-11-05_22_26_56-18262518306838458696]
root: INFO: To access the Dataflow monitoring console, please navigate to 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_22_26_56-18262518306838458696?project=apache-beam-testing
root: INFO: Job 2019-11-05_22_26_56-18262518306838458696 is in state 
JOB_STATE_RUNNING
root: INFO: 2019-11-06T06:26:59.597Z: JOB_MESSAGE_DETAILED: Checking 
permissions granted to controller Service Account.
root: INFO: 2019-11-06T06:27:00.321Z: JOB_MESSAGE_BASIC: Worker configuration: 
n1-standard-4 in us-central1-f.
root: INFO: 2019-11-06T06:27:00.862Z: JOB_MESSAGE_DETAILED: Expanding 
SplittableParDo operations into optimizable parts.
root: INFO: 2019-11-06T06:27:00.868Z: JOB_MESSAGE_DETAILED: Expanding 
CollectionToSingleton operations into optimizable parts.
root: INFO: 2019-11-06T06:27:00.893Z: JOB_MESSAGE_DETAILED: Expanding 
CoGroupByKey operations into optimizable parts.
root: INFO: 2019-11-06T06:27:00.922Z: JOB_MESSAGE_DETAILED: Expanding 
SplittableProcessKeyed operations into optimizable parts.
root: INFO: 2019-11-06T06:27:00.926Z: JOB_MESSAGE_DETAILED: Expanding 
GroupByKey operations into streaming Read/Write steps
root: INFO: 2019-11-06T06:27:00.932Z: JOB_MESSAGE_DEBUG: Annotating graph with 
Autotuner information.
root: INFO: 2019-11-06T06:27:00.964Z: JOB_MESSAGE_DETAILED: Fusing adjacent 
ParDo, Read, Write, and Flatten operations
root: INFO: 2019-11-06T06:27:00.970Z: JOB_MESSAGE_DETAILED: Fusing consumer 
generate_metrics into ReadFromPubSub/Read
root: INFO: 2019-11-06T06:27:00.974Z: JOB_MESSAGE_DETAILED: Fusing consumer 
dump_to_pub/Write/NativeWrite into generate_metrics
root: INFO: 2019-11-06T06:27:00.989Z: JOB_MESSAGE_DEBUG: Adding StepResource 
setup and teardown to workflow graph.
root: INFO: 2019-11-06T06:27:01.026Z: JOB_MESSAGE_DEBUG: Adding workflow start 
and stop steps.
root: INFO: 2019-11-06T06:27:01.076Z: JOB_MESSAGE_DEBUG: Assigning stage ids.
root: INFO: 2019-11-06T06:27:01.229Z: JOB_MESSAGE_DEBUG: Executing wait step 
start2
root: INFO: 2019-11-06T06:27:01.273Z: JOB_MESSAGE_DEBUG: Starting worker pool 
setup.
root: INFO: 2019-11-06T06:27:01.288Z: JOB_MESSAGE_BASIC: Starting 1 workers...
root: INFO: 2019-11-06T06:27:03.743Z: JOB_MESSAGE_BASIC: Executing operation 
ReadFromPubSub/Read+generate_metrics+dump_to_pub/Write/NativeWrite
root: INFO: 2019-11-06T06:27:32.436Z: JOB_MESSAGE_DETAILED: Checking 
permissions granted to controller Service Account.
root: INFO: 2019-11-06T06:27:32.444Z: JOB_MESSAGE_DEBUG: Executing input step 
topology_init_attach_disk_input_step
root: INFO: 2019-11-06T06:27:33.260Z: JOB_MESSAGE_BASIC: Worker configuration: 
n1-standard-4 in us-central1-f.
root: INFO: 2019-11-06T06:27:39.317Z: JOB_MESSAGE_WARNING: Your project already 
contains 100 Dataflow-created metric descriptors and Stackdriver will not 
create new Dataflow custom metrics for this job. Each unique user-defined 
metric name (independent of the DoFn in which it is defined) produces a new 
metric descriptor. To delete old / unused metric descriptors see 
https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list
 and 

Build failed in Jenkins: beam_PreCommit_Python_pytest_Cron #9

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[lostluck] [Go SDK] Correctly return EOFs from boolDecoder


--
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-12 (beam) in workspace 

No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init 
 >  # 
 > timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git fetch --tags --progress https://github.com/apache/beam.git 
 > +refs/heads/*:refs/remotes/origin/*
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # 
 > timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git 
 > +refs/heads/*:refs/remotes/origin/* 
 > +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/*
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 22bd9b16a7f66f91f772fb1e619327c9484cef8f (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 22bd9b16a7f66f91f772fb1e619327c9484cef8f
Commit message: "[Go SDK] Correctly return EOFs from boolDecoder"
 > git rev-list --no-walk 1bff5ae7fab341a06af02ed0c62395f1cf0195ca # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the 
SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ 

 --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g 
-Dorg.gradle.jvmargs=-Xmx4g :pythonPreCommitPytest
Starting a Gradle Daemon, 1 busy and 1 stopped Daemons could not be reused, use 
--status for details
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy
> Task :buildSrc:spotlessGroovyCheck
> Task :buildSrc:spotlessGroovyGradle
> Task :buildSrc:spotlessGroovyGradleCheck
> Task :buildSrc:spotlessCheck
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validateTaskProperties FROM-CACHE
> Task :buildSrc:check
> Task :buildSrc:build
Configuration on demand is an incubating feature.

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task 
':sdks:python:test-suites:tox:py35:preCommitPy35Pytest'.
> Task with path 'lint' not found in project 
> ':sdks:python:test-suites:tox:py35'.

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug 
option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with 
Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See 
https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 10s

Publishing build scan...
https://gradle.com/s/tut6y3cbaxfzm

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org



Build failed in Jenkins: beam_PostCommit_SQL #3131

2019-11-05 Thread Apache Jenkins Server
See 

Changes:


--
[...truncated 1005 B...]
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git 
 > +refs/heads/*:refs/remotes/origin/* 
 > +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/*
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 22bd9b16a7f66f91f772fb1e619327c9484cef8f (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 22bd9b16a7f66f91f772fb1e619327c9484cef8f
Commit message: "[Go SDK] Correctly return EOFs from boolDecoder"
 > git rev-list --no-walk 22bd9b16a7f66f91f772fb1e619327c9484cef8f # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the 
SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $  
--continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g 
-Dorg.gradle.jvmargs=-Xmx4g :sqlPostCommit
Starting a Gradle Daemon, 1 busy and 1 stopped Daemons could not be reused, use 
--status for details
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy
> Task :buildSrc:spotlessGroovyCheck
> Task :buildSrc:spotlessGroovyGradle
> Task :buildSrc:spotlessGroovyGradleCheck
> Task :buildSrc:spotlessCheck
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validateTaskProperties FROM-CACHE
> Task :buildSrc:check
> Task :buildSrc:build
Configuration on demand is an incubating feature.
> Task :sdks:java:core:generateAvroProtocol NO-SOURCE
> Task :runners:core-java:processResources NO-SOURCE
> Task :runners:core-construction-java:processResources NO-SOURCE
> Task :sdks:java:fn-execution:processResources NO-SOURCE
> Task :vendor:sdks-java-extensions-protobuf:processResources NO-SOURCE
> Task :runners:java-fn-execution:processResources NO-SOURCE
> Task :sdks:java:harness:processResources NO-SOURCE
> Task :sdks:java:extensions:google-cloud-platform-core:processResources 
> NO-SOURCE
> Task :runners:local-java:processResources NO-SOURCE
> Task :sdks:java:core:generateAvroJava NO-SOURCE
> Task :runners:direct-java:processResources NO-SOURCE
> Task :model:job-management:extractProto
> Task :sdks:java:extensions:join-library:processResources NO-SOURCE
> Task :model:fn-execution:extractProto
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :sdks:java:io:mongodb:processResources NO-SOURCE
> Task :sdks:java:io:parquet:processResources NO-SOURCE
> Task :sdks:java:io:common:compileJava NO-SOURCE
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :sdks:java:io:common:processResources NO-SOURCE
> Task :sdks:java:io:common:classes UP-TO-DATE
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :sdks:java:extensions:sql:datacatalog:processResources NO-SOURCE
> Task :sdks:java:io:mongodb:processTestResources NO-SOURCE
> Task :sdks:java:extensions:sql:zetasql:processResources NO-SOURCE
> Task :sdks:java:extensions:google-cloud-platform-core:processTestResources 
> NO-SOURCE
> Task :sdks:java:extensions:protobuf:extractProto
> Task :sdks:java:io:google-cloud-platform:processTestResources NO-SOURCE
> Task :sdks:java:extensions:sql:datacatalog:processTestResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processTestResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :model:job-management:processResources
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :model:fn-execution:processResources
> Task :sdks:java:io:common:jar
> Task :sdks:java:io:common:processTestResources NO-SOURCE
> Task :sdks:java:core:processResources
> Task :sdks:java:core:generateTestAvroProtocol NO-SOURCE
> Task :sdks:java:extensions:sql:jdbc:processResources
> Task :sdks:java:extensions:sql:jdbc:processTestResources NO-SOURCE
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :sdks:java:core:generateTestAvroJava
> Task :sdks:java:core:generateTestGrammarSource NO-SOURCE
> Task :sdks:java:core:processTestResources
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromCalciteCore
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromSrc
> Task :model:pipeline:generateProto
> Task 

Build failed in Jenkins: beam_PostCommit_Java_Nexmark_Flink #4225

2019-11-05 Thread Apache Jenkins Server
See 


Changes:


--
[...truncated 9.41 KB...]
> Task :runners:core-construction-java:processResources NO-SOURCE
> Task :runners:direct-java:processResources NO-SOURCE
> Task :runners:flink:1.9:copyResourcesOverrides NO-SOURCE
> Task :model:fn-execution:extractProto
> Task :model:job-management:extractProto
> Task :sdks:java:core:generateAvroJava NO-SOURCE
> Task :sdks:java:extensions:join-library:processResources NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :sdks:java:io:parquet:processResources NO-SOURCE
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :sdks:java:io:mongodb:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:extractProto
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :sdks:java:testing:nexmark:processResources
> Task :model:job-management:processResources
> Task :model:fn-execution:processResources
> Task :runners:flink:1.9:copySourceOverrides
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :runners:flink:1.9:copyTestResourcesOverrides NO-SOURCE
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :runners:flink:1.9:processResources
> Task :sdks:java:build-tools:compileJava FROM-CACHE
> Task :sdks:java:build-tools:processResources
> Task :sdks:java:build-tools:classes
> Task :sdks:java:core:processResources
> Task :sdks:java:build-tools:jar
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromCalciteCore
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromSrc
> Task :model:pipeline:generateProto
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :sdks:java:extensions:sql:generateFmppSources
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:extractIncludeProto
> Task :model:job-management:generateProto
> Task :model:fn-execution:generateProto
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes

> Task :sdks:java:extensions:sql:compileJavacc
Java Compiler Compiler Version 4.0 (Parser Generator)
(type "javacc" with no arguments for help)
Reading from file 

 . . .
Note: UNICODE_INPUT option is specified. Please make sure you create the 
parser/lexer using a Reader with the correct character encoding.
Warning: Lookahead adequacy checking not being performed since option LOOKAHEAD 
is more than 1.  Set option FORCE_LA_CHECK to true to force checking.
File "TokenMgrError.java" does not exist.  Will create one.
File "ParseException.java" does not exist.  Will create one.
File "Token.java" does not exist.  Will create one.
File "SimpleCharStream.java" does not exist.  Will create one.
Parser generated with 0 errors and 1 warnings.

> Task :sdks:java:extensions:sql:processResources
> Task :model:pipeline:shadowJar
> Task :model:job-management:shadowJar
> Task :model:fn-execution:shadowJar
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :sdks:java:core:jar
> Task :sdks:java:extensions:join-library:compileJava FROM-CACHE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :sdks:java:extensions:join-library:classes UP-TO-DATE
> Task :runners:local-java:classes UP-TO-DATE
> Task :sdks:java:extensions:join-library:jar
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:local-java:jar
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :sdks:java:io:mongodb:compileJava FROM-CACHE
> Task :sdks:java:io:mongodb:classes UP-TO-DATE
> Task :vendor:sdks-java-extensions-protobuf:compileJava FROM-CACHE
> Task :vendor:sdks-java-extensions-protobuf:classes UP-TO-DATE
> Task :sdks:java:io:kafka:jar
> Task :sdks:java:io:mongodb:jar
> Task :vendor:sdks-java-extensions-protobuf:shadowJar
> Task :runners:core-construction-java:jar
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:fn-execution:jar
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :sdks:java:io:parquet:compileJava FROM-CACHE
> Task :sdks:java:io:parquet:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task 

beam_PostCommit_Java_PVR_Spark_Batch - Build # 1216 - Aborted

2019-11-05 Thread Apache Jenkins Server
The Apache Jenkins build system has built beam_PostCommit_Java_PVR_Spark_Batch 
(build #1216)

Status: Aborted

Check console output at 
https://builds.apache.org/job/beam_PostCommit_Java_PVR_Spark_Batch/1216/ to 
view the results.

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org

Build failed in Jenkins: beam_PostCommit_Python35_VR_Flink #337

2019-11-05 Thread Apache Jenkins Server
See 


Changes:


--
[...truncated 30.76 MB...]
at 
org.apache.beam.sdk.fn.channel.ManagedChannelFactory.forDescriptor(ManagedChannelFactory.java:44)
at 
org.apache.beam.runners.fnexecution.environment.ExternalEnvironmentFactory.createEnvironment(ExternalEnvironmentFactory.java:112)
at 
org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory$1.load(DefaultJobBundleFactory.java:181)
at 
org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory$1.load(DefaultJobBundleFactory.java:165)
at 
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3528)
at 
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2277)
at 
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2154)
at 
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2044)
at 
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.cache.LocalCache.get(LocalCache.java:3952)
at 
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.cache.LocalCache.getOrLoad(LocalCache.java:3974)
at 
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.cache.LocalCache$LocalLoadingCache.get(LocalCache.java:4958)
at 
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.cache.LocalCache$LocalLoadingCache.getUnchecked(LocalCache.java:4964)
at 
org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory$SimpleStageBundleFactory.(DefaultJobBundleFactory.java:214)
at 
org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory$SimpleStageBundleFactory.(DefaultJobBundleFactory.java:205)
at 
org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory.forStage(DefaultJobBundleFactory.java:188)
at 
org.apache.beam.runners.fnexecution.control.DefaultExecutableStageContext.getStageBundleFactory(DefaultExecutableStageContext.java:42)
at 
org.apache.beam.runners.fnexecution.control.ReferenceCountingExecutableStageContextFactory$WrappedContext.getStageBundleFactory(ReferenceCountingExecutableStageContextFactory.java:198)
at 
org.apache.beam.runners.flink.translation.wrappers.streaming.ExecutableStageDoFnOperator.open(ExecutableStageDoFnOperator.java:194)
at 
org.apache.flink.streaming.runtime.tasks.StreamTask.openAllOperators(StreamTask.java:532)
at 
org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:396)
at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:705)
at org.apache.flink.runtime.taskmanager.Task.run(Task.java:530)
at java.lang.Thread.run(Thread.java:748)

[GroupByKey -> [5]{Map(), assert_that} 
(2/2)] INFO org.apache.flink.runtime.taskmanager.Task - GroupByKey -> 
[5]{Map(), assert_that} (2/2) 
(258e36d9afe11ed885fb653d0517bd91) switched from DEPLOYING to RUNNING.
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Received task 
[4]assert_that/{Create, Group} (2/2).
[GroupByKey -> [5]{Map(), assert_that} 
(2/2)] INFO org.apache.flink.streaming.runtime.tasks.StreamTask - No state 
backend has been configured, using default (Memory / JobManager) 
MemoryStateBackend (data in heap memory / checkpoints to JobManager) 
(checkpoints: 'null', savepoints: 'null', asynchronous: TRUE, maxStateSize: 
5242880)
[flink-akka.actor.default-dispatcher-5] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - GroupByKey -> 
[5]{Map(), assert_that} (2/2) 
(258e36d9afe11ed885fb653d0517bd91) switched from DEPLOYING to RUNNING.
[[4]assert_that/{Create, Group} (1/2)] INFO 
org.apache.flink.runtime.taskmanager.Task - [4]assert_that/{Create, Group} 
(1/2) (a50c4feb441e4c7e42e20b63cb009487) switched from DEPLOYING to RUNNING.
[[4]assert_that/{Create, Group} (1/2)] INFO 
org.apache.flink.streaming.runtime.tasks.StreamTask - No state backend has been 
configured, using default (Memory / JobManager) MemoryStateBackend (data in 
heap memory / checkpoints to JobManager) (checkpoints: 'null', savepoints: 
'null', asynchronous: TRUE, maxStateSize: 5242880)
[flink-akka.actor.default-dispatcher-5] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - 
[4]assert_that/{Create, Group} (1/2) (a50c4feb441e4c7e42e20b63cb009487) 
switched from DEPLOYING to RUNNING.
[[4]assert_that/{Create, Group} (2/2)] INFO 
org.apache.flink.runtime.taskmanager.Task - [4]assert_that/{Create, Group} 
(2/2) (e8502c42e5860318d422f7829789f6fc) switched from CREATED to DEPLOYING.
[[4]assert_that/{Create, Group} (2/2)] INFO 
org.apache.flink.runtime.taskmanager.Task - Creating FileSystem stream leak 
safety net for task [4]assert_that/{Create, 

Build failed in Jenkins: beam_PostCommit_XVR_Flink #858

2019-11-05 Thread Apache Jenkins Server
See 


Changes:


--
[...truncated 4.26 MB...]
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task DataSink 
(DiscardingOutput) ae815cef418d6682db062bb708e6fae8.
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (2/16) (ae815cef418d6682db062bb708e6fae8) switched from 
RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (3/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - MapPartition (MapPartition at 
[3]assert_that/{Group, Unkey, Match}) (3/16) (53146ea64aa2adbd955dfd32f5fce6dc) 
switched from RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (3/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - Freeing task resources for 
MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (3/16) 
(53146ea64aa2adbd955dfd32f5fce6dc).
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (3/16) (65e50ebb5f2303cf68a66db7bf2e048d) switched from 
CREATED to SCHEDULED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (3/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem 
streams are closed for task MapPartition (MapPartition at 
[3]assert_that/{Group, Unkey, Match}) (3/16) (53146ea64aa2adbd955dfd32f5fce6dc) 
[FINISHED]
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task MapPartition 
(MapPartition at [3]assert_that/{Group, Unkey, Match}) 
53146ea64aa2adbd955dfd32f5fce6dc.
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (3/16) (65e50ebb5f2303cf68a66db7bf2e048d) switched from 
SCHEDULED to DEPLOYING.
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - Deploying DataSink 
(DiscardingOutput) (3/16) (attempt #0) to da33c38d-4a84-43ae-9da3-21447555bbe3 
@ localhost (dataPort=-1)
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Received task DataSink 
(DiscardingOutput) (3/16).
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - MapPartition 
(MapPartition at [3]assert_that/{Group, Unkey, Match}) (3/16) 
(53146ea64aa2adbd955dfd32f5fce6dc) switched from RUNNING to FINISHED.
[DataSink (DiscardingOutput) (3/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (3/16) 
(65e50ebb5f2303cf68a66db7bf2e048d) switched from CREATED to DEPLOYING.
[DataSink (DiscardingOutput) (3/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Creating FileSystem stream leak 
safety net for task DataSink (DiscardingOutput) (3/16) 
(65e50ebb5f2303cf68a66db7bf2e048d) [DEPLOYING]
[DataSink (DiscardingOutput) (3/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Loading JAR files for task DataSink 
(DiscardingOutput) (3/16) (65e50ebb5f2303cf68a66db7bf2e048d) [DEPLOYING].
[DataSink (DiscardingOutput) (3/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Registering task at network: 
DataSink (DiscardingOutput) (3/16) (65e50ebb5f2303cf68a66db7bf2e048d) 
[DEPLOYING].
[DataSink (DiscardingOutput) (3/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (3/16) 
(65e50ebb5f2303cf68a66db7bf2e048d) switched from DEPLOYING to RUNNING.
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (3/16) (65e50ebb5f2303cf68a66db7bf2e048d) switched from 
DEPLOYING to RUNNING.
[DataSink (DiscardingOutput) (3/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (3/16) 
(65e50ebb5f2303cf68a66db7bf2e048d) switched from RUNNING to FINISHED.
[DataSink (DiscardingOutput) (3/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Freeing task resources for DataSink 
(DiscardingOutput) (3/16) (65e50ebb5f2303cf68a66db7bf2e048d).
[DataSink (DiscardingOutput) (3/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem streams are 
closed for task DataSink (DiscardingOutput) (3/16) 
(65e50ebb5f2303cf68a66db7bf2e048d) [FINISHED]
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task DataSink 
(DiscardingOutput) 65e50ebb5f2303cf68a66db7bf2e048d.
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 

Jenkins build is back to normal : beam_PreCommit_Java_Cron #2002

2019-11-05 Thread Apache Jenkins Server
See 



-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org



beam_PostCommit_Java_PVR_Spark_Batch - Build # 1209 - Aborted

2019-11-05 Thread Apache Jenkins Server
The Apache Jenkins build system has built beam_PostCommit_Java_PVR_Spark_Batch 
(build #1209)

Status: Aborted

Check console output at 
https://builds.apache.org/job/beam_PostCommit_Java_PVR_Spark_Batch/1209/ to 
view the results.

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org

Build failed in Jenkins: beam_PostCommit_Py_VR_Dataflow #5001

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[mxm] Use better name for Flink current key supplier

[lgajowy] [BEAM-8558] Fix BigQueryIOIT flakiness (#9995)


--
[...truncated 595.82 KB...]
"encoding": {
  "@type": "kind:windowed_value",
  "component_encodings": [
{
  "@type": "kind:pair",
  "component_encodings": [
{
  "@type": "kind:bytes"
},
{
  "@type": 
"FastPrimitivesCoder$eNprYE5OLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqYIfgYGhvi0xJycpMTk7HiwlkJ8pgVkJmfnpEJNYQGawlpbyJZUnKQHACYlLgM=",
  "component_encodings": [
{
  "@type": 
"FastPrimitivesCoder$eNprYE5OLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqYIfgYGhvi0xJycpMTk7HiwlkJ8pgVkJmfnpEJNYQGawlpbyJZUnKQHACYlLgM=",
  "component_encodings": [],
  "pipeline_proto_coder_id": 
"ref_Coder_FastPrimitivesCoder_3"
},
{
  "@type": 
"FastPrimitivesCoder$eNprYE5OLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqYIfgYGhvi0xJycpMTk7HiwlkJ8pgVkJmfnpEJNYQGawlpbyJZUnKQHACYlLgM=",
  "component_encodings": [],
  "pipeline_proto_coder_id": 
"ref_Coder_FastPrimitivesCoder_3"
}
  ],
  "is_pair_like": true,
  "pipeline_proto_coder_id": 
"ref_Coder_FastPrimitivesCoder_3"
}
  ],
  "is_pair_like": true
},
{
  "@type": "kind:global_window"
}
  ],
  "is_wrapper": true
},
"output_name": "out",
"user_name": "Map()/MapToVoidKey1.out"
  }
],
"parallel_input": {
  "@type": "OutputReference",
  "output_name": "out",
  "step_name": "s4"
},
"serialized_fn": "ref_AppliedPTransform_Map()/MapToVoidKey1_24",
"user_name": "Map()/MapToVoidKey1"
  }
}
  ],
  "type": "JOB_TYPE_STREAMING"
}
root: INFO: Create job: 
root: INFO: Created job with id: [2019-11-05_08_20_59-17558144775234027111]
root: INFO: To access the Dataflow monitoring console, please navigate to 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_08_20_59-17558144775234027111?project=apache-beam-testing
root: WARNING: Waiting indefinitely for streaming job.
root: INFO: Job 2019-11-05_08_20_59-17558144775234027111 is in state 
JOB_STATE_RUNNING
root: INFO: 2019-11-05T16:21:03.247Z: JOB_MESSAGE_DETAILED: Checking 
permissions granted to controller Service Account.
root: INFO: 2019-11-05T16:21:04.012Z: JOB_MESSAGE_BASIC: Worker configuration: 
n1-standard-4 in us-central1-f.
root: INFO: 2019-11-05T16:21:04.669Z: JOB_MESSAGE_DETAILED: Expanding 
SplittableParDo operations into optimizable parts.
root: INFO: 2019-11-05T16:21:04.671Z: JOB_MESSAGE_DETAILED: Expanding 
CollectionToSingleton operations into optimizable parts.
root: INFO: 2019-11-05T16:21:04.684Z: JOB_MESSAGE_DETAILED: Expanding 
CoGroupByKey operations into optimizable parts.
root: INFO: 2019-11-05T16:21:04.694Z: JOB_MESSAGE_DEBUG: Combiner lifting 
skipped for step Map()/_UnpickledSideInput(MapToVoidKey1.out.0)/GroupByKey: 
GroupByKey not followed by a combiner.
root: INFO: 2019-11-05T16:21:04.696Z: JOB_MESSAGE_DEBUG: Combiner lifting 
skipped for step Map()/_UnpickledSideInput(MapToVoidKey0.out.0)/GroupByKey: 
GroupByKey not followed by a combiner.
root: INFO: 2019-11-05T16:21:04.706Z: JOB_MESSAGE_DETAILED: Expanding 
SplittableProcessKeyed operations into optimizable parts.
root: INFO: 2019-11-05T16:21:04.709Z: JOB_MESSAGE_DETAILED: Expanding 
GroupByKey operations into streaming Read/Write steps
root: INFO: 2019-11-05T16:21:04.729Z: JOB_MESSAGE_DEBUG: Annotating graph with 
Autotuner information.
root: INFO: 2019-11-05T16:21:04.765Z: JOB_MESSAGE_DETAILED: Fusing adjacent 
ParDo, Read, Write, and Flatten operations
root: INFO: 2019-11-05T16:21:04.772Z: JOB_MESSAGE_DETAILED: Fusing consumer 
Map()/MapToVoidKey0 into side list/Decode 
Values
root: INFO: 2019-11-05T16:21:04.785Z: JOB_MESSAGE_DETAILED: Fusing consumer 
Map()/MapToVoidKey1 into side list/Decode 
Values
root: INFO: 2019-11-05T16:21:04.789Z: JOB_MESSAGE_DETAILED: Fusing consumer 
Map()/MapToVoidKey0 into side list/Decode 
Values
root: INFO: 2019-11-05T16:21:04.794Z: JOB_MESSAGE_DETAILED: Fusing consumer 
Map()/MapToVoidKey1 into side list/Decode 
Values
root: INFO: 2019-11-05T16:21:04.797Z: JOB_MESSAGE_DETAILED: Unzipping flatten 
s16 for input s14.out
root: INFO: 

Build failed in Jenkins: beam_PostCommit_Python35 #909

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[mxm] Use better name for Flink current key supplier

[lgajowy] [BEAM-8558] Fix BigQueryIOIT flakiness (#9995)


--
[...truncated 223.53 KB...]
  "parallelWorkerSettings": {
"baseUrl": "https://dataflow.googleapis.com;,
"servicePath": "https://dataflow.googleapis.com;
  }
},
"workerHarnessContainerImage": 
"gcr.io/cloud-dataflow/v1beta3/python3:beam-master-20191029"
  }
]
  },
  "name": "beamapp-jenkins-1105162447-355862",
  "steps": [
{
  "kind": "ParallelRead",
  "name": "s1",
  "properties": {
"bigquery_export_format": "FORMAT_AVRO",
"bigquery_flatten_results": true,
"bigquery_kms_key": 
"projects/apache-beam-testing/locations/global/keyRings/beam-it/cryptoKeys/test",
"bigquery_query": "SELECT * FROM (SELECT \"apple\" as fruit) UNION ALL 
(SELECT \"orange\" as fruit)",
"bigquery_use_legacy_sql": false,
"display_data": [
  {
"key": "query",
"label": "Query",
"namespace": "apache_beam.io.gcp.bigquery.BigQuerySource",
"type": "STRING",
"value": "SELECT * FROM (SELECT \"apple\" as fruit) UNION ALL 
(SELECT \"orange\" as fruit)"
  },
  {
"key": "validation",
"label": "Validation Enabled",
"namespace": "apache_beam.io.gcp.bigquery.BigQuerySource",
"type": "BOOLEAN",
"value": false
  },
  {
"key": "source",
"label": "Read Source",
"namespace": "apache_beam.io.iobase.Read",
"shortValue": "BigQuerySource",
"type": "STRING",
"value": "apache_beam.io.gcp.bigquery.BigQuerySource"
  }
],
"format": "bigquery",
"output_info": [
  {
"encoding": {
  "@type": "kind:windowed_value",
  "component_encodings": [
{
  "@type": 
"FastPrimitivesCoder$eNprYE5OLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqYIfgYGhvi0xJycpMTk7HiwlkJ8pgVkJmfnpEJNYQGawlpbyJZUnKQHACYlLgM=",
  "component_encodings": [
{
  "@type": 
"FastPrimitivesCoder$eNprYE5OLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqYIfgYGhvi0xJycpMTk7HiwlkJ8pgVkJmfnpEJNYQGawlpbyJZUnKQHACYlLgM=",
  "component_encodings": []
},
{
  "@type": 
"FastPrimitivesCoder$eNprYE5OLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqYIfgYGhvi0xJycpMTk7HiwlkJ8pgVkJmfnpEJNYQGawlpbyJZUnKQHACYlLgM=",
  "component_encodings": []
}
  ],
  "is_pair_like": true
},
{
  "@type": "kind:global_window"
}
  ],
  "is_wrapper": true
},
"output_name": "out",
"user_name": "read.out"
  }
],
"user_name": "read"
  }
},
{
  "kind": "ParallelWrite",
  "name": "s2",
  "properties": {
"bigquery_kms_key": 
"projects/apache-beam-testing/locations/global/keyRings/beam-it/cryptoKeys/test",
"create_disposition": "CREATE_IF_NEEDED",
"dataset": "python_query_to_table_1572971086414",
"display_data": [],
"encoding": {
  "@type": "kind:windowed_value",
  "component_encodings": [
{
  "@type": 
"RowAsDictJsonCoder$eNprYE5OLEhMzkiNT0pNzNXLzNdLTy7QS8pMLyxNLaqML8nPzynmCsovdyx2yUwu8SrOz3POT0kt4ipk0GwsZKwtZErSAwBK5xfp",
  "component_encodings": []
},
{
  "@type": "kind:global_window"
}
  ],
  "is_wrapper": true
},
"format": "bigquery",
"parallel_input": {
  "@type": "OutputReference",
  "output_name": "out",
  "step_name": "s1"
},
"schema": "{\"fields\": [{\"type\": \"STRING\", \"mode\": \"NULLABLE\", 
\"name\": \"fruit\"}]}",
"table": "output_table",
"user_name": "write/NativeWrite",
"write_disposition": "WRITE_EMPTY"
  }
}
  ],
  "type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: 
root: INFO: Created job with id: [2019-11-05_08_25_19-13932524086749525428]
root: INFO: To access the Dataflow monitoring console, please navigate to 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_08_25_19-13932524086749525428?project=apache-beam-testing
- >> end captured logging << -
:1211:
 

Build failed in Jenkins: beam_PostCommit_Java_Nexmark_Flink #4218

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[robertwb] [BEAM-8435] Implement PaneInfo computation for Python.


--
[...truncated 8.45 KB...]
> Task :vendor:sdks-java-extensions-protobuf:processResources NO-SOURCE
> Task :runners:core-java:processResources NO-SOURCE
> Task :runners:java-fn-execution:processResources NO-SOURCE
> Task :sdks:java:harness:processResources NO-SOURCE
> Task :sdks:java:extensions:google-cloud-platform-core:processResources 
> NO-SOURCE
> Task :sdks:java:core:generateAvroJava NO-SOURCE
> Task :runners:flink:1.9:copyResourcesOverrides NO-SOURCE
> Task :runners:direct-java:processResources NO-SOURCE
> Task :model:job-management:extractProto
> Task :model:fn-execution:extractProto
> Task :sdks:java:extensions:join-library:processResources NO-SOURCE
> Task :model:job-management:processResources
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :sdks:java:io:parquet:processResources NO-SOURCE
> Task :model:fn-execution:processResources
> Task :sdks:java:extensions:protobuf:extractProto
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :runners:flink:1.9:copySourceOverrides
> Task :runners:flink:1.9:copyTestResourcesOverrides NO-SOURCE
> Task :sdks:java:testing:nexmark:processResources
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :runners:flink:1.9:processResources
> Task :sdks:java:build-tools:compileJava FROM-CACHE
> Task :sdks:java:build-tools:processResources
> Task :sdks:java:build-tools:classes
> Task :sdks:java:core:processResources
> Task :sdks:java:build-tools:jar
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromCalciteCore
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromSrc
> Task :model:pipeline:generateProto
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :sdks:java:extensions:sql:generateFmppSources
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:generateProto
> Task :model:fn-execution:generateProto
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes

> Task :sdks:java:extensions:sql:compileJavacc
Java Compiler Compiler Version 4.0 (Parser Generator)
(type "javacc" with no arguments for help)
Reading from file 

 . . .
Note: UNICODE_INPUT option is specified. Please make sure you create the 
parser/lexer using a Reader with the correct character encoding.
Warning: Lookahead adequacy checking not being performed since option LOOKAHEAD 
is more than 1.  Set option FORCE_LA_CHECK to true to force checking.
File "TokenMgrError.java" does not exist.  Will create one.
File "ParseException.java" does not exist.  Will create one.
File "Token.java" does not exist.  Will create one.
File "SimpleCharStream.java" does not exist.  Will create one.
Parser generated with 0 errors and 1 warnings.

> Task :sdks:java:extensions:sql:processResources
> Task :model:pipeline:shadowJar
> Task :model:job-management:shadowJar
> Task :model:fn-execution:shadowJar
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :sdks:java:core:jar
> Task :vendor:sdks-java-extensions-protobuf:compileJava FROM-CACHE
> Task :vendor:sdks-java-extensions-protobuf:classes UP-TO-DATE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :sdks:java:extensions:join-library:compileJava FROM-CACHE
> Task :sdks:java:extensions:join-library:classes UP-TO-DATE
> Task :runners:local-java:jar
> Task :sdks:java:extensions:join-library:jar
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :vendor:sdks-java-extensions-protobuf:shadowJar
> Task :sdks:java:io:kafka:jar
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :runners:core-construction-java:jar
> Task :sdks:java:fn-execution:jar
> Task :sdks:java:io:parquet:compileJava FROM-CACHE
> Task :sdks:java:io:parquet:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE

beam_PostCommit_Python37 - Build # 868 - Aborted

2019-11-05 Thread Apache Jenkins Server
The Apache Jenkins build system has built beam_PostCommit_Python37 (build #868)

Status: Aborted

Check console output at 
https://builds.apache.org/job/beam_PostCommit_Python37/868/ to view the results.

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org

Jenkins build is back to normal : beam_PostCommit_Py_VR_Dataflow #5002

2019-11-05 Thread Apache Jenkins Server
See 



-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org



Build failed in Jenkins: beam_PostCommit_Python36 #913

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[robertwb] [BEAM-8435] Implement PaneInfo computation for Python.


--
[...truncated 115.99 KB...]
"label": "Query",
"namespace": "apache_beam.io.gcp.bigquery.BigQuerySource",
"type": "STRING",
"value": "SELECT * FROM (SELECT \"apple\" as fruit) UNION ALL 
(SELECT \"orange\" as fruit)"
  },
  {
"key": "validation",
"label": "Validation Enabled",
"namespace": "apache_beam.io.gcp.bigquery.BigQuerySource",
"type": "BOOLEAN",
"value": false
  }
],
"format": "bigquery",
"output_info": [
  {
"encoding": {
  "@type": "kind:windowed_value",
  "component_encodings": [
{
  "@type": 
"FastPrimitivesCoder$eNprYE5OLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqYIfgYGhvi0xJycpMTk7HiwlkJ8pgVkJmfnpEJNYQGawlpbyJZUnKQHACYlLgM=",
  "component_encodings": [
{
  "@type": 
"FastPrimitivesCoder$eNprYE5OLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqYIfgYGhvi0xJycpMTk7HiwlkJ8pgVkJmfnpEJNYQGawlpbyJZUnKQHACYlLgM=",
  "component_encodings": []
},
{
  "@type": 
"FastPrimitivesCoder$eNprYE5OLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqYIfgYGhvi0xJycpMTk7HiwlkJ8pgVkJmfnpEJNYQGawlpbyJZUnKQHACYlLgM=",
  "component_encodings": []
}
  ],
  "is_pair_like": true
},
{
  "@type": "kind:global_window"
}
  ],
  "is_wrapper": true
},
"output_name": "out",
"user_name": "read.out"
  }
],
"user_name": "read"
  }
},
{
  "kind": "ParallelWrite",
  "name": "s2",
  "properties": {
"create_disposition": "CREATE_IF_NEEDED",
"dataset": "python_query_to_table_15729739549240",
"display_data": [],
"encoding": {
  "@type": "kind:windowed_value",
  "component_encodings": [
{
  "@type": 
"RowAsDictJsonCoder$eNprYE5OLEhMzkiNT0pNzNXLzNdLTy7QS8pMLyxNLaqML8nPzynmCsovdyx2yUwu8SrOz3POT0kt4ipk0GwsZKwtZErSAwBK5xfp",
  "component_encodings": []
},
{
  "@type": "kind:global_window"
}
  ],
  "is_wrapper": true
},
"format": "bigquery",
"parallel_input": {
  "@type": "OutputReference",
  "output_name": "out",
  "step_name": "s1"
},
"schema": "{\"fields\": [{\"name\": \"fruit\", \"type\": \"STRING\", 
\"mode\": \"NULLABLE\"}]}",
"table": "output_table",
"user_name": "write/WriteToBigQuery/NativeWrite",
"write_disposition": "WRITE_EMPTY"
  }
}
  ],
  "type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: 
root: INFO: Created job with id: [2019-11-05_09_12_50-7467189542985167820]
root: INFO: To access the Dataflow monitoring console, please navigate to 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_09_12_50-7467189542985167820?project=apache-beam-testing
root: INFO: Job 2019-11-05_09_12_50-7467189542985167820 is in state 
JOB_STATE_RUNNING
root: INFO: 2019-11-05T17:12:50.072Z: JOB_MESSAGE_DETAILED: Autoscaling is 
enabled for job 2019-11-05_09_12_50-7467189542985167820. The number of workers 
will be between 1 and 1000.
root: INFO: 2019-11-05T17:12:50.072Z: JOB_MESSAGE_DETAILED: Autoscaling was 
automatically enabled for job 2019-11-05_09_12_50-7467189542985167820.
root: INFO: 2019-11-05T17:12:53.490Z: JOB_MESSAGE_DETAILED: Checking 
permissions granted to controller Service Account.
root: INFO: 2019-11-05T17:12:54.288Z: JOB_MESSAGE_BASIC: Worker configuration: 
n1-standard-1 in us-central1-f.
root: INFO: 2019-11-05T17:12:55.015Z: JOB_MESSAGE_DETAILED: Expanding 
CoGroupByKey operations into optimizable parts.
root: INFO: 2019-11-05T17:12:55.054Z: JOB_MESSAGE_DETAILED: Expanding 
GroupByKey operations into optimizable parts.
root: INFO: 2019-11-05T17:12:55.094Z: JOB_MESSAGE_DETAILED: Lifting 
ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2019-11-05T17:12:55.125Z: JOB_MESSAGE_DEBUG: Annotating graph with 
Autotuner information.
root: INFO: 2019-11-05T17:12:55.339Z: JOB_MESSAGE_DETAILED: Fusing adjacent 
ParDo, Read, Write, and Flatten operations
root: INFO: 2019-11-05T17:12:55.378Z: JOB_MESSAGE_DETAILED: Fusing consumer 
write/WriteToBigQuery/NativeWrite into read
root: INFO: 2019-11-05T17:12:55.415Z: JOB_MESSAGE_DEBUG: Workflow config is 
missing a default resource spec.
root: INFO: 

Build failed in Jenkins: beam_PostCommit_Python35 #910

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[robertwb] [BEAM-8435] Implement PaneInfo computation for Python.


--
[...truncated 128.47 KB...]
urllib3.connectionpool: DEBUG: http://metadata.google.internal:80 "GET 
/computeMetadata/v1/instance/service-accounts/844138762903-comp...@developer.gserviceaccount.com/token
 HTTP/1.1" 200 181
urllib3.connectionpool: DEBUG: Starting new HTTPS connection (1): 
www.googleapis.com:443
urllib3.connectionpool: DEBUG: https://www.googleapis.com:443 "POST 
/bigquery/v2/projects/apache-beam-testing/jobs HTTP/1.1" 200 None
urllib3.connectionpool: DEBUG: https://www.googleapis.com:443 "GET 
/bigquery/v2/projects/apache-beam-testing/queries/da857d4f-2260-4c44-8a86-03b7d88521aa?timeoutMs=1=US=0
 HTTP/1.1" 200 None
urllib3.connectionpool: DEBUG: https://www.googleapis.com:443 "GET 
/bigquery/v2/projects/apache-beam-testing/jobs/da857d4f-2260-4c44-8a86-03b7d88521aa?location=US
 HTTP/1.1" 200 None
urllib3.connectionpool: DEBUG: https://www.googleapis.com:443 "GET 
/bigquery/v2/projects/apache-beam-testing/datasets/_7357fab0f784d2a7327ddbe81cdd1f4ca7e429cd/tables/anon35aa1d3f34872e9c0e969bd7a81d4895dba56464/data
 HTTP/1.1" 200 None
root: INFO: Result of query is: []
root: INFO: Deleting dataset python_write_to_table_15729724453683 in project 
apache-beam-testing
- >> end captured logging << -
:1208:
 BeamDeprecationWarning: options is deprecated since First stable release. 
References to .options will not be supported
  self.table_reference.projectId = pcoll.pipeline.options.view_as(
:1208:
 BeamDeprecationWarning: options is deprecated since First stable release. 
References to .options will not be supported
  self.table_reference.projectId = pcoll.pipeline.options.view_as(
:795:
 BeamDeprecationWarning: options is deprecated since First stable release. 
References to .options will not be supported
  temp_location = p.options.view_as(GoogleCloudOptions).temp_location

--
XML: nosetests-postCommitIT-direct-py35.xml
--
XML: 

--
Ran 15 tests in 35.988s

FAILED (SKIP=1, failures=1)

> Task :sdks:python:test-suites:direct:py35:postCommitIT FAILED

> Task :sdks:python:test-suites:dataflow:py35:postCommitIT
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)
:1211:
 BeamDeprecationWarning: options is deprecated since First stable release. 
References to .options will not be supported
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_08_48_08-2142010941516302675?project=apache-beam-testing
  experiments = p.options.view_as(DebugOptions).experiments or []
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_09_03_18-14251067506950311868?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_09_11_15-1555873846204908662?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_09_19_52-9673751567250827863?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_09_28_22-3663977256155970627?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_08_48_03-4394177010130848176?project=apache-beam-testing
:709:
 BeamDeprecationWarning: BigQuerySink is deprecated since 2.11.0. Use 
WriteToBigQuery instead.
  kms_key=transform.kms_key))
Worker logs: 

Build failed in Jenkins: beam_PostCommit_SQL #3127

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[chambers] Fix expected step name in Python PTransformTest test_read_metrics

[chambers] generalized step name matching


--
[...truncated 10.86 KB...]
> Task :model:fn-execution:extractProto
> Task :sdks:java:extensions:google-cloud-platform-core:processTestResources 
> NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processTestResources NO-SOURCE
> Task :sdks:java:extensions:sql:zetasql:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:extractProto
> Task :model:job-management:processResources
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :sdks:java:io:mongodb:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :sdks:java:io:mongodb:processTestResources NO-SOURCE
> Task :sdks:java:io:common:compileJava NO-SOURCE
> Task :sdks:java:testing:test-utils:processTestResources NO-SOURCE
> Task :sdks:java:io:common:processResources NO-SOURCE
> Task :sdks:java:io:common:classes UP-TO-DATE
> Task :sdks:java:extensions:sql:datacatalog:processResources NO-SOURCE
> Task :sdks:java:extensions:sql:datacatalog:processTestResources NO-SOURCE
> Task :model:fn-execution:processResources
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :sdks:java:io:common:jar
> Task :sdks:java:io:common:processTestResources NO-SOURCE
> Task :sdks:java:core:processResources
> Task :sdks:java:core:generateTestAvroProtocol NO-SOURCE
> Task :sdks:java:extensions:sql:jdbc:processResources
> Task :sdks:java:extensions:sql:jdbc:processTestResources NO-SOURCE
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :sdks:java:core:generateTestAvroJava
> Task :sdks:java:core:generateTestGrammarSource NO-SOURCE
> Task :sdks:java:core:processTestResources
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromCalciteCore
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromSrc
> Task :model:pipeline:generateProto
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :sdks:java:extensions:sql:generateFmppSources
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:extractIncludeProto
> Task :model:job-management:generateProto
> Task :model:fn-execution:generateProto
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes

> Task :sdks:java:extensions:sql:compileJavacc
Java Compiler Compiler Version 4.0 (Parser Generator)
(type "javacc" with no arguments for help)
Reading from file 

 . . .
Note: UNICODE_INPUT option is specified. Please make sure you create the 
parser/lexer using a Reader with the correct character encoding.
Warning: Lookahead adequacy checking not being performed since option LOOKAHEAD 
is more than 1.  Set option FORCE_LA_CHECK to true to force checking.
File "TokenMgrError.java" does not exist.  Will create one.
File "ParseException.java" does not exist.  Will create one.
File "Token.java" does not exist.  Will create one.
File "SimpleCharStream.java" does not exist.  Will create one.
Parser generated with 0 errors and 1 warnings.

> Task :sdks:java:extensions:sql:processResources
> Task :sdks:java:extensions:sql:processTestResources
> Task :model:pipeline:shadowJar
> Task :model:job-management:shadowJar
> Task :model:fn-execution:shadowJar
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :vendor:sdks-java-extensions-protobuf:compileJava FROM-CACHE
> Task :vendor:sdks-java-extensions-protobuf:classes UP-TO-DATE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :runners:local-java:jar
> Task :vendor:sdks-java-extensions-protobuf:shadowJar
> Task :sdks:java:core:jar
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :sdks:java:fn-execution:jar
> Task :runners:core-construction-java:jar
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:jar
> Task :runners:core-java:compileJava 

Build failed in Jenkins: beam_PostCommit_XVR_Flink #854

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[chambers] Fix expected step name in Python PTransformTest test_read_metrics

[chambers] generalized step name matching


--
[...truncated 4.27 MB...]
[flink-akka.actor.default-dispatcher-4] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (1/16) (2011fffd660f1288ee7e7c7247ff8ad4) switched from 
RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (13/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - MapPartition (MapPartition at 
[3]assert_that/{Group, Unkey, Match}) (13/16) 
(b4b9f5f23474827c289401ec9ecb9d75) switched from RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (13/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - Freeing task resources for 
MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (13/16) 
(b4b9f5f23474827c289401ec9ecb9d75).
[flink-akka.actor.default-dispatcher-4] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (13/16) (7d241144ae57c0c918f25ed7e83d0169) switched from 
CREATED to SCHEDULED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (13/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem 
streams are closed for task MapPartition (MapPartition at 
[3]assert_that/{Group, Unkey, Match}) (13/16) 
(b4b9f5f23474827c289401ec9ecb9d75) [FINISHED]
[flink-akka.actor.default-dispatcher-8] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task MapPartition 
(MapPartition at [3]assert_that/{Group, Unkey, Match}) 
b4b9f5f23474827c289401ec9ecb9d75.
[flink-akka.actor.default-dispatcher-4] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (13/16) (7d241144ae57c0c918f25ed7e83d0169) switched from 
SCHEDULED to DEPLOYING.
[flink-akka.actor.default-dispatcher-4] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - Deploying DataSink 
(DiscardingOutput) (13/16) (attempt #0) to 1179307f-fd9a-4201-b93e-6400229bf319 
@ localhost (dataPort=-1)
[flink-akka.actor.default-dispatcher-8] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Received task DataSink 
(DiscardingOutput) (13/16).
[flink-akka.actor.default-dispatcher-4] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - MapPartition 
(MapPartition at [3]assert_that/{Group, Unkey, Match}) (13/16) 
(b4b9f5f23474827c289401ec9ecb9d75) switched from RUNNING to FINISHED.
[DataSink (DiscardingOutput) (13/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (13/16) 
(7d241144ae57c0c918f25ed7e83d0169) switched from CREATED to DEPLOYING.
[DataSink (DiscardingOutput) (13/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Creating FileSystem stream leak 
safety net for task DataSink (DiscardingOutput) (13/16) 
(7d241144ae57c0c918f25ed7e83d0169) [DEPLOYING]
[DataSink (DiscardingOutput) (13/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Loading JAR files for task DataSink 
(DiscardingOutput) (13/16) (7d241144ae57c0c918f25ed7e83d0169) [DEPLOYING].
[DataSink (DiscardingOutput) (13/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Registering task at network: 
DataSink (DiscardingOutput) (13/16) (7d241144ae57c0c918f25ed7e83d0169) 
[DEPLOYING].
[DataSink (DiscardingOutput) (13/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (13/16) 
(7d241144ae57c0c918f25ed7e83d0169) switched from DEPLOYING to RUNNING.
[flink-akka.actor.default-dispatcher-8] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (13/16) (7d241144ae57c0c918f25ed7e83d0169) switched from 
DEPLOYING to RUNNING.
[DataSink (DiscardingOutput) (13/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (13/16) 
(7d241144ae57c0c918f25ed7e83d0169) switched from RUNNING to FINISHED.
[DataSink (DiscardingOutput) (13/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Freeing task resources for DataSink 
(DiscardingOutput) (13/16) (7d241144ae57c0c918f25ed7e83d0169).
[DataSink (DiscardingOutput) (13/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem streams are 
closed for task DataSink (DiscardingOutput) (13/16) 
(7d241144ae57c0c918f25ed7e83d0169) [FINISHED]
[flink-akka.actor.default-dispatcher-8] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task DataSink 
(DiscardingOutput) 7d241144ae57c0c918f25ed7e83d0169.
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (13/16) (7d241144ae57c0c918f25ed7e83d0169) switched from 
RUNNING to 

Jenkins build is back to normal : beam_PostCommit_Python36 #916

2019-11-05 Thread Apache Jenkins Server
See 



-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org



Jenkins build is back to normal : beam_PerformanceTests_Kafka_IO #12

2019-11-05 Thread Apache Jenkins Server
See 



-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org



Build failed in Jenkins: beam_PostCommit_Java_Nexmark_Flink #4220

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[kirillkozlov] Created a MongoDbTable and a provider for it

[github] Updated assertion format string to use [] vs ?

[github] [BEAM-8539] Provide an initial definition of all job states and the


--
[...truncated 15.80 KB...]
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :model:pipeline:generateProto
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromCalciteCore
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromSrc
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :sdks:java:extensions:sql:generateFmppSources
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:generateProto
> Task :model:fn-execution:generateProto
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes

> Task :sdks:java:extensions:sql:compileJavacc
Java Compiler Compiler Version 4.0 (Parser Generator)
(type "javacc" with no arguments for help)
Reading from file 

 . . .
Note: UNICODE_INPUT option is specified. Please make sure you create the 
parser/lexer using a Reader with the correct character encoding.
Warning: Lookahead adequacy checking not being performed since option LOOKAHEAD 
is more than 1.  Set option FORCE_LA_CHECK to true to force checking.
File "TokenMgrError.java" does not exist.  Will create one.
File "ParseException.java" does not exist.  Will create one.
File "Token.java" does not exist.  Will create one.
File "SimpleCharStream.java" does not exist.  Will create one.
Parser generated with 0 errors and 1 warnings.

> Task :sdks:java:extensions:sql:processResources
> Task :model:pipeline:shadowJar
> Task :model:job-management:shadowJar
> Task :model:fn-execution:shadowJar
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :sdks:java:extensions:join-library:compileJava FROM-CACHE
> Task :sdks:java:extensions:join-library:classes UP-TO-DATE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :sdks:java:extensions:join-library:jar
> Task :runners:local-java:jar
> Task :sdks:java:io:parquet:compileJava FROM-CACHE
> Task :sdks:java:io:parquet:classes UP-TO-DATE
> Task :sdks:java:io:parquet:jar
> Task :vendor:sdks-java-extensions-protobuf:compileJava FROM-CACHE
> Task :vendor:sdks-java-extensions-protobuf:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :sdks:java:io:kafka:jar
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :sdks:java:fn-execution:jar
> Task :vendor:sdks-java-extensions-protobuf:shadowJar
> Task :runners:core-construction-java:jar
> Task :sdks:java:testing:test-utils:compileJava FROM-CACHE
> Task :sdks:java:testing:test-utils:classes UP-TO-DATE
> Task :sdks:java:testing:test-utils:jar
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :runners:core-java:jar
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:jar
> Task :sdks:java:harness:compileJava FROM-CACHE
> Task :sdks:java:harness:classes UP-TO-DATE
> Task :sdks:java:harness:jar
> Task :sdks:java:core:jar
> Task :sdks:java:io:google-cloud-platform:compileJava FROM-CACHE
> Task :sdks:java:io:google-cloud-platform:classes UP-TO-DATE
> Task :sdks:java:io:google-cloud-platform:jar

> Task :sdks:java:io:mongodb:compileJava
Note: 

 uses or overrides a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.

> Task :sdks:java:io:mongodb:classes
> Task :sdks:java:io:mongodb:jar
> Task :sdks:java:harness:shadowJar
> 

beam_PostCommit_Java_PVR_Spark_Batch - Build # 1212 - Aborted

2019-11-05 Thread Apache Jenkins Server
The Apache Jenkins build system has built beam_PostCommit_Java_PVR_Spark_Batch 
(build #1212)

Status: Aborted

Check console output at 
https://builds.apache.org/job/beam_PostCommit_Java_PVR_Spark_Batch/1212/ to 
view the results.

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org

Jenkins build is back to normal : beam_PostCommit_Python36 #914

2019-11-05 Thread Apache Jenkins Server
See 



-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org



Build failed in Jenkins: beam_PostCommit_Java_Nexmark_Flink #4221

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[chambers] Fix expected step name in Python PTransformTest test_read_metrics

[chambers] generalized step name matching


--
[...truncated 9.97 KB...]
> Task :runners:flink:1.9:copyResourcesOverrides NO-SOURCE
> Task :sdks:java:core:generateAvroJava NO-SOURCE
> Task :sdks:java:extensions:join-library:processResources NO-SOURCE
> Task :model:job-management:extractProto
> Task :model:fn-execution:extractProto
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :sdks:java:io:parquet:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:extractProto
> Task :sdks:java:io:mongodb:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :runners:flink:1.9:copySourceOverrides
> Task :sdks:java:testing:nexmark:processResources
> Task :runners:flink:1.9:copyTestResourcesOverrides NO-SOURCE
> Task :model:job-management:processResources
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :runners:flink:1.9:processResources
> Task :model:fn-execution:processResources
> Task :sdks:java:build-tools:compileJava FROM-CACHE
> Task :sdks:java:build-tools:processResources
> Task :sdks:java:build-tools:classes
> Task :sdks:java:core:processResources
> Task :sdks:java:build-tools:jar
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromCalciteCore
> Task :model:pipeline:generateProto
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromSrc
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :sdks:java:extensions:sql:generateFmppSources
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:generateProto
> Task :model:fn-execution:generateProto
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes

> Task :sdks:java:extensions:sql:compileJavacc
Java Compiler Compiler Version 4.0 (Parser Generator)
(type "javacc" with no arguments for help)
Reading from file 

 . . .
Note: UNICODE_INPUT option is specified. Please make sure you create the 
parser/lexer using a Reader with the correct character encoding.
Warning: Lookahead adequacy checking not being performed since option LOOKAHEAD 
is more than 1.  Set option FORCE_LA_CHECK to true to force checking.
File "TokenMgrError.java" does not exist.  Will create one.
File "ParseException.java" does not exist.  Will create one.
File "Token.java" does not exist.  Will create one.
File "SimpleCharStream.java" does not exist.  Will create one.
Parser generated with 0 errors and 1 warnings.

> Task :sdks:java:extensions:sql:processResources
> Task :model:pipeline:shadowJar
> Task :model:job-management:shadowJar
> Task :model:fn-execution:shadowJar
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :runners:local-java:jar
> Task :vendor:sdks-java-extensions-protobuf:compileJava FROM-CACHE
> Task :vendor:sdks-java-extensions-protobuf:classes UP-TO-DATE
> Task :sdks:java:extensions:join-library:compileJava FROM-CACHE
> Task :sdks:java:extensions:join-library:classes UP-TO-DATE
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :sdks:java:io:mongodb:compileJava FROM-CACHE
> Task :sdks:java:io:mongodb:classes UP-TO-DATE
> Task :sdks:java:extensions:join-library:jar
> Task :vendor:sdks-java-extensions-protobuf:shadowJar
> Task :sdks:java:io:mongodb:jar
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :sdks:java:io:kafka:jar
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:fn-execution:jar
> Task :runners:core-construction-java:jar
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:jar
> Task :sdks:java:core:jar
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> 

beam_PostCommit_Python37 - Build # 869 - Aborted

2019-11-05 Thread Apache Jenkins Server
The Apache Jenkins build system has built beam_PostCommit_Python37 (build #869)

Status: Aborted

Check console output at 
https://builds.apache.org/job/beam_PostCommit_Python37/869/ to view the results.

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org

Build failed in Jenkins: beam_PostCommit_Python36 #915

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[kirillkozlov] Created a MongoDbTable and a provider for it

[github] Updated assertion format string to use [] vs ?

[github] [BEAM-8539] Provide an initial definition of all job states and the


--
[...truncated 129.24 KB...]
"output_name": "out",
"user_name": "ReadFromPubSub/Read.out"
  }
],
"pubsub_subscription": 
"projects/apache-beam-testing/subscriptions/exercise_streaming_metrics_subscription_input49b223b4-81c6-441c-b60f-8dd11e52fcf3",
"user_name": "ReadFromPubSub/Read"
  }
},
{
  "kind": "ParallelDo",
  "name": "s2",
  "properties": {
"display_data": [
  {
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.ParDo",
"shortValue": "StreamingUserMetricsDoFn",
"type": "STRING",
"value": 
"apache_beam.runners.dataflow.dataflow_exercise_streaming_metrics_pipeline.StreamingUserMetricsDoFn"
  }
],
"non_parallel_inputs": {},
"output_info": [
  {
"encoding": {
  "@type": "kind:windowed_value",
  "component_encodings": [
{
  "@type": "kind:bytes"
},
{
  "@type": "kind:global_window"
}
  ],
  "is_wrapper": true
},
"output_name": "out",
"user_name": "generate_metrics.out"
  }
],
"parallel_input": {
  "@type": "OutputReference",
  "output_name": "out",
  "step_name": "s1"
},
"serialized_fn": "ref_AppliedPTransform_generate_metrics_4",
"user_name": "generate_metrics"
  }
},
{
  "kind": "ParallelWrite",
  "name": "s3",
  "properties": {
"display_data": [],
"encoding": {
  "@type": "kind:windowed_value",
  "component_encodings": [
{
  "@type": "kind:bytes"
},
{
  "@type": "kind:global_window"
}
  ],
  "is_wrapper": true
},
"format": "pubsub",
"parallel_input": {
  "@type": "OutputReference",
  "output_name": "out",
  "step_name": "s2"
},
"pubsub_topic": 
"projects/apache-beam-testing/topics/exercise_streaming_metrics_topic_output49b223b4-81c6-441c-b60f-8dd11e52fcf3",
"user_name": "dump_to_pub/Write/NativeWrite"
  }
}
  ],
  "type": "JOB_TYPE_STREAMING"
}
root: INFO: Create job: 
root: INFO: Created job with id: [2019-11-05_11_43_22-9730356978537256421]
root: INFO: To access the Dataflow monitoring console, please navigate to 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_11_43_22-9730356978537256421?project=apache-beam-testing
root: INFO: Job 2019-11-05_11_43_22-9730356978537256421 is in state 
JOB_STATE_RUNNING
root: INFO: 2019-11-05T19:43:25.686Z: JOB_MESSAGE_DETAILED: Checking 
permissions granted to controller Service Account.
root: INFO: 2019-11-05T19:43:26.740Z: JOB_MESSAGE_BASIC: Worker configuration: 
n1-standard-4 in us-central1-f.
root: INFO: 2019-11-05T19:43:27.330Z: JOB_MESSAGE_DETAILED: Expanding 
SplittableParDo operations into optimizable parts.
root: INFO: 2019-11-05T19:43:27.335Z: JOB_MESSAGE_DETAILED: Expanding 
CollectionToSingleton operations into optimizable parts.
root: INFO: 2019-11-05T19:43:27.359Z: JOB_MESSAGE_DETAILED: Expanding 
CoGroupByKey operations into optimizable parts.
root: INFO: 2019-11-05T19:43:27.373Z: JOB_MESSAGE_DETAILED: Expanding 
SplittableProcessKeyed operations into optimizable parts.
root: INFO: 2019-11-05T19:43:27.376Z: JOB_MESSAGE_DETAILED: Expanding 
GroupByKey operations into streaming Read/Write steps
root: INFO: 2019-11-05T19:43:27.379Z: JOB_MESSAGE_DEBUG: Annotating graph with 
Autotuner information.
root: INFO: 2019-11-05T19:43:27.395Z: JOB_MESSAGE_DETAILED: Fusing adjacent 
ParDo, Read, Write, and Flatten operations
root: INFO: 2019-11-05T19:43:27.398Z: JOB_MESSAGE_DETAILED: Fusing consumer 
generate_metrics into ReadFromPubSub/Read
root: INFO: 2019-11-05T19:43:27.401Z: JOB_MESSAGE_DETAILED: Fusing consumer 
dump_to_pub/Write/NativeWrite into generate_metrics
root: INFO: 2019-11-05T19:43:27.410Z: JOB_MESSAGE_DEBUG: Adding StepResource 
setup and teardown to workflow graph.
root: INFO: 2019-11-05T19:43:27.435Z: JOB_MESSAGE_DEBUG: Adding workflow start 
and stop steps.
root: INFO: 2019-11-05T19:43:27.485Z: JOB_MESSAGE_DEBUG: Assigning stage ids.
root: INFO: 2019-11-05T19:43:27.637Z: JOB_MESSAGE_DEBUG: Executing wait step 
start2
root: INFO: 2019-11-05T19:43:27.652Z: JOB_MESSAGE_DEBUG: Starting worker pool 
setup.
root: INFO: 2019-11-05T19:43:27.658Z: 

beam_PostCommit_Java_PVR_Spark_Batch - Build # 1211 - Aborted

2019-11-05 Thread Apache Jenkins Server
The Apache Jenkins build system has built beam_PostCommit_Java_PVR_Spark_Batch 
(build #1211)

Status: Aborted

Check console output at 
https://builds.apache.org/job/beam_PostCommit_Java_PVR_Spark_Batch/1211/ to 
view the results.

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org

Jenkins build is back to normal : beam_PreCommit_Java_Cron #2000

2019-11-05 Thread Apache Jenkins Server
See 



-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org



Build failed in Jenkins: beam_PostCommit_XVR_Flink #853

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[kirillkozlov] Created a MongoDbTable and a provider for it

[github] Updated assertion format string to use [] vs ?


--
[...truncated 4.28 MB...]
[flink-akka.actor.default-dispatcher-7] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (5/16) (8331118505c70b4f039e9c44c5ad881a) switched from 
RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (13/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - MapPartition (MapPartition at 
[3]assert_that/{Group, Unkey, Match}) (13/16) 
(50fc322b86aeb8f771c9874ea9d9f149) switched from RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (13/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - Freeing task resources for 
MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (13/16) 
(50fc322b86aeb8f771c9874ea9d9f149).
[flink-akka.actor.default-dispatcher-5] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (13/16) (8918fbc802da72190add4a10f011d7ee) switched from 
CREATED to SCHEDULED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (13/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem 
streams are closed for task MapPartition (MapPartition at 
[3]assert_that/{Group, Unkey, Match}) (13/16) 
(50fc322b86aeb8f771c9874ea9d9f149) [FINISHED]
[flink-akka.actor.default-dispatcher-7] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task MapPartition 
(MapPartition at [3]assert_that/{Group, Unkey, Match}) 
50fc322b86aeb8f771c9874ea9d9f149.
[flink-akka.actor.default-dispatcher-5] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (13/16) (8918fbc802da72190add4a10f011d7ee) switched from 
SCHEDULED to DEPLOYING.
[flink-akka.actor.default-dispatcher-5] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - Deploying DataSink 
(DiscardingOutput) (13/16) (attempt #0) to 924dae49-404b-46d9-ae28-e4383ccbc183 
@ localhost (dataPort=-1)
[flink-akka.actor.default-dispatcher-5] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - MapPartition 
(MapPartition at [3]assert_that/{Group, Unkey, Match}) (13/16) 
(50fc322b86aeb8f771c9874ea9d9f149) switched from RUNNING to FINISHED.
[flink-akka.actor.default-dispatcher-7] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Received task DataSink 
(DiscardingOutput) (13/16).
[DataSink (DiscardingOutput) (13/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (13/16) 
(8918fbc802da72190add4a10f011d7ee) switched from CREATED to DEPLOYING.
[DataSink (DiscardingOutput) (13/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Creating FileSystem stream leak 
safety net for task DataSink (DiscardingOutput) (13/16) 
(8918fbc802da72190add4a10f011d7ee) [DEPLOYING]
[DataSink (DiscardingOutput) (13/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Loading JAR files for task DataSink 
(DiscardingOutput) (13/16) (8918fbc802da72190add4a10f011d7ee) [DEPLOYING].
[DataSink (DiscardingOutput) (13/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Registering task at network: 
DataSink (DiscardingOutput) (13/16) (8918fbc802da72190add4a10f011d7ee) 
[DEPLOYING].
[DataSink (DiscardingOutput) (13/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (13/16) 
(8918fbc802da72190add4a10f011d7ee) switched from DEPLOYING to RUNNING.
[flink-akka.actor.default-dispatcher-7] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (13/16) (8918fbc802da72190add4a10f011d7ee) switched from 
DEPLOYING to RUNNING.
[DataSink (DiscardingOutput) (13/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (13/16) 
(8918fbc802da72190add4a10f011d7ee) switched from RUNNING to FINISHED.
[DataSink (DiscardingOutput) (13/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Freeing task resources for DataSink 
(DiscardingOutput) (13/16) (8918fbc802da72190add4a10f011d7ee).
[DataSink (DiscardingOutput) (13/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem streams are 
closed for task DataSink (DiscardingOutput) (13/16) 
(8918fbc802da72190add4a10f011d7ee) [FINISHED]
[flink-akka.actor.default-dispatcher-7] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task DataSink 
(DiscardingOutput) 8918fbc802da72190add4a10f011d7ee.
[flink-akka.actor.default-dispatcher-7] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (13/16) (8918fbc802da72190add4a10f011d7ee) switched from 
RUNNING to 

Build failed in Jenkins: beam_PostCommit_XVR_Flink #851

2019-11-05 Thread Apache Jenkins Server
See 


Changes:


--
[...truncated 4.28 MB...]
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (12/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - MapPartition (MapPartition at 
[3]assert_that/{Group, Unkey, Match}) (12/16) 
(7bfef59d9c2794a2cfb0426580754594) switched from RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (12/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - Freeing task resources for 
MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (12/16) 
(7bfef59d9c2794a2cfb0426580754594).
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (12/16) (ea5f55efafd9465fdeb1462b3ec861e8) switched from 
CREATED to SCHEDULED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (12/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem 
streams are closed for task MapPartition (MapPartition at 
[3]assert_that/{Group, Unkey, Match}) (12/16) 
(7bfef59d9c2794a2cfb0426580754594) [FINISHED]
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task MapPartition 
(MapPartition at [3]assert_that/{Group, Unkey, Match}) 
7bfef59d9c2794a2cfb0426580754594.
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (12/16) (ea5f55efafd9465fdeb1462b3ec861e8) switched from 
SCHEDULED to DEPLOYING.
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - Deploying DataSink 
(DiscardingOutput) (12/16) (attempt #0) to 1e519951-edad-48ee-8f8a-a6c0d5d29758 
@ localhost (dataPort=-1)
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Received task DataSink 
(DiscardingOutput) (12/16).
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - MapPartition 
(MapPartition at [3]assert_that/{Group, Unkey, Match}) (12/16) 
(7bfef59d9c2794a2cfb0426580754594) switched from RUNNING to FINISHED.
[DataSink (DiscardingOutput) (12/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (12/16) 
(ea5f55efafd9465fdeb1462b3ec861e8) switched from CREATED to DEPLOYING.
[DataSink (DiscardingOutput) (12/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Creating FileSystem stream leak 
safety net for task DataSink (DiscardingOutput) (12/16) 
(ea5f55efafd9465fdeb1462b3ec861e8) [DEPLOYING]
[DataSink (DiscardingOutput) (12/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Loading JAR files for task DataSink 
(DiscardingOutput) (12/16) (ea5f55efafd9465fdeb1462b3ec861e8) [DEPLOYING].
[DataSink (DiscardingOutput) (12/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Registering task at network: 
DataSink (DiscardingOutput) (12/16) (ea5f55efafd9465fdeb1462b3ec861e8) 
[DEPLOYING].
[DataSink (DiscardingOutput) (12/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (12/16) 
(ea5f55efafd9465fdeb1462b3ec861e8) switched from DEPLOYING to RUNNING.
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (12/16) (ea5f55efafd9465fdeb1462b3ec861e8) switched from 
DEPLOYING to RUNNING.
[DataSink (DiscardingOutput) (12/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (12/16) 
(ea5f55efafd9465fdeb1462b3ec861e8) switched from RUNNING to FINISHED.
[DataSink (DiscardingOutput) (12/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Freeing task resources for DataSink 
(DiscardingOutput) (12/16) (ea5f55efafd9465fdeb1462b3ec861e8).
[DataSink (DiscardingOutput) (12/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem streams are 
closed for task DataSink (DiscardingOutput) (12/16) 
(ea5f55efafd9465fdeb1462b3ec861e8) [FINISHED]
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task DataSink 
(DiscardingOutput) ea5f55efafd9465fdeb1462b3ec861e8.
[flink-akka.actor.default-dispatcher-7] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (12/16) (ea5f55efafd9465fdeb1462b3ec861e8) switched from 
RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (15/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - MapPartition (MapPartition at 
[3]assert_that/{Group, Unkey, Match}) (15/16) 
(87e1ebdcda98c05740da21e69fe24775) switched from RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, 

Build failed in Jenkins: beam_PostCommit_XVR_Flink #852

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[github] [BEAM-8539] Provide an initial definition of all job states and the


--
[...truncated 4.28 MB...]
[flink-akka.actor.default-dispatcher-5] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task DataSink 
(DiscardingOutput) 5b37ee02b37c538b02bd75159ed48a5a.
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (13/16) (5b37ee02b37c538b02bd75159ed48a5a) switched from 
RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (14/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - MapPartition (MapPartition at 
[3]assert_that/{Group, Unkey, Match}) (14/16) 
(b36b2f459aca25873c06dc2118eb6761) switched from RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (14/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - Freeing task resources for 
MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (14/16) 
(b36b2f459aca25873c06dc2118eb6761).
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (14/16) (3be23d49dcb03286da7ffc312f822018) switched from 
CREATED to SCHEDULED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (14/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem 
streams are closed for task MapPartition (MapPartition at 
[3]assert_that/{Group, Unkey, Match}) (14/16) 
(b36b2f459aca25873c06dc2118eb6761) [FINISHED]
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (14/16) (3be23d49dcb03286da7ffc312f822018) switched from 
SCHEDULED to DEPLOYING.
[flink-akka.actor.default-dispatcher-5] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task MapPartition 
(MapPartition at [3]assert_that/{Group, Unkey, Match}) 
b36b2f459aca25873c06dc2118eb6761.
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - Deploying DataSink 
(DiscardingOutput) (14/16) (attempt #0) to bc0bc027-bd08-48cb-9953-80789572eac2 
@ localhost (dataPort=-1)
[flink-akka.actor.default-dispatcher-5] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Received task DataSink 
(DiscardingOutput) (14/16).
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - MapPartition 
(MapPartition at [3]assert_that/{Group, Unkey, Match}) (14/16) 
(b36b2f459aca25873c06dc2118eb6761) switched from RUNNING to FINISHED.
[DataSink (DiscardingOutput) (14/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (14/16) 
(3be23d49dcb03286da7ffc312f822018) switched from CREATED to DEPLOYING.
[DataSink (DiscardingOutput) (14/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Creating FileSystem stream leak 
safety net for task DataSink (DiscardingOutput) (14/16) 
(3be23d49dcb03286da7ffc312f822018) [DEPLOYING]
[DataSink (DiscardingOutput) (14/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Loading JAR files for task DataSink 
(DiscardingOutput) (14/16) (3be23d49dcb03286da7ffc312f822018) [DEPLOYING].
[DataSink (DiscardingOutput) (14/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Registering task at network: 
DataSink (DiscardingOutput) (14/16) (3be23d49dcb03286da7ffc312f822018) 
[DEPLOYING].
[DataSink (DiscardingOutput) (14/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (14/16) 
(3be23d49dcb03286da7ffc312f822018) switched from DEPLOYING to RUNNING.
[flink-akka.actor.default-dispatcher-5] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (14/16) (3be23d49dcb03286da7ffc312f822018) switched from 
DEPLOYING to RUNNING.
[DataSink (DiscardingOutput) (14/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (14/16) 
(3be23d49dcb03286da7ffc312f822018) switched from RUNNING to FINISHED.
[DataSink (DiscardingOutput) (14/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Freeing task resources for DataSink 
(DiscardingOutput) (14/16) (3be23d49dcb03286da7ffc312f822018).
[DataSink (DiscardingOutput) (14/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem streams are 
closed for task DataSink (DiscardingOutput) (14/16) 
(3be23d49dcb03286da7ffc312f822018) [FINISHED]
[flink-akka.actor.default-dispatcher-5] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task DataSink 
(DiscardingOutput) 

beam_PostCommit_Java_PVR_Spark_Batch - Build # 1210 - Aborted

2019-11-05 Thread Apache Jenkins Server
The Apache Jenkins build system has built beam_PostCommit_Java_PVR_Spark_Batch 
(build #1210)

Status: Aborted

Check console output at 
https://builds.apache.org/job/beam_PostCommit_Java_PVR_Spark_Batch/1210/ to 
view the results.

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org

Build failed in Jenkins: beam_PostCommit_SQL #3126

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[kirillkozlov] Created a MongoDbTable and a provider for it

[github] Updated assertion format string to use [] vs ?


--
[...truncated 3.25 KB...]
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy
> Task :buildSrc:spotlessGroovyCheck
> Task :buildSrc:spotlessGroovyGradle
> Task :buildSrc:spotlessGroovyGradleCheck
> Task :buildSrc:spotlessCheck
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validateTaskProperties FROM-CACHE
> Task :buildSrc:check
> Task :buildSrc:build
Configuration on demand is an incubating feature.
> Task :sdks:java:extensions:google-cloud-platform-core:processResources 
> NO-SOURCE
> Task :sdks:java:core:generateAvroProtocol NO-SOURCE
> Task :sdks:java:harness:processResources NO-SOURCE
> Task :sdks:java:fn-execution:processResources NO-SOURCE
> Task :runners:core-construction-java:processResources NO-SOURCE
> Task :runners:core-java:processResources NO-SOURCE
> Task :vendor:sdks-java-extensions-protobuf:processResources NO-SOURCE
> Task :sdks:java:core:generateAvroJava NO-SOURCE
> Task :runners:java-fn-execution:processResources NO-SOURCE
> Task :runners:local-java:processResources NO-SOURCE
> Task :sdks:java:extensions:join-library:processResources NO-SOURCE
> Task :sdks:java:io:parquet:processResources NO-SOURCE
> Task :model:job-management:extractProto
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :sdks:java:io:mongodb:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:extractProto
> Task :model:fn-execution:extractProto
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :sdks:java:io:mongodb:processTestResources NO-SOURCE
> Task :sdks:java:io:common:compileJava NO-SOURCE
> Task :runners:direct-java:processResources NO-SOURCE
> Task :sdks:java:extensions:sql:zetasql:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :sdks:java:extensions:sql:datacatalog:processResources NO-SOURCE
> Task :sdks:java:io:common:processResources NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processTestResources NO-SOURCE
> Task :sdks:java:extensions:google-cloud-platform-core:processTestResources 
> NO-SOURCE
> Task :sdks:java:testing:test-utils:processTestResources NO-SOURCE
> Task :sdks:java:io:common:classes UP-TO-DATE
> Task :sdks:java:extensions:sql:datacatalog:processTestResources NO-SOURCE
> Task :model:fn-execution:processResources
> Task :sdks:java:io:common:jar
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :model:job-management:processResources
> Task :sdks:java:io:common:processTestResources NO-SOURCE
> Task :sdks:java:extensions:sql:jdbc:processResources
> Task :sdks:java:extensions:sql:jdbc:processTestResources NO-SOURCE
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :sdks:java:core:processResources
> Task :sdks:java:core:generateTestAvroProtocol NO-SOURCE
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :model:pipeline:generateProto
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromCalciteCore
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromSrc
> Task :sdks:java:core:generateTestAvroJava
> Task :sdks:java:core:generateTestGrammarSource NO-SOURCE
> Task :sdks:java:core:processTestResources
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:extractIncludeProto
> Task :sdks:java:extensions:sql:generateFmppSources
> Task :model:job-management:generateProto
> Task :model:fn-execution:generateProto
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes

> Task :sdks:java:extensions:sql:compileJavacc
Java Compiler Compiler Version 4.0 (Parser Generator)
(type "javacc" with no arguments for help)
Reading from file 

 . . .
Note: UNICODE_INPUT option is specified. Please make sure you create the 
parser/lexer using a Reader with the correct character encoding.
Warning: Lookahead adequacy checking not being performed since option LOOKAHEAD 
is more than 1.  Set option FORCE_LA_CHECK to true to force checking.
File "TokenMgrError.java" does not exist.  Will create one.
File "ParseException.java" does not exist.  Will create one.
File "Token.java" does not exist.  Will create one.
File "SimpleCharStream.java" does not 

beam_sonarqube_report - Build # 998 - Aborted

2019-11-05 Thread Apache Jenkins Server
The Apache Jenkins build system has built beam_sonarqube_report (build #998)

Status: Aborted

Check console output at 
https://builds.apache.org/job/beam_sonarqube_report/998/ to view the results.

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org

beam_PostCommit_Python37 - Build # 870 - Aborted

2019-11-05 Thread Apache Jenkins Server
The Apache Jenkins build system has built beam_PostCommit_Python37 (build #870)

Status: Aborted

Check console output at 
https://builds.apache.org/job/beam_PostCommit_Python37/870/ to view the results.

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org

Build failed in Jenkins: beam_PreCommit_Python_pytest_Cron #7

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[mxm] Use better name for Flink current key supplier

[lukasz.gajowy] [BEAM-4420] Make it possible for kafka test k8s setup to pass 
namespace

[michal.walenia] [BEAM-7303] Rename runners.reference package to 
runners.portability

[lukasz.gajowy] [BEAM-4420] Refactor multiple things:

[lukasz.gajowy] [BEAM-4420] Create jenkins job for Kafka IOIT

[lukasz.gajowy] [BEAM-4420] Extract common Jenkins code

[lukasz.gajowy] [BEAM-4420] Fix cancelling timeouting Dataflow jobs

[robertwb] [BEAM-8435] Implement PaneInfo computation for Python.

[lgajowy] [BEAM-8558] Fix BigQueryIOIT flakiness (#9995)


--
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-7 (beam) in workspace 

No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init 
 >  # 
 > timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git fetch --tags --progress https://github.com/apache/beam.git 
 > +refs/heads/*:refs/remotes/origin/*
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # 
 > timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git 
 > +refs/heads/*:refs/remotes/origin/* 
 > +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/*
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 16fbe89104fd9ef71179627f4cebff8b0ea9cd90 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 16fbe89104fd9ef71179627f4cebff8b0ea9cd90
Commit message: "Merge pull request #9836 [BEAM-8435] Implement PaneInfo 
computation for Python."
 > git rev-list --no-walk 9f566394102a5a52684311ceea962f1720c16b1a # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the 
SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ 

 --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g 
-Dorg.gradle.jvmargs=-Xmx4g :pythonPreCommitPytest
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for 
details
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy
> Task :buildSrc:spotlessGroovyCheck
> Task :buildSrc:spotlessGroovyGradle
> Task :buildSrc:spotlessGroovyGradleCheck
> Task :buildSrc:spotlessCheck
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validateTaskProperties FROM-CACHE
> Task :buildSrc:check
> Task :buildSrc:build
Configuration on demand is an incubating feature.

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task 
':sdks:python:test-suites:tox:py35:preCommitPy35Pytest'.
> Task with path 'lint' not found in project 
> ':sdks:python:test-suites:tox:py35'.

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug 
option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with 
Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See 
https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 11s

Publishing build scan...
https://gradle.com/s/b5uhtyv36xk6e

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org



Build failed in Jenkins: beam_PostCommit_Java_Nexmark_Flink #4219

2019-11-05 Thread Apache Jenkins Server
See 


Changes:


--
[...truncated 8.42 KB...]
> Task :sdks:java:extensions:google-cloud-platform-core:processResources 
> NO-SOURCE
> Task :sdks:java:fn-execution:processResources NO-SOURCE
> Task :sdks:java:harness:processResources NO-SOURCE
> Task :vendor:sdks-java-extensions-protobuf:processResources NO-SOURCE
> Task :runners:java-fn-execution:processResources NO-SOURCE
> Task :sdks:java:core:generateAvroJava NO-SOURCE
> Task :runners:flink:1.9:copyResourcesOverrides NO-SOURCE
> Task :runners:direct-java:processResources NO-SOURCE
> Task :sdks:java:extensions:join-library:processResources NO-SOURCE
> Task :model:job-management:extractProto
> Task :sdks:java:extensions:protobuf:extractProto
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :sdks:java:io:parquet:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :model:job-management:processResources
> Task :model:fn-execution:extractProto
> Task :sdks:java:testing:nexmark:processResources
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :runners:flink:1.9:copySourceOverrides
> Task :runners:flink:1.9:copyTestResourcesOverrides NO-SOURCE
> Task :runners:flink:1.9:processResources
> Task :model:fn-execution:processResources
> Task :sdks:java:build-tools:compileJava FROM-CACHE
> Task :sdks:java:build-tools:processResources
> Task :sdks:java:build-tools:classes
> Task :sdks:java:core:processResources
> Task :sdks:java:build-tools:jar
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :model:pipeline:generateProto
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromCalciteCore
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromSrc
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :sdks:java:extensions:sql:generateFmppSources
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:generateProto
> Task :model:fn-execution:generateProto
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes

> Task :sdks:java:extensions:sql:compileJavacc
Java Compiler Compiler Version 4.0 (Parser Generator)
(type "javacc" with no arguments for help)
Reading from file 

 . . .
Note: UNICODE_INPUT option is specified. Please make sure you create the 
parser/lexer using a Reader with the correct character encoding.
Warning: Lookahead adequacy checking not being performed since option LOOKAHEAD 
is more than 1.  Set option FORCE_LA_CHECK to true to force checking.
File "TokenMgrError.java" does not exist.  Will create one.
File "ParseException.java" does not exist.  Will create one.
File "Token.java" does not exist.  Will create one.
File "SimpleCharStream.java" does not exist.  Will create one.
Parser generated with 0 errors and 1 warnings.

> Task :sdks:java:extensions:sql:processResources
> Task :model:pipeline:shadowJar
> Task :model:job-management:shadowJar
> Task :model:fn-execution:shadowJar
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :sdks:java:extensions:join-library:compileJava FROM-CACHE
> Task :sdks:java:extensions:join-library:classes UP-TO-DATE
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :vendor:sdks-java-extensions-protobuf:compileJava FROM-CACHE
> Task :vendor:sdks-java-extensions-protobuf:classes UP-TO-DATE
> Task :sdks:java:extensions:join-library:jar
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :runners:local-java:jar
> Task :sdks:java:io:kafka:jar
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :vendor:sdks-java-extensions-protobuf:shadowJar
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:core:jar
> Task :sdks:java:fn-execution:jar
> Task :runners:core-construction-java:jar
> Task :sdks:java:io:parquet:compileJava FROM-CACHE
> Task :sdks:java:io:parquet:classes UP-TO-DATE
> Task :sdks:java:io:parquet:jar
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> 

Jenkins build is back to normal : beam_PostCommit_Py_VR_Dataflow #5004

2019-11-05 Thread Apache Jenkins Server
See 



-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org



beam_PreCommit_Python_Cron - Build # 1999 - Aborted

2019-11-05 Thread Apache Jenkins Server
The Apache Jenkins build system has built beam_PreCommit_Python_Cron (build 
#1999)

Status: Aborted

Check console output at 
https://builds.apache.org/job/beam_PreCommit_Python_Cron/1999/ to view the 
results.

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org

Build failed in Jenkins: beam_PostCommit_Python36 #908

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[aryan.naraghi] Fix a bug related to zero-row responses


--
[...truncated 127.06 KB...]
},
"output_name": "out",
"user_name": "ReadFromPubSub/Read.out"
  }
],
"pubsub_subscription": 
"projects/apache-beam-testing/subscriptions/exercise_streaming_metrics_subscription_input7ccc1685-b563-4438-ba78-d0ec21aa9dad",
"user_name": "ReadFromPubSub/Read"
  }
},
{
  "kind": "ParallelDo",
  "name": "s2",
  "properties": {
"display_data": [
  {
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.ParDo",
"shortValue": "StreamingUserMetricsDoFn",
"type": "STRING",
"value": 
"apache_beam.runners.dataflow.dataflow_exercise_streaming_metrics_pipeline.StreamingUserMetricsDoFn"
  }
],
"non_parallel_inputs": {},
"output_info": [
  {
"encoding": {
  "@type": "kind:windowed_value",
  "component_encodings": [
{
  "@type": "kind:bytes"
},
{
  "@type": "kind:global_window"
}
  ],
  "is_wrapper": true
},
"output_name": "out",
"user_name": "generate_metrics.out"
  }
],
"parallel_input": {
  "@type": "OutputReference",
  "output_name": "out",
  "step_name": "s1"
},
"serialized_fn": "ref_AppliedPTransform_generate_metrics_4",
"user_name": "generate_metrics"
  }
},
{
  "kind": "ParallelWrite",
  "name": "s3",
  "properties": {
"display_data": [],
"encoding": {
  "@type": "kind:windowed_value",
  "component_encodings": [
{
  "@type": "kind:bytes"
},
{
  "@type": "kind:global_window"
}
  ],
  "is_wrapper": true
},
"format": "pubsub",
"parallel_input": {
  "@type": "OutputReference",
  "output_name": "out",
  "step_name": "s2"
},
"pubsub_topic": 
"projects/apache-beam-testing/topics/exercise_streaming_metrics_topic_output7ccc1685-b563-4438-ba78-d0ec21aa9dad",
"user_name": "dump_to_pub/Write/NativeWrite"
  }
}
  ],
  "type": "JOB_TYPE_STREAMING"
}
root: INFO: Create job: 
root: INFO: Created job with id: [2019-11-05_00_19_09-10380762630637283640]
root: INFO: To access the Dataflow monitoring console, please navigate to 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_00_19_09-10380762630637283640?project=apache-beam-testing
root: INFO: Job 2019-11-05_00_19_09-10380762630637283640 is in state 
JOB_STATE_RUNNING
root: INFO: 2019-11-05T08:19:13.228Z: JOB_MESSAGE_DETAILED: Checking 
permissions granted to controller Service Account.
root: INFO: 2019-11-05T08:19:14.172Z: JOB_MESSAGE_BASIC: Worker configuration: 
n1-standard-4 in us-central1-f.
root: INFO: 2019-11-05T08:19:14.783Z: JOB_MESSAGE_DETAILED: Expanding 
SplittableParDo operations into optimizable parts.
root: INFO: 2019-11-05T08:19:14.785Z: JOB_MESSAGE_DETAILED: Expanding 
CollectionToSingleton operations into optimizable parts.
root: INFO: 2019-11-05T08:19:14.793Z: JOB_MESSAGE_DETAILED: Expanding 
CoGroupByKey operations into optimizable parts.
root: INFO: 2019-11-05T08:19:14.801Z: JOB_MESSAGE_DETAILED: Expanding 
SplittableProcessKeyed operations into optimizable parts.
root: INFO: 2019-11-05T08:19:14.803Z: JOB_MESSAGE_DETAILED: Expanding 
GroupByKey operations into streaming Read/Write steps
root: INFO: 2019-11-05T08:19:14.805Z: JOB_MESSAGE_DEBUG: Annotating graph with 
Autotuner information.
root: INFO: 2019-11-05T08:19:14.819Z: JOB_MESSAGE_DETAILED: Fusing adjacent 
ParDo, Read, Write, and Flatten operations
root: INFO: 2019-11-05T08:19:14.821Z: JOB_MESSAGE_DETAILED: Fusing consumer 
generate_metrics into ReadFromPubSub/Read
root: INFO: 2019-11-05T08:19:14.823Z: JOB_MESSAGE_DETAILED: Fusing consumer 
dump_to_pub/Write/NativeWrite into generate_metrics
root: INFO: 2019-11-05T08:19:14.830Z: JOB_MESSAGE_DEBUG: Adding StepResource 
setup and teardown to workflow graph.
root: INFO: 2019-11-05T08:19:14.840Z: JOB_MESSAGE_DEBUG: Adding workflow start 
and stop steps.
root: INFO: 2019-11-05T08:19:14.875Z: JOB_MESSAGE_DEBUG: Assigning stage ids.
root: INFO: 2019-11-05T08:19:15.032Z: JOB_MESSAGE_DEBUG: Executing wait step 
start2
root: INFO: 2019-11-05T08:19:15.044Z: JOB_MESSAGE_DEBUG: Starting worker pool 
setup.
root: INFO: 2019-11-05T08:19:15.048Z: JOB_MESSAGE_BASIC: Starting 1 workers...
root: INFO: 2019-11-05T08:19:18.480Z: JOB_MESSAGE_BASIC: Executing operation 

Build failed in Jenkins: beam_PostCommit_Java_Nexmark_Flink #4213

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[lukasz.gajowy] [BEAM-8432] Enable building kotlin examples with different java 
versions


--
[...truncated 8.99 KB...]
> Task :sdks:java:harness:processResources NO-SOURCE
> Task :sdks:java:fn-execution:processResources NO-SOURCE
> Task :runners:core-construction-java:processResources NO-SOURCE
> Task :sdks:java:core:generateAvroJava NO-SOURCE
> Task :runners:direct-java:processResources NO-SOURCE
> Task :runners:flink:1.9:copyResourcesOverrides NO-SOURCE
> Task :model:fn-execution:extractProto
> Task :model:job-management:extractProto
> Task :sdks:java:extensions:join-library:processResources NO-SOURCE
> Task :sdks:java:io:parquet:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:extractProto
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :sdks:java:testing:nexmark:processResources
> Task :model:job-management:processResources
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :model:fn-execution:processResources
> Task :runners:flink:1.9:copySourceOverrides
> Task :runners:flink:1.9:copyTestResourcesOverrides NO-SOURCE
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :runners:flink:1.9:processResources
> Task :sdks:java:build-tools:compileJava FROM-CACHE
> Task :sdks:java:build-tools:processResources
> Task :sdks:java:build-tools:classes
> Task :sdks:java:core:processResources
> Task :sdks:java:build-tools:jar
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromCalciteCore
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromSrc
> Task :model:pipeline:generateProto
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :sdks:java:extensions:sql:generateFmppSources
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:generateProto
> Task :model:fn-execution:generateProto
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes

> Task :sdks:java:extensions:sql:compileJavacc
Java Compiler Compiler Version 4.0 (Parser Generator)
(type "javacc" with no arguments for help)
Reading from file 

 . . .
Warning: Lookahead adequacy checking not being performed since option LOOKAHEAD 
is more than 1.  Set option FORCE_LA_CHECK to true to force checking.
Note: UNICODE_INPUT option is specified. Please make sure you create the 
parser/lexer using a Reader with the correct character encoding.
File "TokenMgrError.java" does not exist.  Will create one.
File "ParseException.java" does not exist.  Will create one.
File "Token.java" does not exist.  Will create one.
File "SimpleCharStream.java" does not exist.  Will create one.
Parser generated with 0 errors and 1 warnings.

> Task :sdks:java:extensions:sql:processResources
> Task :model:pipeline:shadowJar
> Task :model:job-management:shadowJar
> Task :model:fn-execution:shadowJar
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :sdks:java:extensions:join-library:compileJava FROM-CACHE
> Task :sdks:java:extensions:join-library:classes UP-TO-DATE
> Task :vendor:sdks-java-extensions-protobuf:compileJava FROM-CACHE
> Task :vendor:sdks-java-extensions-protobuf:classes UP-TO-DATE
> Task :sdks:java:extensions:join-library:jar
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :runners:local-java:jar
> Task :sdks:java:io:kafka:jar
> Task :vendor:sdks-java-extensions-protobuf:shadowJar
> Task :sdks:java:core:jar
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :runners:core-construction-java:jar
> Task :sdks:java:fn-execution:jar
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :sdks:java:io:parquet:compileJava FROM-CACHE
> Task :sdks:java:io:parquet:classes UP-TO-DATE
> Task :sdks:java:io:parquet:jar
> Task 

Jenkins build is back to normal : beam_PreCommit_CommunityMetrics_Cron #1476

2019-11-05 Thread Apache Jenkins Server
See 



-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org



Build failed in Jenkins: beam_PreCommit_Java_Cron #1999

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[mxm] [BEAM-8549] Do not use keyed operator state for checkpoint buffering

[lukasz.gajowy] [BEAM-8432] Enable building kotlin examples with different java 
versions

[aryan.naraghi] Fix a bug related to zero-row responses


--
[...truncated 502.22 KB...]
> Task :sdks:java:extensions:sql:zetasql:testClasses UP-TO-DATE
> Task :sdks:java:extensions:sql:zetasql:analyzeTestClassesDependencies SKIPPED
> Task :sdks:java:extensions:sql:zetasql:analyzeDependencies SKIPPED
> Task :sdks:java:extensions:kryo:check
> Task :sdks:java:extensions:kryo:build
> Task :sdks:java:extensions:sketching:check
> Task :sdks:java:extensions:sketching:build
> Task :sdks:java:extensions:sketching:buildDependents
> Task :sdks:java:io:amazon-web-services2:check
> Task :sdks:java:io:amazon-web-services2:build
> Task :sdks:java:io:amazon-web-services2:buildDependents

> Task :sdks:java:extensions:sql:jdbc:spotbugsMain
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in 
[jar:file:/home/jenkins/.gradle/caches/modules-2/files-2.1/org.slf4j/slf4j-simple/1.7.25/8dacf9514f0c707cbbcdd6fd699e8940d42fb54e/slf4j-simple-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in 
[jar:file:/home/jenkins/.gradle/wrapper/dists/gradle-5.2.1-all/bviwmvmbexq6idcscbicws5me/gradle-5.2.1/lib/gradle-logging-5.2.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.SimpleLoggerFactory]

> Task :sdks:java:extensions:sql:zetasql:checkstyleMain
[main] INFO org.gradle.internal.nativeintegration.services.NativeServices - 
Initialized native services in: /home/jenkins/.gradle/native
> Task :sdks:java:extensions:sql:zetasql:checkstyleTest

> Task :sdks:java:extensions:sql:jdbc:spotbugsMain
SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder".
SLF4J: Defaulting to no-operation (NOP) logger implementation
SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further 
details.

> Task :sdks:java:extensions:sql:datacatalog:javadoc
javadoc: warning - Multiple sources of package comments found for package 
"javax.annotation"
1 warning

> Task :sdks:java:extensions:zetasketch:check
> Task :sdks:java:extensions:zetasketch:build
> Task :sdks:java:extensions:zetasketch:buildDependents
[main] INFO org.gradle.internal.nativeintegration.services.NativeServices - 
Initialized native services in: /home/jenkins/.gradle/native

> Task :sdks:java:extensions:sql:datacatalog:spotbugsMain
SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder".
SLF4J: Defaulting to no-operation (NOP) logger implementation
SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further 
details.

> Task :sdks:java:extensions:sql:zetasql:javadoc
> Task :sdks:java:io:cassandra:test
> Task :sdks:java:io:cassandra:check
> Task :sdks:java:io:cassandra:build
> Task :sdks:java:io:cassandra:buildDependents

> Task :sdks:java:extensions:sql:zetasql:spotbugsMain
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in 
[jar:file:/home/jenkins/.gradle/caches/modules-2/files-2.1/org.slf4j/slf4j-simple/1.7.25/8dacf9514f0c707cbbcdd6fd699e8940d42fb54e/slf4j-simple-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in 
[jar:file:/home/jenkins/.gradle/wrapper/dists/gradle-5.2.1-all/bviwmvmbexq6idcscbicws5me/gradle-5.2.1/lib/gradle-logging-5.2.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.SimpleLoggerFactory]

[main] INFO org.gradle.internal.nativeintegration.services.NativeServices - 
Initialized native services in: /home/jenkins/.gradle/native

> Task :sdks:java:extensions:sql:zetasql:spotbugsMain
SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder".
SLF4J: Defaulting to no-operation (NOP) logger implementation
SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further 
details.

> Task :sdks:java:io:amqp:test
> Task :sdks:java:extensions:sql:datacatalog:test
> Task :sdks:java:extensions:sql:datacatalog:check
> Task :sdks:java:extensions:sql:datacatalog:build
> Task :sdks:java:extensions:sql:datacatalog:buildDependents
> Task :runners:google-cloud-dataflow-java:buildDependents
> Task :sdks:java:io:amazon-web-services:check
> Task :sdks:java:io:amazon-web-services:build
> Task :sdks:java:io:amazon-web-services:buildDependents
> Task :runners:direct-java:test
> Task :sdks:java:extensions:euphoria:check
> Task :sdks:java:extensions:euphoria:build
> Task :sdks:java:extensions:euphoria:buildDependents
> Task :sdks:java:extensions:kryo:buildDependents
> Task :sdks:java:io:clickhouse:test
> Task :sdks:java:extensions:sorter:check
> Task :sdks:java:extensions:sorter:build
> Task 

Jenkins build is back to normal : beam_PostCommit_Py_VR_Dataflow #4999

2019-11-05 Thread Apache Jenkins Server
See 



-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org



Jenkins build is back to normal : beam_PerformanceTests_WordCountIT_Py36 #668

2019-11-05 Thread Apache Jenkins Server
See 



-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org



Build failed in Jenkins: beam_PostCommit_Python35 #906

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[lukasz.gajowy] [BEAM-8432] Enable building kotlin examples with different java 
versions


--
[...truncated 129.69 KB...]
  "is_wrapper": true
},
"output_name": "out",
"user_name": "ReadFromPubSub/Read.out"
  }
],
"pubsub_subscription": 
"projects/apache-beam-testing/subscriptions/exercise_streaming_metrics_subscription_inputfd07c16c-b805-45fd-ab9d-075715f3c816",
"user_name": "ReadFromPubSub/Read"
  }
},
{
  "kind": "ParallelDo",
  "name": "s2",
  "properties": {
"display_data": [
  {
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.ParDo",
"shortValue": "StreamingUserMetricsDoFn",
"type": "STRING",
"value": 
"apache_beam.runners.dataflow.dataflow_exercise_streaming_metrics_pipeline.StreamingUserMetricsDoFn"
  }
],
"non_parallel_inputs": {},
"output_info": [
  {
"encoding": {
  "@type": "kind:windowed_value",
  "component_encodings": [
{
  "@type": "kind:bytes"
},
{
  "@type": "kind:global_window"
}
  ],
  "is_wrapper": true
},
"output_name": "out",
"user_name": "generate_metrics.out"
  }
],
"parallel_input": {
  "@type": "OutputReference",
  "output_name": "out",
  "step_name": "s1"
},
"serialized_fn": "ref_AppliedPTransform_generate_metrics_4",
"user_name": "generate_metrics"
  }
},
{
  "kind": "ParallelWrite",
  "name": "s3",
  "properties": {
"display_data": [],
"encoding": {
  "@type": "kind:windowed_value",
  "component_encodings": [
{
  "@type": "kind:bytes"
},
{
  "@type": "kind:global_window"
}
  ],
  "is_wrapper": true
},
"format": "pubsub",
"parallel_input": {
  "@type": "OutputReference",
  "output_name": "out",
  "step_name": "s2"
},
"pubsub_topic": 
"projects/apache-beam-testing/topics/exercise_streaming_metrics_topic_outputfd07c16c-b805-45fd-ab9d-075715f3c816",
"user_name": "dump_to_pub/Write/NativeWrite"
  }
}
  ],
  "type": "JOB_TYPE_STREAMING"
}
root: INFO: Create job: 
root: INFO: Created job with id: [2019-11-05_01_11_46-4036322759951075359]
root: INFO: To access the Dataflow monitoring console, please navigate to 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_01_11_46-4036322759951075359?project=apache-beam-testing
root: INFO: Job 2019-11-05_01_11_46-4036322759951075359 is in state 
JOB_STATE_RUNNING
root: INFO: 2019-11-05T09:11:50.383Z: JOB_MESSAGE_DETAILED: Checking 
permissions granted to controller Service Account.
root: INFO: 2019-11-05T09:11:51.474Z: JOB_MESSAGE_BASIC: Worker configuration: 
n1-standard-4 in us-central1-f.
root: INFO: 2019-11-05T09:11:52.127Z: JOB_MESSAGE_DETAILED: Expanding 
SplittableParDo operations into optimizable parts.
root: INFO: 2019-11-05T09:11:52.130Z: JOB_MESSAGE_DETAILED: Expanding 
CollectionToSingleton operations into optimizable parts.
root: INFO: 2019-11-05T09:11:52.140Z: JOB_MESSAGE_DETAILED: Expanding 
CoGroupByKey operations into optimizable parts.
root: INFO: 2019-11-05T09:11:52.151Z: JOB_MESSAGE_DETAILED: Expanding 
SplittableProcessKeyed operations into optimizable parts.
root: INFO: 2019-11-05T09:11:52.154Z: JOB_MESSAGE_DETAILED: Expanding 
GroupByKey operations into streaming Read/Write steps
root: INFO: 2019-11-05T09:11:52.158Z: JOB_MESSAGE_DEBUG: Annotating graph with 
Autotuner information.
root: INFO: 2019-11-05T09:11:52.177Z: JOB_MESSAGE_DETAILED: Fusing adjacent 
ParDo, Read, Write, and Flatten operations
root: INFO: 2019-11-05T09:11:52.179Z: JOB_MESSAGE_DETAILED: Fusing consumer 
generate_metrics into ReadFromPubSub/Read
root: INFO: 2019-11-05T09:11:52.182Z: JOB_MESSAGE_DETAILED: Fusing consumer 
dump_to_pub/Write/NativeWrite into generate_metrics
root: INFO: 2019-11-05T09:11:52.193Z: JOB_MESSAGE_DEBUG: Adding StepResource 
setup and teardown to workflow graph.
root: INFO: 2019-11-05T09:11:52.207Z: JOB_MESSAGE_DEBUG: Adding workflow start 
and stop steps.
root: INFO: 2019-11-05T09:11:52.232Z: JOB_MESSAGE_DEBUG: Assigning stage ids.
root: INFO: 2019-11-05T09:11:52.522Z: JOB_MESSAGE_DEBUG: Executing wait step 
start2
root: INFO: 2019-11-05T09:11:52.593Z: JOB_MESSAGE_DEBUG: Starting worker pool 
setup.
root: INFO: 2019-11-05T09:11:52.606Z: JOB_MESSAGE_BASIC: Starting 1 workers...
root: INFO: 

beam_PostCommit_Java_PVR_Spark_Batch - Build # 1206 - Aborted

2019-11-05 Thread Apache Jenkins Server
The Apache Jenkins build system has built beam_PostCommit_Java_PVR_Spark_Batch 
(build #1206)

Status: Aborted

Check console output at 
https://builds.apache.org/job/beam_PostCommit_Java_PVR_Spark_Batch/1206/ to 
view the results.

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org

beam_PostCommit_Python37 - Build # 865 - Aborted

2019-11-05 Thread Apache Jenkins Server
The Apache Jenkins build system has built beam_PostCommit_Python37 (build #865)

Status: Aborted

Check console output at 
https://builds.apache.org/job/beam_PostCommit_Python37/865/ to view the results.

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org

Build failed in Jenkins: beam_PostCommit_XVR_Flink #846

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[mxm] [BEAM-8549] Do not use keyed operator state for checkpoint buffering


--
[...truncated 4.27 MB...]
[flink-akka.actor.default-dispatcher-3] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (15/16) (62bdd58f89be47bed7e609eb4b72cff2) switched from 
RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - MapPartition (MapPartition at 
[3]assert_that/{Group, Unkey, Match}) (2/16) (04aadff016569262d6e25cb500bb6170) 
switched from RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - Freeing task resources for 
MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/16) 
(04aadff016569262d6e25cb500bb6170).
[flink-akka.actor.default-dispatcher-3] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (2/16) (bd52d1911cb667897dc9e0c5a1e8ae5a) switched from 
CREATED to SCHEDULED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem 
streams are closed for task MapPartition (MapPartition at 
[3]assert_that/{Group, Unkey, Match}) (2/16) (04aadff016569262d6e25cb500bb6170) 
[FINISHED]
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task MapPartition 
(MapPartition at [3]assert_that/{Group, Unkey, Match}) 
04aadff016569262d6e25cb500bb6170.
[flink-akka.actor.default-dispatcher-3] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (2/16) (bd52d1911cb667897dc9e0c5a1e8ae5a) switched from 
SCHEDULED to DEPLOYING.
[flink-akka.actor.default-dispatcher-3] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - Deploying DataSink 
(DiscardingOutput) (2/16) (attempt #0) to fa9a5842-0b26-4188-9f65-3d21b8d9b4fa 
@ localhost (dataPort=-1)
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Received task DataSink 
(DiscardingOutput) (2/16).
[flink-akka.actor.default-dispatcher-3] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - MapPartition 
(MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/16) 
(04aadff016569262d6e25cb500bb6170) switched from RUNNING to FINISHED.
[DataSink (DiscardingOutput) (2/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (2/16) 
(bd52d1911cb667897dc9e0c5a1e8ae5a) switched from CREATED to DEPLOYING.
[DataSink (DiscardingOutput) (2/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Creating FileSystem stream leak 
safety net for task DataSink (DiscardingOutput) (2/16) 
(bd52d1911cb667897dc9e0c5a1e8ae5a) [DEPLOYING]
[DataSink (DiscardingOutput) (2/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Loading JAR files for task DataSink 
(DiscardingOutput) (2/16) (bd52d1911cb667897dc9e0c5a1e8ae5a) [DEPLOYING].
[DataSink (DiscardingOutput) (2/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Registering task at network: 
DataSink (DiscardingOutput) (2/16) (bd52d1911cb667897dc9e0c5a1e8ae5a) 
[DEPLOYING].
[DataSink (DiscardingOutput) (2/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (2/16) 
(bd52d1911cb667897dc9e0c5a1e8ae5a) switched from DEPLOYING to RUNNING.
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (2/16) (bd52d1911cb667897dc9e0c5a1e8ae5a) switched from 
DEPLOYING to RUNNING.
[DataSink (DiscardingOutput) (2/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (2/16) 
(bd52d1911cb667897dc9e0c5a1e8ae5a) switched from RUNNING to FINISHED.
[DataSink (DiscardingOutput) (2/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Freeing task resources for DataSink 
(DiscardingOutput) (2/16) (bd52d1911cb667897dc9e0c5a1e8ae5a).
[DataSink (DiscardingOutput) (2/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem streams are 
closed for task DataSink (DiscardingOutput) (2/16) 
(bd52d1911cb667897dc9e0c5a1e8ae5a) [FINISHED]
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task DataSink 
(DiscardingOutput) bd52d1911cb667897dc9e0c5a1e8ae5a.
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (2/16) (bd52d1911cb667897dc9e0c5a1e8ae5a) switched from 
RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, 

Jenkins build is back to normal : beam_PreCommit_Portable_Python_Cron #1351

2019-11-05 Thread Apache Jenkins Server
See 



-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org



Jenkins build is back to normal : beam_PostCommit_Python35 #907

2019-11-05 Thread Apache Jenkins Server
See 



-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org



Build failed in Jenkins: beam_PostCommit_Python2 #902

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[mxm] [BEAM-8549] Do not use keyed operator state for checkpoint buffering


--
[...truncated 964.87 KB...]
[flink-akka.actor.default-dispatcher-3] INFO 
org.apache.flink.runtime.taskexecutor.JobLeaderService - Cannot reconnect to 
job 687c79aef9cd946611892f7abaa410cf because it is not registered.
[flink-akka.actor.default-dispatcher-3] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Stopping TaskExecutor 
akka://flink/user/taskmanager_0.
[flink-akka.actor.default-dispatcher-3] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Close ResourceManager 
connection 5cc316ac652f0006a88db3c0da3169c8.
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.resourcemanager.StandaloneResourceManager - Closing 
TaskExecutor connection d145c248-f765-42e3-be45-50ed3792a88d because: The 
TaskExecutor is shutting down.
[flink-akka.actor.default-dispatcher-3] INFO 
org.apache.flink.runtime.taskexecutor.JobLeaderService - Stop job leader 
service.
[flink-akka.actor.default-dispatcher-3] INFO 
org.apache.flink.runtime.state.TaskExecutorLocalStateStoresManager - Shutting 
down TaskExecutorLocalStateStoresManager.
[flink-akka.actor.default-dispatcher-3] INFO 
org.apache.flink.runtime.io.disk.FileChannelManagerImpl - FileChannelManager 
removed spill file directory /tmp/flink-io-edf29f84-0907-4625-9390-b8e80fed7d61
[flink-akka.actor.default-dispatcher-3] INFO 
org.apache.flink.runtime.io.network.NettyShuffleEnvironment - Shutting down the 
network environment and its components.
[ForkJoinPool.commonPool-worker-2] INFO 
org.apache.flink.runtime.dispatcher.DispatcherRestEndpoint - Removing cache 
directory /tmp/flink-web-ui
[ForkJoinPool.commonPool-worker-2] INFO 
org.apache.flink.runtime.dispatcher.DispatcherRestEndpoint - Shut down complete.
[flink-akka.actor.default-dispatcher-7] INFO 
org.apache.flink.runtime.resourcemanager.StandaloneResourceManager - Shut down 
cluster because application is in CANCELED, diagnostics 
DispatcherResourceManagerComponent has been closed..
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.dispatcher.StandaloneDispatcher - Stopping dispatcher 
akka://flink/user/dispatcher.
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.dispatcher.StandaloneDispatcher - Stopping all 
currently running jobs of dispatcher akka://flink/user/dispatcher.
[flink-akka.actor.default-dispatcher-8] INFO 
org.apache.flink.runtime.resourcemanager.slotmanager.SlotManagerImpl - Closing 
the SlotManager.
[flink-akka.actor.default-dispatcher-8] INFO 
org.apache.flink.runtime.resourcemanager.slotmanager.SlotManagerImpl - 
Suspending the SlotManager.
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.rest.handler.legacy.backpressure.StackTraceSampleCoordinator
 - Shutting down stack trace sample coordinator.
[flink-akka.actor.default-dispatcher-3] INFO 
org.apache.flink.runtime.io.disk.FileChannelManagerImpl - FileChannelManager 
removed spill file directory 
/tmp/flink-netty-shuffle-8c35ece5-b656-44cd-ba09-46fb88d10ec7
[flink-akka.actor.default-dispatcher-3] INFO 
org.apache.flink.runtime.taskexecutor.KvStateService - Shutting down the 
kvState service and its components.
[flink-akka.actor.default-dispatcher-3] INFO 
org.apache.flink.runtime.taskexecutor.JobLeaderService - Stop job leader 
service.
[flink-akka.actor.default-dispatcher-3] INFO 
org.apache.flink.runtime.filecache.FileCache - removed file cache directory 
/tmp/flink-dist-cache-fa1afd80-bf4e-4991-848b-0d5a51e84b33
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.dispatcher.StandaloneDispatcher - Stopped dispatcher 
akka://flink/user/dispatcher.
[flink-akka.actor.default-dispatcher-3] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Stopped TaskExecutor 
akka://flink/user/taskmanager_0.
[flink-akka.actor.default-dispatcher-3] INFO 
org.apache.flink.runtime.rpc.akka.AkkaRpcService - Stopping Akka RPC service.
[flink-metrics-2] INFO akka.remote.RemoteActorRefProvider$RemotingTerminator - 
Shutting down remote daemon.
[flink-metrics-2] INFO akka.remote.RemoteActorRefProvider$RemotingTerminator - 
Remote daemon shut down; proceeding with flushing remote transports.
[flink-metrics-2] INFO akka.remote.RemoteActorRefProvider$RemotingTerminator - 
Remoting shut down.
[flink-metrics-2] INFO org.apache.flink.runtime.rpc.akka.AkkaRpcService - 
Stopping Akka RPC service.
[flink-metrics-2] INFO org.apache.flink.runtime.rpc.akka.AkkaRpcService - 
Stopped Akka RPC service.
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.blob.PermanentBlobCache - Shutting down BLOB cache
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.blob.TransientBlobCache - Shutting down BLOB cache
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.blob.BlobServer - Stopped 

beam_PostCommit_Python37 - Build # 866 - Aborted

2019-11-05 Thread Apache Jenkins Server
The Apache Jenkins build system has built beam_PostCommit_Python37 (build #866)

Status: Aborted

Check console output at 
https://builds.apache.org/job/beam_PostCommit_Python37/866/ to view the results.

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org

beam_sonarqube_report - Build # 996 - Aborted

2019-11-05 Thread Apache Jenkins Server
The Apache Jenkins build system has built beam_sonarqube_report (build #996)

Status: Aborted

Check console output at 
https://builds.apache.org/job/beam_sonarqube_report/996/ to view the results.

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org

beam_PostCommit_Python37 - Build # 864 - Aborted

2019-11-05 Thread Apache Jenkins Server
The Apache Jenkins build system has built beam_PostCommit_Python37 (build #864)

Status: Aborted

Check console output at 
https://builds.apache.org/job/beam_PostCommit_Python37/864/ to view the results.

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org

Jenkins build is back to normal : beam_PostCommit_Python36 #909

2019-11-05 Thread Apache Jenkins Server
See 



-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org



Build failed in Jenkins: beam_PostCommit_Py_VR_Dataflow #4997

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[aryan.naraghi] Fix a bug related to zero-row responses


--
[...truncated 219.93 KB...]
(unset)

> Task :sdks:python:test-suites:dataflow:py37:validatesRunnerStreamingTests
(unset)
(unset)

> Task :sdks:python:test-suites:dataflow:py36:validatesRunnerStreamingTests
(unset)
(unset)
(unset)

> Task :sdks:python:test-suites:dataflow:py37:validatesRunnerStreamingTests
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)

> Task :sdks:python:test-suites:dataflow:py36:validatesRunnerStreamingTests
(unset)
(unset)
(unset)
(unset)

> Task :sdks:python:test-suites:dataflow:py35:validatesRunnerStreamingTests
(unset)
(unset)
(unset)
(unset)
(unset)
(unset)

> Task :sdks:python:test-suites:dataflow:py2:validatesRunnerStreamingTests
test_as_list_and_as_dict_side_inputs 
(apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
(unset)
test_as_dict_twice (apache_beam.transforms.sideinputs_test.SideInputsTest) ... 
ok
test_default_value_singleton_side_input 
(apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_as_singleton_with_different_defaults 
(apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok

> Task :sdks:python:test-suites:dataflow:py35:validatesRunnerStreamingTests
(unset)
(unset)

> Task :sdks:python:test-suites:dataflow:py2:validatesRunnerStreamingTests
test_as_list_twice (apache_beam.transforms.sideinputs_test.SideInputsTest) ... 
ok
test_as_singleton_without_unique_labels 
(apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_empty_singleton_side_input 
(apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_flattened_side_input 
(apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok

> Task :sdks:python:test-suites:dataflow:py37:validatesRunnerStreamingTests
(unset)

> Task :sdks:python:test-suites:dataflow:py36:validatesRunnerStreamingTests
(unset)

> Task :sdks:python:test-suites:dataflow:py35:validatesRunnerStreamingTests
(unset)

> Task :sdks:python:test-suites:dataflow:py2:validatesRunnerStreamingTests
test_iterable_side_input 
(apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok

--
XML: nosetests-validatesRunnerStreamingTests-df.xml
--
XML: 

--
Ran 17 tests in 1306.126s

OK
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-04_23_45_53-18320771905635718723?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-04_23_52_46-1411167559952717?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-05_00_00_04-4853874799857115208?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-04_23_45_55-15303580306201900460?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-04_23_52_54-1845751760414228930?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-04_23_45_51-15864219214209614471?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-04_23_53_04-4758348086310889050?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-04_23_45_53-12377409482151374542?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-04_23_53_06-2285273010050723309?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-04_23_45_52-10227474540064498261?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-04_23_53_36-7801123343357370662?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-04_23_45_53-1791593666816297783?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-04_23_53_12-14481775608919920660?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-04_23_45_53-11587219118972271270?project=apache-beam-testing
Worker logs: 

beam_PostCommit_Java_PVR_Spark_Batch - Build # 1205 - Aborted

2019-11-05 Thread Apache Jenkins Server
The Apache Jenkins build system has built beam_PostCommit_Java_PVR_Spark_Batch 
(build #1205)

Status: Aborted

Check console output at 
https://builds.apache.org/job/beam_PostCommit_Java_PVR_Spark_Batch/1205/ to 
view the results.

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org

Build failed in Jenkins: beam_PostCommit_Py_VR_Dataflow #4998

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[lukasz.gajowy] [BEAM-8432] Enable building kotlin examples with different java 
versions


--
[...truncated 227.25 KB...]
"location": 
"storage.googleapis.com/temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1105085733-853834.1572944253.854138/pickled_main_session",
"name": "pickled_main_session"
  },
  {
"location": 
"storage.googleapis.com/temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1105085733-853834.1572944253.854138/dataflow_python_sdk.tar",
"name": "dataflow_python_sdk.tar"
  },
  {
"location": 
"storage.googleapis.com/temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1105085733-853834.1572944253.854138/dataflow-worker.jar",
"name": "dataflow-worker.jar"
  }
],
"taskrunnerSettings": {
  "parallelWorkerSettings": {
"baseUrl": "https://dataflow.googleapis.com;,
"servicePath": "https://dataflow.googleapis.com;
  }
},
"workerHarnessContainerImage": 
"gcr.io/cloud-dataflow/v1beta3/python3-fnapi:beam-master-20191029"
  }
]
  },
  "name": "beamapp-jenkins-1105085733-853834",
  "steps": [
{
  "kind": "ParallelRead",
  "name": "s1",
  "properties": {
"display_data": [
  {
"key": "with_attributes",
"label": "With Attributes",
"namespace": "apache_beam.io.gcp.pubsub._PubSubSource",
"type": "BOOLEAN",
"value": false
  },
  {
"key": "subscription",
"label": "Pubsub Subscription",
"namespace": "apache_beam.io.gcp.pubsub._PubSubSource",
"type": "STRING",
"value": 
"projects/apache-beam-testing/subscriptions/exercise_streaming_metrics_subscription_inputf803e25d-2bcd-4c37-967e-9e966095929f"
  },
  {
"key": "source",
"label": "Read Source",
"namespace": "apache_beam.io.iobase.Read",
"shortValue": "_PubSubSource",
"type": "STRING",
"value": "apache_beam.io.gcp.pubsub._PubSubSource"
  }
],
"format": "pubsub",
"output_info": [
  {
"encoding": {
  "@type": "kind:windowed_value",
  "component_encodings": [
{
  "@type": "kind:bytes"
},
{
  "@type": "kind:global_window"
}
  ],
  "is_wrapper": true
},
"output_name": "out",
"user_name": "ReadFromPubSub/Read.out"
  }
],
"pubsub_subscription": 
"projects/apache-beam-testing/subscriptions/exercise_streaming_metrics_subscription_inputf803e25d-2bcd-4c37-967e-9e966095929f",
"user_name": "ReadFromPubSub/Read"
  }
},
{
  "kind": "ParallelDo",
  "name": "s2",
  "properties": {
"display_data": [
  {
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.ParDo",
"shortValue": "StreamingUserMetricsDoFn",
"type": "STRING",
"value": 
"apache_beam.runners.dataflow.dataflow_exercise_streaming_metrics_pipeline.StreamingUserMetricsDoFn"
  }
],
"non_parallel_inputs": {},
"output_info": [
  {
"encoding": {
  "@type": "kind:windowed_value",
  "component_encodings": [
{
  "@type": "kind:bytes"
},
{
  "@type": "kind:global_window"
}
  ],
  "is_wrapper": true
},
"output_name": "out",
"user_name": "generate_metrics.out"
  }
],
"parallel_input": {
  "@type": "OutputReference",
  "output_name": "out",
  "step_name": "s1"
},
"serialized_fn": "ref_AppliedPTransform_generate_metrics_4",
"user_name": "generate_metrics"
  }
},
{
  "kind": "ParallelWrite",
  "name": "s3",
  "properties": {
"display_data": [],
"encoding": {
  "@type": "kind:windowed_value",
  "component_encodings": [
{
  "@type": "kind:bytes"
},
{
  "@type": "kind:global_window"
}
  ],
  "is_wrapper": true
},
"format": "pubsub",
"parallel_input": {
  "@type": "OutputReference",
  "output_name": "out",
  "step_name": "s2"
},
"pubsub_topic": 

beam_PostCommit_Java_ValidatesRunner_PortabilityApi_Dataflow - Build # 2376 - Aborted

2019-11-05 Thread Apache Jenkins Server
The Apache Jenkins build system has built 
beam_PostCommit_Java_ValidatesRunner_PortabilityApi_Dataflow (build #2376)

Status: Aborted

Check console output at 
https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_PortabilityApi_Dataflow/2376/
 to view the results.

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org

Build failed in Jenkins: beam_PostCommit_Python35 #905

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[aryan.naraghi] Fix a bug related to zero-row responses


--
[...truncated 334.81 KB...]
 fields: ['language']>
root: INFO: Matching ['language'] to ['language']
root: INFO: Attempting to perform query SELECT name, language FROM 
python_bq_streaming_inserts_15729397952671.output_table1 to BQ
google.auth.transport._http_client: DEBUG: Making request: GET 
http://169.254.169.254
google.auth.transport._http_client: DEBUG: Making request: GET 
http://metadata.google.internal/computeMetadata/v1/project/project-id
urllib3.util.retry: DEBUG: Converted retries value: 3 -> Retry(total=3, 
connect=None, read=None, redirect=None, status=None)
google.auth.transport.requests: DEBUG: Making request: GET 
http://metadata.google.internal/computeMetadata/v1/instance/service-accounts/default/?recursive=true
urllib3.connectionpool: DEBUG: Starting new HTTP connection (1): 
metadata.google.internal:80
urllib3.connectionpool: DEBUG: http://metadata.google.internal:80 "GET 
/computeMetadata/v1/instance/service-accounts/default/?recursive=true HTTP/1.1" 
200 144
google.auth.transport.requests: DEBUG: Making request: GET 
http://metadata.google.internal/computeMetadata/v1/instance/service-accounts/844138762903-comp...@developer.gserviceaccount.com/token
urllib3.connectionpool: DEBUG: http://metadata.google.internal:80 "GET 
/computeMetadata/v1/instance/service-accounts/844138762903-comp...@developer.gserviceaccount.com/token
 HTTP/1.1" 200 181
urllib3.connectionpool: DEBUG: Starting new HTTPS connection (1): 
www.googleapis.com:443
urllib3.connectionpool: DEBUG: https://www.googleapis.com:443 "POST 
/bigquery/v2/projects/apache-beam-testing/jobs HTTP/1.1" 200 None
urllib3.connectionpool: DEBUG: https://www.googleapis.com:443 "GET 
/bigquery/v2/projects/apache-beam-testing/queries/454d528e-c27c-4884-9713-ff8ef4ac0da2?maxResults=0=1=US
 HTTP/1.1" 200 None
urllib3.connectionpool: DEBUG: https://www.googleapis.com:443 "GET 
/bigquery/v2/projects/apache-beam-testing/jobs/454d528e-c27c-4884-9713-ff8ef4ac0da2?location=US
 HTTP/1.1" 200 None
urllib3.connectionpool: DEBUG: https://www.googleapis.com:443 "GET 
/bigquery/v2/projects/apache-beam-testing/datasets/_7357fab0f784d2a7327ddbe81cdd1f4ca7e429cd/tables/anon3dee4dc6_aabb_4a7d_9f70_9a8b5ed8211d/data
 HTTP/1.1" 200 None
root: INFO: Result of query is: [('spark', 'scala'), ('beam', 'go'), ('flink', 
'scala'), ('spark', 'py'), ('beam', 'java'), ('beam', 'py'), ('flink', 'java'), 
('spark', 'scala')]
root: INFO: Attempting to perform query SELECT name, language FROM 
python_bq_streaming_inserts_15729397952671.output_table2 to BQ
google.auth.transport._http_client: DEBUG: Making request: GET 
http://169.254.169.254
google.auth.transport._http_client: DEBUG: Making request: GET 
http://metadata.google.internal/computeMetadata/v1/project/project-id
urllib3.util.retry: DEBUG: Converted retries value: 3 -> Retry(total=3, 
connect=None, read=None, redirect=None, status=None)
google.auth.transport.requests: DEBUG: Making request: GET 
http://metadata.google.internal/computeMetadata/v1/instance/service-accounts/default/?recursive=true
urllib3.connectionpool: DEBUG: Starting new HTTP connection (1): 
metadata.google.internal:80
urllib3.connectionpool: DEBUG: http://metadata.google.internal:80 "GET 
/computeMetadata/v1/instance/service-accounts/default/?recursive=true HTTP/1.1" 
200 144
google.auth.transport.requests: DEBUG: Making request: GET 
http://metadata.google.internal/computeMetadata/v1/instance/service-accounts/844138762903-comp...@developer.gserviceaccount.com/token
urllib3.connectionpool: DEBUG: http://metadata.google.internal:80 "GET 
/computeMetadata/v1/instance/service-accounts/844138762903-comp...@developer.gserviceaccount.com/token
 HTTP/1.1" 200 181
urllib3.connectionpool: DEBUG: Starting new HTTPS connection (1): 
www.googleapis.com:443
urllib3.connectionpool: DEBUG: https://www.googleapis.com:443 "POST 
/bigquery/v2/projects/apache-beam-testing/jobs HTTP/1.1" 200 None
urllib3.connectionpool: DEBUG: https://www.googleapis.com:443 "GET 
/bigquery/v2/projects/apache-beam-testing/queries/21147188-5b59-40b3-8832-a724fa11f4fe?maxResults=0=1=US
 HTTP/1.1" 200 None
urllib3.connectionpool: DEBUG: https://www.googleapis.com:443 "GET 
/bigquery/v2/projects/apache-beam-testing/jobs/21147188-5b59-40b3-8832-a724fa11f4fe?location=US
 HTTP/1.1" 200 None
urllib3.connectionpool: DEBUG: https://www.googleapis.com:443 "GET 
/bigquery/v2/projects/apache-beam-testing/datasets/_7357fab0f784d2a7327ddbe81cdd1f4ca7e429cd/tables/anon8a9cd30d19a2585a5823f151c0f6c12a2a58ac09/data
 HTTP/1.1" 200 None
root: INFO: Result of query is: [('beam', 'go'), ('beam', 'py'), ('spark', 
'py'), ('beam', 'java'), ('flink', 'java'), ('flink', 'scala'), ('spark', 
'scala'), ('spark', 'scala'), ('beam', 'go'), ('beam', 'py'), ('spark', 'py'), 
('beam', 'java'), ('flink', 

Build failed in Jenkins: beam_PostCommit_XVR_Flink #845

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[lukasz.gajowy] [BEAM-8432] Enable building kotlin examples with different java 
versions


--
[...truncated 4.28 MB...]
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task DataSink 
(DiscardingOutput) a1c2e527b705e9d5dc74a730effad9f9.
[flink-akka.actor.default-dispatcher-5] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (9/16) (a1c2e527b705e9d5dc74a730effad9f9) switched from 
RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - MapPartition (MapPartition at 
[3]assert_that/{Group, Unkey, Match}) (2/16) (a6cfa18c7a10970376c52ac4d3343552) 
switched from RUNNING to FINISHED.
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (2/16) (ad60fbd413a655ca0a308fd098318e87) switched from 
CREATED to SCHEDULED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - Freeing task resources for 
MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/16) 
(a6cfa18c7a10970376c52ac4d3343552).
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem 
streams are closed for task MapPartition (MapPartition at 
[3]assert_that/{Group, Unkey, Match}) (2/16) (a6cfa18c7a10970376c52ac4d3343552) 
[FINISHED]
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (2/16) (ad60fbd413a655ca0a308fd098318e87) switched from 
SCHEDULED to DEPLOYING.
[flink-akka.actor.default-dispatcher-5] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task MapPartition 
(MapPartition at [3]assert_that/{Group, Unkey, Match}) 
a6cfa18c7a10970376c52ac4d3343552.
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - Deploying DataSink 
(DiscardingOutput) (2/16) (attempt #0) to 60168c57-0d28-4cb1-92a6-7a6d3ad17357 
@ localhost (dataPort=-1)
[flink-akka.actor.default-dispatcher-6] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - MapPartition 
(MapPartition at [3]assert_that/{Group, Unkey, Match}) (2/16) 
(a6cfa18c7a10970376c52ac4d3343552) switched from RUNNING to FINISHED.
[flink-akka.actor.default-dispatcher-5] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Received task DataSink 
(DiscardingOutput) (2/16).
[DataSink (DiscardingOutput) (2/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (2/16) 
(ad60fbd413a655ca0a308fd098318e87) switched from CREATED to DEPLOYING.
[DataSink (DiscardingOutput) (2/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Creating FileSystem stream leak 
safety net for task DataSink (DiscardingOutput) (2/16) 
(ad60fbd413a655ca0a308fd098318e87) [DEPLOYING]
[DataSink (DiscardingOutput) (2/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Loading JAR files for task DataSink 
(DiscardingOutput) (2/16) (ad60fbd413a655ca0a308fd098318e87) [DEPLOYING].
[DataSink (DiscardingOutput) (2/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Registering task at network: 
DataSink (DiscardingOutput) (2/16) (ad60fbd413a655ca0a308fd098318e87) 
[DEPLOYING].
[DataSink (DiscardingOutput) (2/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (2/16) 
(ad60fbd413a655ca0a308fd098318e87) switched from DEPLOYING to RUNNING.
[flink-akka.actor.default-dispatcher-5] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (2/16) (ad60fbd413a655ca0a308fd098318e87) switched from 
DEPLOYING to RUNNING.
[DataSink (DiscardingOutput) (2/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (2/16) 
(ad60fbd413a655ca0a308fd098318e87) switched from RUNNING to FINISHED.
[DataSink (DiscardingOutput) (2/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Freeing task resources for DataSink 
(DiscardingOutput) (2/16) (ad60fbd413a655ca0a308fd098318e87).
[DataSink (DiscardingOutput) (2/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem streams are 
closed for task DataSink (DiscardingOutput) (2/16) 
(ad60fbd413a655ca0a308fd098318e87) [FINISHED]
[flink-akka.actor.default-dispatcher-5] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task DataSink 
(DiscardingOutput) ad60fbd413a655ca0a308fd098318e87.

Build failed in Jenkins: beam_PostCommit_Java_Nexmark_Flink #4217

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[mxm] Use better name for Flink current key supplier

[lgajowy] [BEAM-8558] Fix BigQueryIOIT flakiness (#9995)


--
[...truncated 10.41 KB...]
> Task :model:fn-execution:extractProto
> Task :model:job-management:extractProto
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:extractProto
> Task :sdks:java:extensions:join-library:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :sdks:java:io:parquet:processResources NO-SOURCE
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :model:job-management:processResources
> Task :sdks:java:testing:nexmark:processResources
> Task :model:fn-execution:processResources
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :runners:flink:1.9:copySourceOverrides
> Task :runners:flink:1.9:copyTestResourcesOverrides NO-SOURCE
> Task :runners:flink:1.9:processResources
> Task :sdks:java:build-tools:compileJava FROM-CACHE
> Task :sdks:java:build-tools:processResources
> Task :sdks:java:build-tools:classes
> Task :sdks:java:core:processResources
> Task :sdks:java:build-tools:jar
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :model:pipeline:generateProto
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromCalciteCore
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromSrc
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :sdks:java:extensions:sql:generateFmppSources
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:extractIncludeProto
> Task :model:job-management:generateProto
> Task :model:fn-execution:generateProto
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes

> Task :sdks:java:extensions:sql:compileJavacc
Java Compiler Compiler Version 4.0 (Parser Generator)
(type "javacc" with no arguments for help)
Reading from file 

 . . .
Note: UNICODE_INPUT option is specified. Please make sure you create the 
parser/lexer using a Reader with the correct character encoding.
Warning: Lookahead adequacy checking not being performed since option LOOKAHEAD 
is more than 1.  Set option FORCE_LA_CHECK to true to force checking.
File "TokenMgrError.java" does not exist.  Will create one.
File "ParseException.java" does not exist.  Will create one.
File "Token.java" does not exist.  Will create one.
File "SimpleCharStream.java" does not exist.  Will create one.
Parser generated with 0 errors and 1 warnings.

> Task :sdks:java:extensions:sql:processResources
> Task :model:pipeline:shadowJar
> Task :model:job-management:shadowJar
> Task :model:fn-execution:shadowJar
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :sdks:java:extensions:join-library:compileJava FROM-CACHE
> Task :vendor:sdks-java-extensions-protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:join-library:classes UP-TO-DATE
> Task :vendor:sdks-java-extensions-protobuf:classes UP-TO-DATE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :sdks:java:extensions:join-library:jar
> Task :runners:local-java:jar
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :vendor:sdks-java-extensions-protobuf:shadowJar
> Task :sdks:java:io:kafka:jar
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:core:jar
> Task :sdks:java:fn-execution:jar
> Task :runners:core-construction-java:jar
> Task :sdks:java:io:parquet:compileJava FROM-CACHE
> Task :sdks:java:io:parquet:classes UP-TO-DATE
> Task :sdks:java:io:parquet:jar
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:jar
> Task 

Build failed in Jenkins: beam_PostCommit_Java_Nexmark_Flink #4214

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[mxm] [BEAM-8549] Do not use keyed operator state for checkpoint buffering


--
[...truncated 9.93 KB...]
> Task :runners:direct-java:processResources NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :sdks:java:core:generateAvroJava NO-SOURCE
> Task :model:job-management:extractProto
> Task :model:fn-execution:extractProto
> Task :sdks:java:testing:test-utils:processResources NO-SOURCE
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :sdks:java:io:parquet:processResources NO-SOURCE
> Task :sdks:java:extensions:join-library:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:extractProto
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :model:job-management:processResources
> Task :model:fn-execution:processResources
> Task :runners:flink:1.9:copySourceOverrides
> Task :sdks:java:testing:nexmark:processResources
> Task :runners:flink:1.9:copyTestResourcesOverrides NO-SOURCE
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :runners:flink:1.9:processResources
> Task :sdks:java:build-tools:compileJava FROM-CACHE
> Task :sdks:java:build-tools:processResources
> Task :sdks:java:build-tools:classes
> Task :sdks:java:core:processResources
> Task :sdks:java:build-tools:jar
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromCalciteCore
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromSrc
> Task :model:pipeline:generateProto
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :sdks:java:extensions:sql:generateFmppSources
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:extractIncludeProto
> Task :model:job-management:generateProto
> Task :model:fn-execution:generateProto
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes

> Task :sdks:java:extensions:sql:compileJavacc
Java Compiler Compiler Version 4.0 (Parser Generator)
(type "javacc" with no arguments for help)
Reading from file 

 . . .
Note: UNICODE_INPUT option is specified. Please make sure you create the 
parser/lexer using a Reader with the correct character encoding.
Warning: Lookahead adequacy checking not being performed since option LOOKAHEAD 
is more than 1.  Set option FORCE_LA_CHECK to true to force checking.
File "TokenMgrError.java" does not exist.  Will create one.
File "ParseException.java" does not exist.  Will create one.
File "Token.java" does not exist.  Will create one.
File "SimpleCharStream.java" does not exist.  Will create one.
Parser generated with 0 errors and 1 warnings.

> Task :sdks:java:extensions:sql:processResources
> Task :model:pipeline:shadowJar
> Task :model:job-management:shadowJar
> Task :model:fn-execution:shadowJar
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :sdks:java:core:jar
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :vendor:sdks-java-extensions-protobuf:compileJava FROM-CACHE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :vendor:sdks-java-extensions-protobuf:classes UP-TO-DATE
> Task :runners:local-java:classes UP-TO-DATE
> Task :runners:local-java:jar
> Task :sdks:java:io:kafka:jar
> Task :sdks:java:extensions:join-library:compileJava FROM-CACHE
> Task :sdks:java:extensions:join-library:classes UP-TO-DATE
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :sdks:java:extensions:join-library:jar
> Task :vendor:sdks-java-extensions-protobuf:shadowJar
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :runners:core-construction-java:jar
> Task :sdks:java:fn-execution:jar
> Task :sdks:java:io:parquet:compileJava FROM-CACHE
> Task :sdks:java:io:parquet:classes UP-TO-DATE
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :sdks:java:io:parquet:jar
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :runners:core-java:jar
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task 

Build failed in Jenkins: beam_PreCommit_Python_pytest_Cron #6

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[mxm] [BEAM-8549] Do not use keyed operator state for checkpoint buffering

[lukasz.gajowy] [BEAM-8432] Enable building kotlin examples with different java 
versions

[aryan.naraghi] Fix a bug related to zero-row responses


--
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-13 (beam) in workspace 

No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init 
 >  # 
 > timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git fetch --tags --progress https://github.com/apache/beam.git 
 > +refs/heads/*:refs/remotes/origin/*
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # 
 > timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git 
 > +refs/heads/*:refs/remotes/origin/* 
 > +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/*
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 9f566394102a5a52684311ceea962f1720c16b1a (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 9f566394102a5a52684311ceea962f1720c16b1a
Commit message: "Merge pull request #9980 from mxm/BEAM-8549"
 > git rev-list --no-walk a4141aceadabbfc1cdfdf35a119547e389ab2631 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the 
SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1

[EnvInject] - Variables injected successfully.
[Gradle] - Launching build.
[src] $ 

 --continue --max-workers=12 -Dorg.gradle.jvmargs=-Xms2g 
-Dorg.gradle.jvmargs=-Xmx4g :pythonPreCommitPytest
Starting a Gradle Daemon, 1 busy Daemon could not be reused, use --status for 
details
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy
> Task :buildSrc:spotlessGroovyCheck
> Task :buildSrc:spotlessGroovyGradle
> Task :buildSrc:spotlessGroovyGradleCheck
> Task :buildSrc:spotlessCheck
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validateTaskProperties FROM-CACHE
> Task :buildSrc:check
> Task :buildSrc:build
Configuration on demand is an incubating feature.

FAILURE: Build failed with an exception.

* What went wrong:
Could not determine the dependencies of task 
':sdks:python:test-suites:tox:py35:preCommitPy35Pytest'.
> Task with path 'lint' not found in project 
> ':sdks:python:test-suites:tox:py35'.

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug 
option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with 
Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See 
https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 10s

Publishing build scan...
https://gradle.com/s/yef2vxu2cszfy

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org



Jenkins build is back to normal : beam_BiqQueryIO_Batch_Performance_Test_Java #134

2019-11-05 Thread Apache Jenkins Server
See 



-
To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org
For additional commands, e-mail: builds-h...@beam.apache.org



Build failed in Jenkins: beam_PostCommit_XVR_Flink #849

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[mxm] Use better name for Flink current key supplier

[lgajowy] [BEAM-8558] Fix BigQueryIOIT flakiness (#9995)


--
[...truncated 4.28 MB...]
[flink-akka.actor.default-dispatcher-7] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - MapPartition 
(MapPartition at [3]assert_that/{Group, Unkey, Match}) (7/16) 
(135204ab5b6e1b516d51dcd07d503748) switched from RUNNING to FINISHED.
[flink-akka.actor.default-dispatcher-3] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Received task DataSink 
(DiscardingOutput) (7/16).
[DataSink (DiscardingOutput) (7/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (7/16) 
(2bc83139e462f03f648819a17e35e189) switched from CREATED to DEPLOYING.
[DataSink (DiscardingOutput) (7/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Creating FileSystem stream leak 
safety net for task DataSink (DiscardingOutput) (7/16) 
(2bc83139e462f03f648819a17e35e189) [DEPLOYING]
[DataSink (DiscardingOutput) (7/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Loading JAR files for task DataSink 
(DiscardingOutput) (7/16) (2bc83139e462f03f648819a17e35e189) [DEPLOYING].
[DataSink (DiscardingOutput) (7/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Registering task at network: 
DataSink (DiscardingOutput) (7/16) (2bc83139e462f03f648819a17e35e189) 
[DEPLOYING].
[DataSink (DiscardingOutput) (7/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (7/16) 
(2bc83139e462f03f648819a17e35e189) switched from DEPLOYING to RUNNING.
[flink-akka.actor.default-dispatcher-3] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (7/16) (2bc83139e462f03f648819a17e35e189) switched from 
DEPLOYING to RUNNING.
[DataSink (DiscardingOutput) (7/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (7/16) 
(2bc83139e462f03f648819a17e35e189) switched from RUNNING to FINISHED.
[DataSink (DiscardingOutput) (7/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Freeing task resources for DataSink 
(DiscardingOutput) (7/16) (2bc83139e462f03f648819a17e35e189).
[DataSink (DiscardingOutput) (7/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem streams are 
closed for task DataSink (DiscardingOutput) (7/16) 
(2bc83139e462f03f648819a17e35e189) [FINISHED]
[flink-akka.actor.default-dispatcher-8] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task DataSink 
(DiscardingOutput) 2bc83139e462f03f648819a17e35e189.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (12/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - MapPartition (MapPartition at 
[3]assert_that/{Group, Unkey, Match}) (12/16) 
(5baf44b0f6af68ccd62c3421265103d4) switched from RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (12/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - Freeing task resources for 
MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (12/16) 
(5baf44b0f6af68ccd62c3421265103d4).
[flink-akka.actor.default-dispatcher-3] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (12/16) (27ae9a606e0a140c1e0e7ea086bd6584) switched from 
CREATED to SCHEDULED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (12/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem 
streams are closed for task MapPartition (MapPartition at 
[3]assert_that/{Group, Unkey, Match}) (12/16) 
(5baf44b0f6af68ccd62c3421265103d4) [FINISHED]
[flink-akka.actor.default-dispatcher-8] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task MapPartition 
(MapPartition at [3]assert_that/{Group, Unkey, Match}) 
5baf44b0f6af68ccd62c3421265103d4.
[flink-akka.actor.default-dispatcher-3] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (12/16) (27ae9a606e0a140c1e0e7ea086bd6584) switched from 
SCHEDULED to DEPLOYING.
[flink-akka.actor.default-dispatcher-3] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - Deploying DataSink 
(DiscardingOutput) (12/16) (attempt #0) to 066a5b9f-6f90-45a4-baa8-8fd529aba440 
@ localhost (dataPort=-1)
[flink-akka.actor.default-dispatcher-3] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (7/16) (2bc83139e462f03f648819a17e35e189) switched from 
RUNNING to FINISHED.
[flink-akka.actor.default-dispatcher-8] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Received task DataSink 
(DiscardingOutput) (12/16).
[flink-akka.actor.default-dispatcher-3] INFO 

Build failed in Jenkins: beam_PostCommit_XVR_Flink #847

2019-11-05 Thread Apache Jenkins Server
See 


Changes:

[michal.walenia] [BEAM-7303] Rename runners.reference package to 
runners.portability


--
[...truncated 4.28 MB...]
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task DataSink 
(DiscardingOutput) e19fcec98e38bb04f04f414958910da7.
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (9/16) (e19fcec98e38bb04f04f414958910da7) switched from 
RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (4/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - MapPartition (MapPartition at 
[3]assert_that/{Group, Unkey, Match}) (4/16) (2f3645edeb052402aa39137d3a5b06c8) 
switched from RUNNING to FINISHED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (4/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - Freeing task resources for 
MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (4/16) 
(2f3645edeb052402aa39137d3a5b06c8).
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (4/16) (200278373fbd7c38ef75118d2cd413b1) switched from 
CREATED to SCHEDULED.
[MapPartition (MapPartition at [3]assert_that/{Group, Unkey, Match}) (4/16)] 
INFO org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem 
streams are closed for task MapPartition (MapPartition at 
[3]assert_that/{Group, Unkey, Match}) (4/16) (2f3645edeb052402aa39137d3a5b06c8) 
[FINISHED]
[flink-akka.actor.default-dispatcher-5] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task MapPartition 
(MapPartition at [3]assert_that/{Group, Unkey, Match}) 
2f3645edeb052402aa39137d3a5b06c8.
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (4/16) (200278373fbd7c38ef75118d2cd413b1) switched from 
SCHEDULED to DEPLOYING.
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - Deploying DataSink 
(DiscardingOutput) (4/16) (attempt #0) to f2e88916-aab4-4892-8a76-8ed023f6da77 
@ localhost (dataPort=-1)
[flink-akka.actor.default-dispatcher-5] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Received task DataSink 
(DiscardingOutput) (4/16).
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - MapPartition 
(MapPartition at [3]assert_that/{Group, Unkey, Match}) (4/16) 
(2f3645edeb052402aa39137d3a5b06c8) switched from RUNNING to FINISHED.
[DataSink (DiscardingOutput) (4/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (4/16) 
(200278373fbd7c38ef75118d2cd413b1) switched from CREATED to DEPLOYING.
[DataSink (DiscardingOutput) (4/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Creating FileSystem stream leak 
safety net for task DataSink (DiscardingOutput) (4/16) 
(200278373fbd7c38ef75118d2cd413b1) [DEPLOYING]
[DataSink (DiscardingOutput) (4/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Loading JAR files for task DataSink 
(DiscardingOutput) (4/16) (200278373fbd7c38ef75118d2cd413b1) [DEPLOYING].
[DataSink (DiscardingOutput) (4/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Registering task at network: 
DataSink (DiscardingOutput) (4/16) (200278373fbd7c38ef75118d2cd413b1) 
[DEPLOYING].
[DataSink (DiscardingOutput) (4/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (4/16) 
(200278373fbd7c38ef75118d2cd413b1) switched from DEPLOYING to RUNNING.
[flink-akka.actor.default-dispatcher-5] INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph - DataSink 
(DiscardingOutput) (4/16) (200278373fbd7c38ef75118d2cd413b1) switched from 
DEPLOYING to RUNNING.
[DataSink (DiscardingOutput) (4/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - DataSink (DiscardingOutput) (4/16) 
(200278373fbd7c38ef75118d2cd413b1) switched from RUNNING to FINISHED.
[DataSink (DiscardingOutput) (4/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Freeing task resources for DataSink 
(DiscardingOutput) (4/16) (200278373fbd7c38ef75118d2cd413b1).
[DataSink (DiscardingOutput) (4/16)] INFO 
org.apache.flink.runtime.taskmanager.Task - Ensuring all FileSystem streams are 
closed for task DataSink (DiscardingOutput) (4/16) 
(200278373fbd7c38ef75118d2cd413b1) [FINISHED]
[flink-akka.actor.default-dispatcher-2] INFO 
org.apache.flink.runtime.taskexecutor.TaskExecutor - Un-registering task and 
sending final execution state FINISHED to JobManager for task DataSink 
(DiscardingOutput) 200278373fbd7c38ef75118d2cd413b1.

  1   2   >