[
https://issues.apache.org/jira/browse/BEAM-2719?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Sathish Jayaraman updated BEAM-2719:
------------------------------------
Description:
Hi,
The Beam job submitted for execution via spark-submit does not get past the
Evaluating ParMultiDo step. The compile execution runs fine when given
--runner=SparkRunner as parameter. But if I bundle the jar & submit it using
spark-submit, there were no executors getting assigned. I tried to submit with
both master spark-url & YARN but no luck in getting it executed past that step.
Below is the command I used to submit, the console log & job log from YARN.
I tried executing in both local single node cluster & in Azure HDInsight
cluster, the result is the same. So I guess there is nothing wrong in the Spark
configuration & could be a bug.
*Command:*
{code}
$ ~/spark/bin/spark-submit --class org.apache.beam.examples.WordCount --master
yarn --executor-memory 2G --num-executors 2
target/word-count-beam-0.1-shaded.jar --runner=SparkRunner --inputFile=pom.xml
--output=counts
{code}
*Terminal log*
{code}
17/08/03 15:18:15 WARN util.NativeCodeLoader: Unable to load native-hadoop
library for your platform... using builtin-java classes where applicable
17/08/03 15:18:16 INFO spark.SparkRunner: Executing pipeline using the
SparkRunner.
17/08/03 15:18:16 INFO translation.SparkContextFactory: Creating a brand new
Spark Context.
17/08/03 15:18:16 INFO spark.SparkContext: Running Spark version 2.1.1
17/08/03 15:18:17 INFO spark.SecurityManager: Changing view acls to:
sathishjayaraman
17/08/03 15:18:17 INFO spark.SecurityManager: Changing modify acls to:
sathishjayaraman
17/08/03 15:18:17 INFO spark.SecurityManager: Changing view acls groups to:
17/08/03 15:18:17 INFO spark.SecurityManager: Changing modify acls groups to:
17/08/03 15:18:17 INFO spark.SecurityManager: SecurityManager: authentication
disabled; ui acls disabled; users with view permissions:
Set(sathishjayaraman); groups with view permissions: Set(); users with modify
permissions: Set(sathishjayaraman); groups with modify permissions: Set()
17/08/03 15:18:17 INFO util.Utils: Successfully started service 'sparkDriver'
on port 51207.
17/08/03 15:18:17 INFO spark.SparkEnv: Registering MapOutputTracker
17/08/03 15:18:17 INFO spark.SparkEnv: Registering BlockManagerMaster
17/08/03 15:18:17 INFO storage.BlockManagerMasterEndpoint: Using
org.apache.spark.storage.DefaultTopologyMapper for getting topology information
17/08/03 15:18:17 INFO storage.BlockManagerMasterEndpoint:
BlockManagerMasterEndpoint up
17/08/03 15:18:17 INFO storage.DiskBlockManager: Created local directory at
/private/var/folders/d3/d1mrkc4s023d3qv1jr6w4cg00000gp/T/blockmgr-92d9827e-49e0-4035-b206-6fb4c24aa34c
17/08/03 15:18:17 INFO memory.MemoryStore: MemoryStore started with capacity
366.3 MB
17/08/03 15:18:17 INFO spark.SparkEnv: Registering OutputCommitCoordinator
17/08/03 15:18:18 INFO util.log: Logging initialized @5489ms
17/08/03 15:18:18 INFO server.Server: jetty-9.2.z-SNAPSHOT
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@2525a5b8{/jobs,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@3458eca5{/jobs/json,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@1e0fdb2f{/jobs/job,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@3c380bd8{/jobs/job/json,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@34b87182{/stages,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@47768e74{/stages/json,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@2740e316{/stages/stage,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@5b5a4aed{/stages/stage/json,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@2c991465{/stages/pool,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@5fef2aac{/stages/pool/json,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@7f973a14{/storage,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@76130a29{/storage/json,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@124d02b2{/storage/rdd,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@3451f01d{/storage/rdd/json,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@72503b19{/environment,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@1cfc2538{/environment/json,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@66deec87{/executors,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@5d342959{/executors/json,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@2ecf5915{/executors/threadDump,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@53e76c11{/executors/threadDump/json,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@42cc183e{/static,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@3e4e8fdf{/,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@6a1d6ef2{/api,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@2721044{/jobs/job/kill,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@205df5dc{/stages/stage/kill,null,AVAILABLE,@Spark}
17/08/03 15:18:18 WARN component.AbstractLifeCycle: FAILED
Spark@1756f7cc{HTTP/1.1}{0.0.0.0:4040}: java.net.BindException: Address already
in use
java.net.BindException: Address already in use
at sun.nio.ch.Net.bind0(Native Method)
at sun.nio.ch.Net.bind(Net.java:433)
at sun.nio.ch.Net.bind(Net.java:425)
at
sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:223)
at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:74)
at
org.spark_project.jetty.server.ServerConnector.open(ServerConnector.java:321)
at
org.spark_project.jetty.server.AbstractNetworkConnector.doStart(AbstractNetworkConnector.java:80)
at
org.spark_project.jetty.server.ServerConnector.doStart(ServerConnector.java:236)
at
org.spark_project.jetty.util.component.AbstractLifeCycle.start(AbstractLifeCycle.java:68)
at org.spark_project.jetty.server.Server.doStart(Server.java:366)
at
org.spark_project.jetty.util.component.AbstractLifeCycle.start(AbstractLifeCycle.java:68)
at
org.apache.spark.ui.JettyUtils$.org$apache$spark$ui$JettyUtils$$connect$1(JettyUtils.scala:365)
at org.apache.spark.ui.JettyUtils$$anonfun$5.apply(JettyUtils.scala:375)
at org.apache.spark.ui.JettyUtils$$anonfun$5.apply(JettyUtils.scala:375)
at
org.apache.spark.util.Utils$$anonfun$startServiceOnPort$1.apply$mcVI$sp(Utils.scala:2213)
at scala.collection.immutable.Range.foreach$mVc$sp(Range.scala:160)
at org.apache.spark.util.Utils$.startServiceOnPort(Utils.scala:2204)
at
org.apache.spark.ui.JettyUtils$.startJettyServer(JettyUtils.scala:375)
at org.apache.spark.ui.WebUI.bind(WebUI.scala:130)
at
org.apache.spark.SparkContext$$anonfun$10.apply(SparkContext.scala:460)
at
org.apache.spark.SparkContext$$anonfun$10.apply(SparkContext.scala:460)
at scala.Option.foreach(Option.scala:257)
at org.apache.spark.SparkContext.<init>(SparkContext.scala:460)
at
org.apache.spark.api.java.JavaSparkContext.<init>(JavaSparkContext.scala:58)
at
org.apache.beam.runners.spark.translation.SparkContextFactory.createSparkContext(SparkContextFactory.java:100)
at
org.apache.beam.runners.spark.translation.SparkContextFactory.getSparkContext(SparkContextFactory.java:69)
at org.apache.beam.runners.spark.SparkRunner.run(SparkRunner.java:195)
at org.apache.beam.runners.spark.SparkRunner.run(SparkRunner.java:85)
at org.apache.beam.sdk.Pipeline.run(Pipeline.java:295)
at org.apache.beam.sdk.Pipeline.run(Pipeline.java:281)
at org.apache.beam.examples.WordCount.main(WordCount.java:184)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at
org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:743)
at
org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:187)
at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:212)
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:126)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
17/08/03 15:18:18 WARN component.AbstractLifeCycle: FAILED
org.spark_project.jetty.server.Server@506aabf6: java.net.BindException: Address
already in use
java.net.BindException: Address already in use
at sun.nio.ch.Net.bind0(Native Method)
at sun.nio.ch.Net.bind(Net.java:433)
at sun.nio.ch.Net.bind(Net.java:425)
at
sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:223)
at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:74)
at
org.spark_project.jetty.server.ServerConnector.open(ServerConnector.java:321)
at
org.spark_project.jetty.server.AbstractNetworkConnector.doStart(AbstractNetworkConnector.java:80)
at
org.spark_project.jetty.server.ServerConnector.doStart(ServerConnector.java:236)
at
org.spark_project.jetty.util.component.AbstractLifeCycle.start(AbstractLifeCycle.java:68)
at org.spark_project.jetty.server.Server.doStart(Server.java:366)
at
org.spark_project.jetty.util.component.AbstractLifeCycle.start(AbstractLifeCycle.java:68)
at
org.apache.spark.ui.JettyUtils$.org$apache$spark$ui$JettyUtils$$connect$1(JettyUtils.scala:365)
at org.apache.spark.ui.JettyUtils$$anonfun$5.apply(JettyUtils.scala:375)
at org.apache.spark.ui.JettyUtils$$anonfun$5.apply(JettyUtils.scala:375)
at
org.apache.spark.util.Utils$$anonfun$startServiceOnPort$1.apply$mcVI$sp(Utils.scala:2213)
at scala.collection.immutable.Range.foreach$mVc$sp(Range.scala:160)
at org.apache.spark.util.Utils$.startServiceOnPort(Utils.scala:2204)
at
org.apache.spark.ui.JettyUtils$.startJettyServer(JettyUtils.scala:375)
at org.apache.spark.ui.WebUI.bind(WebUI.scala:130)
at
org.apache.spark.SparkContext$$anonfun$10.apply(SparkContext.scala:460)
at
org.apache.spark.SparkContext$$anonfun$10.apply(SparkContext.scala:460)
at scala.Option.foreach(Option.scala:257)
at org.apache.spark.SparkContext.<init>(SparkContext.scala:460)
at
org.apache.spark.api.java.JavaSparkContext.<init>(JavaSparkContext.scala:58)
at
org.apache.beam.runners.spark.translation.SparkContextFactory.createSparkContext(SparkContextFactory.java:100)
at
org.apache.beam.runners.spark.translation.SparkContextFactory.getSparkContext(SparkContextFactory.java:69)
at org.apache.beam.runners.spark.SparkRunner.run(SparkRunner.java:195)
at org.apache.beam.runners.spark.SparkRunner.run(SparkRunner.java:85)
at org.apache.beam.sdk.Pipeline.run(Pipeline.java:295)
at org.apache.beam.sdk.Pipeline.run(Pipeline.java:281)
at org.apache.beam.examples.WordCount.main(WordCount.java:184)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at
org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:743)
at
org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:187)
at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:212)
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:126)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
17/08/03 15:18:18 INFO server.ServerConnector: Stopped
Spark@1756f7cc{HTTP/1.1}{0.0.0.0:4040}
17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
o.s.j.s.ServletContextHandler@205df5dc{/stages/stage/kill,null,UNAVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
o.s.j.s.ServletContextHandler@2721044{/jobs/job/kill,null,UNAVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
o.s.j.s.ServletContextHandler@6a1d6ef2{/api,null,UNAVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
o.s.j.s.ServletContextHandler@3e4e8fdf{/,null,UNAVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
o.s.j.s.ServletContextHandler@42cc183e{/static,null,UNAVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
o.s.j.s.ServletContextHandler@53e76c11{/executors/threadDump/json,null,UNAVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
o.s.j.s.ServletContextHandler@2ecf5915{/executors/threadDump,null,UNAVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
o.s.j.s.ServletContextHandler@5d342959{/executors/json,null,UNAVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
o.s.j.s.ServletContextHandler@66deec87{/executors,null,UNAVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
o.s.j.s.ServletContextHandler@1cfc2538{/environment/json,null,UNAVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
o.s.j.s.ServletContextHandler@72503b19{/environment,null,UNAVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
o.s.j.s.ServletContextHandler@3451f01d{/storage/rdd/json,null,UNAVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
o.s.j.s.ServletContextHandler@124d02b2{/storage/rdd,null,UNAVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
o.s.j.s.ServletContextHandler@76130a29{/storage/json,null,UNAVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
o.s.j.s.ServletContextHandler@7f973a14{/storage,null,UNAVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
o.s.j.s.ServletContextHandler@5fef2aac{/stages/pool/json,null,UNAVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
o.s.j.s.ServletContextHandler@2c991465{/stages/pool,null,UNAVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
o.s.j.s.ServletContextHandler@5b5a4aed{/stages/stage/json,null,UNAVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
o.s.j.s.ServletContextHandler@2740e316{/stages/stage,null,UNAVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
o.s.j.s.ServletContextHandler@47768e74{/stages/json,null,UNAVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
o.s.j.s.ServletContextHandler@34b87182{/stages,null,UNAVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
o.s.j.s.ServletContextHandler@3c380bd8{/jobs/job/json,null,UNAVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
o.s.j.s.ServletContextHandler@1e0fdb2f{/jobs/job,null,UNAVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
o.s.j.s.ServletContextHandler@3458eca5{/jobs/json,null,UNAVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
o.s.j.s.ServletContextHandler@2525a5b8{/jobs,null,UNAVAILABLE,@Spark}
17/08/03 15:18:18 WARN util.Utils: Service 'SparkUI' could not bind on port
4040. Attempting port 4041.
17/08/03 15:18:18 INFO server.Server: jetty-9.2.z-SNAPSHOT
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@2525a5b8{/jobs,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@3458eca5{/jobs/json,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@1e0fdb2f{/jobs/job,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@3c380bd8{/jobs/job/json,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@34b87182{/stages,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@47768e74{/stages/json,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@2740e316{/stages/stage,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@5b5a4aed{/stages/stage/json,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@2c991465{/stages/pool,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@5fef2aac{/stages/pool/json,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@7f973a14{/storage,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@76130a29{/storage/json,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@124d02b2{/storage/rdd,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@3451f01d{/storage/rdd/json,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@72503b19{/environment,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@1cfc2538{/environment/json,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@66deec87{/executors,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@5d342959{/executors/json,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@2ecf5915{/executors/threadDump,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@53e76c11{/executors/threadDump/json,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@42cc183e{/static,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@3e4e8fdf{/,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@6a1d6ef2{/api,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@2721044{/jobs/job/kill,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@205df5dc{/stages/stage/kill,null,AVAILABLE,@Spark}
17/08/03 15:18:18 INFO server.ServerConnector: Started
Spark@3ba348ca{HTTP/1.1}{0.0.0.0:4041}
17/08/03 15:18:18 INFO server.Server: Started @5724ms
17/08/03 15:18:18 INFO util.Utils: Successfully started service 'SparkUI' on
port 4041.
17/08/03 15:18:18 INFO ui.SparkUI: Bound SparkUI to 0.0.0.0, and started at
http://192.168.0.7:4041
17/08/03 15:18:18 INFO spark.SparkContext: Added JAR
file:/Users/sathishjayaraman/java_projects/beamexample/word-count-beam/target/word-count-beam-0.1-shaded.jar
at spark://192.168.0.7:51207/jars/word-count-beam-0.1-shaded.jar with
timestamp 1501753698326
17/08/03 15:18:18 INFO client.RMProxy: Connecting to ResourceManager at
/0.0.0.0:8032
17/08/03 15:18:18 INFO yarn.Client: Requesting a new application from cluster
with 1 NodeManagers
17/08/03 15:18:18 INFO yarn.Client: Verifying our application has not requested
more than the maximum memory capability of the cluster (8192 MB per container)
17/08/03 15:18:18 INFO yarn.Client: Will allocate AM container, with 896 MB
memory including 384 MB overhead
17/08/03 15:18:18 INFO yarn.Client: Setting up container launch context for our
AM
17/08/03 15:18:18 INFO yarn.Client: Setting up the launch environment for our
AM container
17/08/03 15:18:18 INFO yarn.Client: Preparing resources for our AM container
17/08/03 15:18:20 WARN yarn.Client: Neither spark.yarn.jars nor
spark.yarn.archive is set, falling back to uploading libraries under SPARK_HOME.
17/08/03 15:18:23 INFO yarn.Client: Uploading resource
file:/private/var/folders/d3/d1mrkc4s023d3qv1jr6w4cg00000gp/T/spark-1c8deda6-98af-4ac3-8719-14ca7c90ddfc/__spark_libs__5010012016677506288.zip
->
hdfs://localhost:9000/user/sathishjayaraman/.sparkStaging/application_1501749993704_0001/__spark_libs__5010012016677506288.zip
17/08/03 15:18:25 INFO yarn.Client: Uploading resource
file:/private/var/folders/d3/d1mrkc4s023d3qv1jr6w4cg00000gp/T/spark-1c8deda6-98af-4ac3-8719-14ca7c90ddfc/__spark_conf__8657584076288234522.zip
->
hdfs://localhost:9000/user/sathishjayaraman/.sparkStaging/application_1501749993704_0001/__spark_conf__.zip
17/08/03 15:18:25 INFO spark.SecurityManager: Changing view acls to:
sathishjayaraman
17/08/03 15:18:25 INFO spark.SecurityManager: Changing modify acls to:
sathishjayaraman
17/08/03 15:18:25 INFO spark.SecurityManager: Changing view acls groups to:
17/08/03 15:18:25 INFO spark.SecurityManager: Changing modify acls groups to:
17/08/03 15:18:25 INFO spark.SecurityManager: SecurityManager: authentication
disabled; ui acls disabled; users with view permissions:
Set(sathishjayaraman); groups with view permissions: Set(); users with modify
permissions: Set(sathishjayaraman); groups with modify permissions: Set()
17/08/03 15:18:25 INFO yarn.Client: Submitting application
application_1501749993704_0001 to ResourceManager
17/08/03 15:18:26 INFO impl.YarnClientImpl: Submitted application
application_1501749993704_0001
17/08/03 15:18:26 INFO cluster.SchedulerExtensionServices: Starting Yarn
extension services with app application_1501749993704_0001 and attemptId None
17/08/03 15:18:27 INFO yarn.Client: Application report for
application_1501749993704_0001 (state: ACCEPTED)
17/08/03 15:18:27 INFO yarn.Client:
client token: N/A
diagnostics: AM container is launched, waiting for AM container to
Register with RM
ApplicationMaster host: N/A
ApplicationMaster RPC port: -1
queue: default
start time: 1501753705959
final status: UNDEFINED
tracking URL:
http://Quartics-MacBook-Pro.local:8088/proxy/application_1501749993704_0001/
user: sathishjayaraman
17/08/03 15:18:28 INFO yarn.Client: Application report for
application_1501749993704_0001 (state: ACCEPTED)
17/08/03 15:18:29 INFO yarn.Client: Application report for
application_1501749993704_0001 (state: ACCEPTED)
17/08/03 15:18:30 INFO yarn.Client: Application report for
application_1501749993704_0001 (state: ACCEPTED)
17/08/03 15:18:31 INFO yarn.Client: Application report for
application_1501749993704_0001 (state: ACCEPTED)
17/08/03 15:18:32 INFO yarn.Client: Application report for
application_1501749993704_0001 (state: ACCEPTED)
17/08/03 15:18:33 INFO yarn.Client: Application report for
application_1501749993704_0001 (state: ACCEPTED)
17/08/03 15:18:34 INFO yarn.Client: Application report for
application_1501749993704_0001 (state: ACCEPTED)
17/08/03 15:18:35 INFO yarn.Client: Application report for
application_1501749993704_0001 (state: ACCEPTED)
17/08/03 15:18:35 INFO cluster.YarnSchedulerBackend$YarnSchedulerEndpoint:
ApplicationMaster registered as NettyRpcEndpointRef(null)
17/08/03 15:18:35 INFO cluster.YarnClientSchedulerBackend: Add WebUI Filter.
org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter, Map(PROXY_HOSTS ->
Quartics-MacBook-Pro.local, PROXY_URI_BASES ->
http://Quartics-MacBook-Pro.local:8088/proxy/application_1501749993704_0001),
/proxy/application_1501749993704_0001
17/08/03 15:18:35 INFO ui.JettyUtils: Adding filter:
org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter
17/08/03 15:18:36 INFO yarn.Client: Application report for
application_1501749993704_0001 (state: RUNNING)
17/08/03 15:18:36 INFO yarn.Client:
client token: N/A
diagnostics: N/A
ApplicationMaster host: 192.168.0.7
ApplicationMaster RPC port: 0
queue: default
start time: 1501753705959
final status: UNDEFINED
tracking URL:
http://Quartics-MacBook-Pro.local:8088/proxy/application_1501749993704_0001/
user: sathishjayaraman
17/08/03 15:18:36 INFO cluster.YarnClientSchedulerBackend: Application
application_1501749993704_0001 has started running.
17/08/03 15:18:36 INFO util.Utils: Successfully started service
'org.apache.spark.network.netty.NettyBlockTransferService' on port 51221.
17/08/03 15:18:36 INFO netty.NettyBlockTransferService: Server created on
192.168.0.7:51221
17/08/03 15:18:36 INFO storage.BlockManager: Using
org.apache.spark.storage.RandomBlockReplicationPolicy for block replication
policy
17/08/03 15:18:36 INFO storage.BlockManagerMaster: Registering BlockManager
BlockManagerId(driver, 192.168.0.7, 51221, None)
17/08/03 15:18:36 INFO storage.BlockManagerMasterEndpoint: Registering block
manager 192.168.0.7:51221 with 366.3 MB RAM, BlockManagerId(driver,
192.168.0.7, 51221, None)
17/08/03 15:18:36 INFO storage.BlockManagerMaster: Registered BlockManager
BlockManagerId(driver, 192.168.0.7, 51221, None)
17/08/03 15:18:36 INFO storage.BlockManager: Initialized BlockManager:
BlockManagerId(driver, 192.168.0.7, 51221, None)
17/08/03 15:18:36 INFO handler.ContextHandler: Started
o.s.j.s.ServletContextHandler@350bbd5d{/metrics/json,null,AVAILABLE,@Spark}
17/08/03 15:18:41 INFO cluster.YarnSchedulerBackend$YarnDriverEndpoint:
Registered executor NettyRpcEndpointRef(null) (192.168.0.7:51226) with ID 1
17/08/03 15:18:41 INFO storage.BlockManagerMasterEndpoint: Registering block
manager 192.168.0.7:51229 with 912.3 MB RAM, BlockManagerId(1, 192.168.0.7,
51229, None)
17/08/03 15:18:42 INFO cluster.YarnSchedulerBackend$YarnDriverEndpoint:
Registered executor NettyRpcEndpointRef(null) (192.168.0.7:51228) with ID 2
17/08/03 15:18:42 INFO cluster.YarnClientSchedulerBackend: SchedulerBackend is
ready for scheduling beginning after reached minRegisteredResourcesRatio: 0.8
17/08/03 15:18:42 INFO storage.BlockManagerMasterEndpoint: Registering block
manager 192.168.0.7:51230 with 912.3 MB RAM, BlockManagerId(2, 192.168.0.7,
51230, None)
17/08/03 15:18:42 INFO Configuration.deprecation: dfs.data.dir is deprecated.
Instead, use dfs.datanode.data.dir
17/08/03 15:18:42 INFO Configuration.deprecation: dfs.name.dir is deprecated.
Instead, use dfs.namenode.name.dir
17/08/03 15:18:42 INFO spark.SparkRunner$Evaluator: Entering
directly-translatable composite transform:
'WordCount.CountWords/Count.PerElement/Combine.perKey(Count)'
17/08/03 15:18:42 INFO spark.SparkRunner$Evaluator: Entering
directly-translatable composite transform:
'WriteCounts/WriteFiles/View.AsIterable'
17/08/03 15:18:42 INFO spark.SparkRunner$Evaluator: Entering
directly-translatable composite transform:
'WriteCounts/WriteFiles/Create.Values'
17/08/03 15:18:42 INFO metrics.MetricsAccumulator: Instantiated metrics
accumulator:
org.apache.beam.runners.core.metrics.MetricsContainerStepMap@2b1a901d
17/08/03 15:18:42 INFO aggregators.AggregatorsAccumulator: Instantiated
aggregators accumulator:
17/08/03 15:18:42 INFO spark.SparkRunner$Evaluator: Evaluating
Read(CompressedSource)
17/08/03 15:18:42 INFO spark.SparkRunner$Evaluator: Evaluating
ParMultiDo(ExtractWords)
{code}
*YARN Job jog*
{code}
17/08/03 13:00:33 INFO client.RMProxy: Connecting to ResourceManager at
/0.0.0.0:8030
17/08/03 13:00:33 INFO yarn.YarnRMClient: Registering the ApplicationMaster
17/08/03 13:00:34 INFO yarn.YarnAllocator: Will request 2 executor
container(s), each with 1 core(s) and 2432 MB memory (including 384 MB of
overhead)
17/08/03 13:00:34 INFO yarn.YarnAllocator: Submitted 2 unlocalized container
requests.
17/08/03 13:00:34 INFO yarn.ApplicationMaster: Started progress reporter thread
with (heartbeat : 3000, initial allocation : 200) intervals
17/08/03 13:00:35 INFO impl.AMRMClientImpl: Received new token for :
192.168.0.7:50173
17/08/03 13:00:35 INFO yarn.YarnAllocator: Launching container
container_1501744514957_0003_01_000002 on host 192.168.0.7
17/08/03 13:00:35 INFO yarn.YarnAllocator: Received 1 containers from YARN,
launching executors on 1 of them.
17/08/03 13:00:35 INFO impl.ContainerManagementProtocolProxy:
yarn.client.max-cached-nodemanagers-proxies : 0
17/08/03 13:00:35 INFO impl.ContainerManagementProtocolProxy: Opening proxy :
192.168.0.7:50173
17/08/03 13:00:37 INFO yarn.YarnAllocator: Launching container
container_1501744514957_0003_01_000003 on host 192.168.0.7
17/08/03 13:00:37 INFO yarn.YarnAllocator: Received 1 containers from YARN,
launching executors on 1 of them.
17/08/03 13:00:37 INFO impl.ContainerManagementProtocolProxy:
yarn.client.max-cached-nodemanagers-proxies : 0
17/08/03 13:00:37 INFO impl.ContainerManagementProtocolProxy: Opening proxy :
192.168.0.7:50173
{code}
was:
Hi,
The Beam job submitted for execution via spark-submit does not get past the
Evaluating ParMultiDo step. The compile execution runs fine when given
--runner=SparkRunner as parameter. But if I bundle the jar & submit it using
spark-submit, there were no executors getting assigned. I tried to submit with
both master spark-url & YARN but no luck in getting it executed past that step.
Below is the command I used to submit & job log from YARN.
I tried executing in both local single node cluster & in Azure HDInsight
cluster, the result is the same. So I guess there is nothing wrong in the Spark
configuration & could be a bug.
{code}
$ ~/spark/bin/spark-submit --class org.apache.beam.examples.WordCount --master
yarn --executor-memory 2G --num-executors 2
target/word-count-beam-0.1-shaded.jar --runner=SparkRunner --inputFile=pom.xml
--output=counts
{code}
{code}
17/08/03 13:00:33 INFO client.RMProxy: Connecting to ResourceManager at
/0.0.0.0:8030
17/08/03 13:00:33 INFO yarn.YarnRMClient: Registering the ApplicationMaster
17/08/03 13:00:34 INFO yarn.YarnAllocator: Will request 2 executor
container(s), each with 1 core(s) and 2432 MB memory (including 384 MB of
overhead)
17/08/03 13:00:34 INFO yarn.YarnAllocator: Submitted 2 unlocalized container
requests.
17/08/03 13:00:34 INFO yarn.ApplicationMaster: Started progress reporter thread
with (heartbeat : 3000, initial allocation : 200) intervals
17/08/03 13:00:35 INFO impl.AMRMClientImpl: Received new token for :
192.168.0.7:50173
17/08/03 13:00:35 INFO yarn.YarnAllocator: Launching container
container_1501744514957_0003_01_000002 on host 192.168.0.7
17/08/03 13:00:35 INFO yarn.YarnAllocator: Received 1 containers from YARN,
launching executors on 1 of them.
17/08/03 13:00:35 INFO impl.ContainerManagementProtocolProxy:
yarn.client.max-cached-nodemanagers-proxies : 0
17/08/03 13:00:35 INFO impl.ContainerManagementProtocolProxy: Opening proxy :
192.168.0.7:50173
17/08/03 13:00:37 INFO yarn.YarnAllocator: Launching container
container_1501744514957_0003_01_000003 on host 192.168.0.7
17/08/03 13:00:37 INFO yarn.YarnAllocator: Received 1 containers from YARN,
launching executors on 1 of them.
17/08/03 13:00:37 INFO impl.ContainerManagementProtocolProxy:
yarn.client.max-cached-nodemanagers-proxies : 0
17/08/03 13:00:37 INFO impl.ContainerManagementProtocolProxy: Opening proxy :
192.168.0.7:50173
{code}
> Beam job hangs at Evaluating ParMultiDo when submitted via spark-runner
> ------------------------------------------------------------------------
>
> Key: BEAM-2719
> URL: https://issues.apache.org/jira/browse/BEAM-2719
> Project: Beam
> Issue Type: Bug
> Components: runner-spark
> Affects Versions: 2.0.0
> Environment: OSX / i5 / 10GB
> Reporter: Sathish Jayaraman
> Assignee: Jean-Baptiste Onofré
>
> Hi,
> The Beam job submitted for execution via spark-submit does not get past the
> Evaluating ParMultiDo step. The compile execution runs fine when given
> --runner=SparkRunner as parameter. But if I bundle the jar & submit it using
> spark-submit, there were no executors getting assigned. I tried to submit
> with both master spark-url & YARN but no luck in getting it executed past
> that step. Below is the command I used to submit, the console log & job log
> from YARN.
> I tried executing in both local single node cluster & in Azure HDInsight
> cluster, the result is the same. So I guess there is nothing wrong in the
> Spark configuration & could be a bug.
> *Command:*
> {code}
> $ ~/spark/bin/spark-submit --class org.apache.beam.examples.WordCount
> --master yarn --executor-memory 2G --num-executors 2
> target/word-count-beam-0.1-shaded.jar --runner=SparkRunner
> --inputFile=pom.xml --output=counts
> {code}
> *Terminal log*
> {code}
> 17/08/03 15:18:15 WARN util.NativeCodeLoader: Unable to load native-hadoop
> library for your platform... using builtin-java classes where applicable
> 17/08/03 15:18:16 INFO spark.SparkRunner: Executing pipeline using the
> SparkRunner.
> 17/08/03 15:18:16 INFO translation.SparkContextFactory: Creating a brand new
> Spark Context.
> 17/08/03 15:18:16 INFO spark.SparkContext: Running Spark version 2.1.1
> 17/08/03 15:18:17 INFO spark.SecurityManager: Changing view acls to:
> sathishjayaraman
> 17/08/03 15:18:17 INFO spark.SecurityManager: Changing modify acls to:
> sathishjayaraman
> 17/08/03 15:18:17 INFO spark.SecurityManager: Changing view acls groups to:
> 17/08/03 15:18:17 INFO spark.SecurityManager: Changing modify acls groups to:
> 17/08/03 15:18:17 INFO spark.SecurityManager: SecurityManager: authentication
> disabled; ui acls disabled; users with view permissions:
> Set(sathishjayaraman); groups with view permissions: Set(); users with
> modify permissions: Set(sathishjayaraman); groups with modify permissions:
> Set()
> 17/08/03 15:18:17 INFO util.Utils: Successfully started service 'sparkDriver'
> on port 51207.
> 17/08/03 15:18:17 INFO spark.SparkEnv: Registering MapOutputTracker
> 17/08/03 15:18:17 INFO spark.SparkEnv: Registering BlockManagerMaster
> 17/08/03 15:18:17 INFO storage.BlockManagerMasterEndpoint: Using
> org.apache.spark.storage.DefaultTopologyMapper for getting topology
> information
> 17/08/03 15:18:17 INFO storage.BlockManagerMasterEndpoint:
> BlockManagerMasterEndpoint up
> 17/08/03 15:18:17 INFO storage.DiskBlockManager: Created local directory at
> /private/var/folders/d3/d1mrkc4s023d3qv1jr6w4cg00000gp/T/blockmgr-92d9827e-49e0-4035-b206-6fb4c24aa34c
> 17/08/03 15:18:17 INFO memory.MemoryStore: MemoryStore started with capacity
> 366.3 MB
> 17/08/03 15:18:17 INFO spark.SparkEnv: Registering OutputCommitCoordinator
> 17/08/03 15:18:18 INFO util.log: Logging initialized @5489ms
> 17/08/03 15:18:18 INFO server.Server: jetty-9.2.z-SNAPSHOT
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@2525a5b8{/jobs,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@3458eca5{/jobs/json,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@1e0fdb2f{/jobs/job,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@3c380bd8{/jobs/job/json,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@34b87182{/stages,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@47768e74{/stages/json,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@2740e316{/stages/stage,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@5b5a4aed{/stages/stage/json,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@2c991465{/stages/pool,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@5fef2aac{/stages/pool/json,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@7f973a14{/storage,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@76130a29{/storage/json,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@124d02b2{/storage/rdd,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@3451f01d{/storage/rdd/json,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@72503b19{/environment,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@1cfc2538{/environment/json,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@66deec87{/executors,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@5d342959{/executors/json,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@2ecf5915{/executors/threadDump,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@53e76c11{/executors/threadDump/json,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@42cc183e{/static,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@3e4e8fdf{/,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@6a1d6ef2{/api,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@2721044{/jobs/job/kill,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@205df5dc{/stages/stage/kill,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 WARN component.AbstractLifeCycle: FAILED
> Spark@1756f7cc{HTTP/1.1}{0.0.0.0:4040}: java.net.BindException: Address
> already in use
> java.net.BindException: Address already in use
> at sun.nio.ch.Net.bind0(Native Method)
> at sun.nio.ch.Net.bind(Net.java:433)
> at sun.nio.ch.Net.bind(Net.java:425)
> at
> sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:223)
> at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:74)
> at
> org.spark_project.jetty.server.ServerConnector.open(ServerConnector.java:321)
> at
> org.spark_project.jetty.server.AbstractNetworkConnector.doStart(AbstractNetworkConnector.java:80)
> at
> org.spark_project.jetty.server.ServerConnector.doStart(ServerConnector.java:236)
> at
> org.spark_project.jetty.util.component.AbstractLifeCycle.start(AbstractLifeCycle.java:68)
> at org.spark_project.jetty.server.Server.doStart(Server.java:366)
> at
> org.spark_project.jetty.util.component.AbstractLifeCycle.start(AbstractLifeCycle.java:68)
> at
> org.apache.spark.ui.JettyUtils$.org$apache$spark$ui$JettyUtils$$connect$1(JettyUtils.scala:365)
> at org.apache.spark.ui.JettyUtils$$anonfun$5.apply(JettyUtils.scala:375)
> at org.apache.spark.ui.JettyUtils$$anonfun$5.apply(JettyUtils.scala:375)
> at
> org.apache.spark.util.Utils$$anonfun$startServiceOnPort$1.apply$mcVI$sp(Utils.scala:2213)
> at scala.collection.immutable.Range.foreach$mVc$sp(Range.scala:160)
> at org.apache.spark.util.Utils$.startServiceOnPort(Utils.scala:2204)
> at
> org.apache.spark.ui.JettyUtils$.startJettyServer(JettyUtils.scala:375)
> at org.apache.spark.ui.WebUI.bind(WebUI.scala:130)
> at
> org.apache.spark.SparkContext$$anonfun$10.apply(SparkContext.scala:460)
> at
> org.apache.spark.SparkContext$$anonfun$10.apply(SparkContext.scala:460)
> at scala.Option.foreach(Option.scala:257)
> at org.apache.spark.SparkContext.<init>(SparkContext.scala:460)
> at
> org.apache.spark.api.java.JavaSparkContext.<init>(JavaSparkContext.scala:58)
> at
> org.apache.beam.runners.spark.translation.SparkContextFactory.createSparkContext(SparkContextFactory.java:100)
> at
> org.apache.beam.runners.spark.translation.SparkContextFactory.getSparkContext(SparkContextFactory.java:69)
> at org.apache.beam.runners.spark.SparkRunner.run(SparkRunner.java:195)
> at org.apache.beam.runners.spark.SparkRunner.run(SparkRunner.java:85)
> at org.apache.beam.sdk.Pipeline.run(Pipeline.java:295)
> at org.apache.beam.sdk.Pipeline.run(Pipeline.java:281)
> at org.apache.beam.examples.WordCount.main(WordCount.java:184)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498)
> at
> org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:743)
> at
> org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:187)
> at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:212)
> at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:126)
> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> 17/08/03 15:18:18 WARN component.AbstractLifeCycle: FAILED
> org.spark_project.jetty.server.Server@506aabf6: java.net.BindException:
> Address already in use
> java.net.BindException: Address already in use
> at sun.nio.ch.Net.bind0(Native Method)
> at sun.nio.ch.Net.bind(Net.java:433)
> at sun.nio.ch.Net.bind(Net.java:425)
> at
> sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:223)
> at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:74)
> at
> org.spark_project.jetty.server.ServerConnector.open(ServerConnector.java:321)
> at
> org.spark_project.jetty.server.AbstractNetworkConnector.doStart(AbstractNetworkConnector.java:80)
> at
> org.spark_project.jetty.server.ServerConnector.doStart(ServerConnector.java:236)
> at
> org.spark_project.jetty.util.component.AbstractLifeCycle.start(AbstractLifeCycle.java:68)
> at org.spark_project.jetty.server.Server.doStart(Server.java:366)
> at
> org.spark_project.jetty.util.component.AbstractLifeCycle.start(AbstractLifeCycle.java:68)
> at
> org.apache.spark.ui.JettyUtils$.org$apache$spark$ui$JettyUtils$$connect$1(JettyUtils.scala:365)
> at org.apache.spark.ui.JettyUtils$$anonfun$5.apply(JettyUtils.scala:375)
> at org.apache.spark.ui.JettyUtils$$anonfun$5.apply(JettyUtils.scala:375)
> at
> org.apache.spark.util.Utils$$anonfun$startServiceOnPort$1.apply$mcVI$sp(Utils.scala:2213)
> at scala.collection.immutable.Range.foreach$mVc$sp(Range.scala:160)
> at org.apache.spark.util.Utils$.startServiceOnPort(Utils.scala:2204)
> at
> org.apache.spark.ui.JettyUtils$.startJettyServer(JettyUtils.scala:375)
> at org.apache.spark.ui.WebUI.bind(WebUI.scala:130)
> at
> org.apache.spark.SparkContext$$anonfun$10.apply(SparkContext.scala:460)
> at
> org.apache.spark.SparkContext$$anonfun$10.apply(SparkContext.scala:460)
> at scala.Option.foreach(Option.scala:257)
> at org.apache.spark.SparkContext.<init>(SparkContext.scala:460)
> at
> org.apache.spark.api.java.JavaSparkContext.<init>(JavaSparkContext.scala:58)
> at
> org.apache.beam.runners.spark.translation.SparkContextFactory.createSparkContext(SparkContextFactory.java:100)
> at
> org.apache.beam.runners.spark.translation.SparkContextFactory.getSparkContext(SparkContextFactory.java:69)
> at org.apache.beam.runners.spark.SparkRunner.run(SparkRunner.java:195)
> at org.apache.beam.runners.spark.SparkRunner.run(SparkRunner.java:85)
> at org.apache.beam.sdk.Pipeline.run(Pipeline.java:295)
> at org.apache.beam.sdk.Pipeline.run(Pipeline.java:281)
> at org.apache.beam.examples.WordCount.main(WordCount.java:184)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498)
> at
> org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:743)
> at
> org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:187)
> at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:212)
> at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:126)
> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> 17/08/03 15:18:18 INFO server.ServerConnector: Stopped
> Spark@1756f7cc{HTTP/1.1}{0.0.0.0:4040}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
> o.s.j.s.ServletContextHandler@205df5dc{/stages/stage/kill,null,UNAVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
> o.s.j.s.ServletContextHandler@2721044{/jobs/job/kill,null,UNAVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
> o.s.j.s.ServletContextHandler@6a1d6ef2{/api,null,UNAVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
> o.s.j.s.ServletContextHandler@3e4e8fdf{/,null,UNAVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
> o.s.j.s.ServletContextHandler@42cc183e{/static,null,UNAVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
> o.s.j.s.ServletContextHandler@53e76c11{/executors/threadDump/json,null,UNAVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
> o.s.j.s.ServletContextHandler@2ecf5915{/executors/threadDump,null,UNAVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
> o.s.j.s.ServletContextHandler@5d342959{/executors/json,null,UNAVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
> o.s.j.s.ServletContextHandler@66deec87{/executors,null,UNAVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
> o.s.j.s.ServletContextHandler@1cfc2538{/environment/json,null,UNAVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
> o.s.j.s.ServletContextHandler@72503b19{/environment,null,UNAVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
> o.s.j.s.ServletContextHandler@3451f01d{/storage/rdd/json,null,UNAVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
> o.s.j.s.ServletContextHandler@124d02b2{/storage/rdd,null,UNAVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
> o.s.j.s.ServletContextHandler@76130a29{/storage/json,null,UNAVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
> o.s.j.s.ServletContextHandler@7f973a14{/storage,null,UNAVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
> o.s.j.s.ServletContextHandler@5fef2aac{/stages/pool/json,null,UNAVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
> o.s.j.s.ServletContextHandler@2c991465{/stages/pool,null,UNAVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
> o.s.j.s.ServletContextHandler@5b5a4aed{/stages/stage/json,null,UNAVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
> o.s.j.s.ServletContextHandler@2740e316{/stages/stage,null,UNAVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
> o.s.j.s.ServletContextHandler@47768e74{/stages/json,null,UNAVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
> o.s.j.s.ServletContextHandler@34b87182{/stages,null,UNAVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
> o.s.j.s.ServletContextHandler@3c380bd8{/jobs/job/json,null,UNAVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
> o.s.j.s.ServletContextHandler@1e0fdb2f{/jobs/job,null,UNAVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
> o.s.j.s.ServletContextHandler@3458eca5{/jobs/json,null,UNAVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Stopped
> o.s.j.s.ServletContextHandler@2525a5b8{/jobs,null,UNAVAILABLE,@Spark}
> 17/08/03 15:18:18 WARN util.Utils: Service 'SparkUI' could not bind on port
> 4040. Attempting port 4041.
> 17/08/03 15:18:18 INFO server.Server: jetty-9.2.z-SNAPSHOT
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@2525a5b8{/jobs,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@3458eca5{/jobs/json,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@1e0fdb2f{/jobs/job,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@3c380bd8{/jobs/job/json,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@34b87182{/stages,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@47768e74{/stages/json,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@2740e316{/stages/stage,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@5b5a4aed{/stages/stage/json,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@2c991465{/stages/pool,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@5fef2aac{/stages/pool/json,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@7f973a14{/storage,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@76130a29{/storage/json,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@124d02b2{/storage/rdd,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@3451f01d{/storage/rdd/json,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@72503b19{/environment,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@1cfc2538{/environment/json,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@66deec87{/executors,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@5d342959{/executors/json,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@2ecf5915{/executors/threadDump,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@53e76c11{/executors/threadDump/json,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@42cc183e{/static,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@3e4e8fdf{/,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@6a1d6ef2{/api,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@2721044{/jobs/job/kill,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@205df5dc{/stages/stage/kill,null,AVAILABLE,@Spark}
> 17/08/03 15:18:18 INFO server.ServerConnector: Started
> Spark@3ba348ca{HTTP/1.1}{0.0.0.0:4041}
> 17/08/03 15:18:18 INFO server.Server: Started @5724ms
> 17/08/03 15:18:18 INFO util.Utils: Successfully started service 'SparkUI' on
> port 4041.
> 17/08/03 15:18:18 INFO ui.SparkUI: Bound SparkUI to 0.0.0.0, and started at
> http://192.168.0.7:4041
> 17/08/03 15:18:18 INFO spark.SparkContext: Added JAR
> file:/Users/sathishjayaraman/java_projects/beamexample/word-count-beam/target/word-count-beam-0.1-shaded.jar
> at spark://192.168.0.7:51207/jars/word-count-beam-0.1-shaded.jar with
> timestamp 1501753698326
> 17/08/03 15:18:18 INFO client.RMProxy: Connecting to ResourceManager at
> /0.0.0.0:8032
> 17/08/03 15:18:18 INFO yarn.Client: Requesting a new application from cluster
> with 1 NodeManagers
> 17/08/03 15:18:18 INFO yarn.Client: Verifying our application has not
> requested more than the maximum memory capability of the cluster (8192 MB per
> container)
> 17/08/03 15:18:18 INFO yarn.Client: Will allocate AM container, with 896 MB
> memory including 384 MB overhead
> 17/08/03 15:18:18 INFO yarn.Client: Setting up container launch context for
> our AM
> 17/08/03 15:18:18 INFO yarn.Client: Setting up the launch environment for our
> AM container
> 17/08/03 15:18:18 INFO yarn.Client: Preparing resources for our AM container
> 17/08/03 15:18:20 WARN yarn.Client: Neither spark.yarn.jars nor
> spark.yarn.archive is set, falling back to uploading libraries under
> SPARK_HOME.
> 17/08/03 15:18:23 INFO yarn.Client: Uploading resource
> file:/private/var/folders/d3/d1mrkc4s023d3qv1jr6w4cg00000gp/T/spark-1c8deda6-98af-4ac3-8719-14ca7c90ddfc/__spark_libs__5010012016677506288.zip
> ->
> hdfs://localhost:9000/user/sathishjayaraman/.sparkStaging/application_1501749993704_0001/__spark_libs__5010012016677506288.zip
> 17/08/03 15:18:25 INFO yarn.Client: Uploading resource
> file:/private/var/folders/d3/d1mrkc4s023d3qv1jr6w4cg00000gp/T/spark-1c8deda6-98af-4ac3-8719-14ca7c90ddfc/__spark_conf__8657584076288234522.zip
> ->
> hdfs://localhost:9000/user/sathishjayaraman/.sparkStaging/application_1501749993704_0001/__spark_conf__.zip
> 17/08/03 15:18:25 INFO spark.SecurityManager: Changing view acls to:
> sathishjayaraman
> 17/08/03 15:18:25 INFO spark.SecurityManager: Changing modify acls to:
> sathishjayaraman
> 17/08/03 15:18:25 INFO spark.SecurityManager: Changing view acls groups to:
> 17/08/03 15:18:25 INFO spark.SecurityManager: Changing modify acls groups to:
> 17/08/03 15:18:25 INFO spark.SecurityManager: SecurityManager: authentication
> disabled; ui acls disabled; users with view permissions:
> Set(sathishjayaraman); groups with view permissions: Set(); users with
> modify permissions: Set(sathishjayaraman); groups with modify permissions:
> Set()
> 17/08/03 15:18:25 INFO yarn.Client: Submitting application
> application_1501749993704_0001 to ResourceManager
> 17/08/03 15:18:26 INFO impl.YarnClientImpl: Submitted application
> application_1501749993704_0001
> 17/08/03 15:18:26 INFO cluster.SchedulerExtensionServices: Starting Yarn
> extension services with app application_1501749993704_0001 and attemptId None
> 17/08/03 15:18:27 INFO yarn.Client: Application report for
> application_1501749993704_0001 (state: ACCEPTED)
> 17/08/03 15:18:27 INFO yarn.Client:
> client token: N/A
> diagnostics: AM container is launched, waiting for AM container to
> Register with RM
> ApplicationMaster host: N/A
> ApplicationMaster RPC port: -1
> queue: default
> start time: 1501753705959
> final status: UNDEFINED
> tracking URL:
> http://Quartics-MacBook-Pro.local:8088/proxy/application_1501749993704_0001/
> user: sathishjayaraman
> 17/08/03 15:18:28 INFO yarn.Client: Application report for
> application_1501749993704_0001 (state: ACCEPTED)
> 17/08/03 15:18:29 INFO yarn.Client: Application report for
> application_1501749993704_0001 (state: ACCEPTED)
> 17/08/03 15:18:30 INFO yarn.Client: Application report for
> application_1501749993704_0001 (state: ACCEPTED)
> 17/08/03 15:18:31 INFO yarn.Client: Application report for
> application_1501749993704_0001 (state: ACCEPTED)
> 17/08/03 15:18:32 INFO yarn.Client: Application report for
> application_1501749993704_0001 (state: ACCEPTED)
> 17/08/03 15:18:33 INFO yarn.Client: Application report for
> application_1501749993704_0001 (state: ACCEPTED)
> 17/08/03 15:18:34 INFO yarn.Client: Application report for
> application_1501749993704_0001 (state: ACCEPTED)
> 17/08/03 15:18:35 INFO yarn.Client: Application report for
> application_1501749993704_0001 (state: ACCEPTED)
> 17/08/03 15:18:35 INFO cluster.YarnSchedulerBackend$YarnSchedulerEndpoint:
> ApplicationMaster registered as NettyRpcEndpointRef(null)
> 17/08/03 15:18:35 INFO cluster.YarnClientSchedulerBackend: Add WebUI Filter.
> org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter, Map(PROXY_HOSTS
> -> Quartics-MacBook-Pro.local, PROXY_URI_BASES ->
> http://Quartics-MacBook-Pro.local:8088/proxy/application_1501749993704_0001),
> /proxy/application_1501749993704_0001
> 17/08/03 15:18:35 INFO ui.JettyUtils: Adding filter:
> org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter
> 17/08/03 15:18:36 INFO yarn.Client: Application report for
> application_1501749993704_0001 (state: RUNNING)
> 17/08/03 15:18:36 INFO yarn.Client:
> client token: N/A
> diagnostics: N/A
> ApplicationMaster host: 192.168.0.7
> ApplicationMaster RPC port: 0
> queue: default
> start time: 1501753705959
> final status: UNDEFINED
> tracking URL:
> http://Quartics-MacBook-Pro.local:8088/proxy/application_1501749993704_0001/
> user: sathishjayaraman
> 17/08/03 15:18:36 INFO cluster.YarnClientSchedulerBackend: Application
> application_1501749993704_0001 has started running.
> 17/08/03 15:18:36 INFO util.Utils: Successfully started service
> 'org.apache.spark.network.netty.NettyBlockTransferService' on port 51221.
> 17/08/03 15:18:36 INFO netty.NettyBlockTransferService: Server created on
> 192.168.0.7:51221
> 17/08/03 15:18:36 INFO storage.BlockManager: Using
> org.apache.spark.storage.RandomBlockReplicationPolicy for block replication
> policy
> 17/08/03 15:18:36 INFO storage.BlockManagerMaster: Registering BlockManager
> BlockManagerId(driver, 192.168.0.7, 51221, None)
> 17/08/03 15:18:36 INFO storage.BlockManagerMasterEndpoint: Registering block
> manager 192.168.0.7:51221 with 366.3 MB RAM, BlockManagerId(driver,
> 192.168.0.7, 51221, None)
> 17/08/03 15:18:36 INFO storage.BlockManagerMaster: Registered BlockManager
> BlockManagerId(driver, 192.168.0.7, 51221, None)
> 17/08/03 15:18:36 INFO storage.BlockManager: Initialized BlockManager:
> BlockManagerId(driver, 192.168.0.7, 51221, None)
> 17/08/03 15:18:36 INFO handler.ContextHandler: Started
> o.s.j.s.ServletContextHandler@350bbd5d{/metrics/json,null,AVAILABLE,@Spark}
> 17/08/03 15:18:41 INFO cluster.YarnSchedulerBackend$YarnDriverEndpoint:
> Registered executor NettyRpcEndpointRef(null) (192.168.0.7:51226) with ID 1
> 17/08/03 15:18:41 INFO storage.BlockManagerMasterEndpoint: Registering block
> manager 192.168.0.7:51229 with 912.3 MB RAM, BlockManagerId(1, 192.168.0.7,
> 51229, None)
> 17/08/03 15:18:42 INFO cluster.YarnSchedulerBackend$YarnDriverEndpoint:
> Registered executor NettyRpcEndpointRef(null) (192.168.0.7:51228) with ID 2
> 17/08/03 15:18:42 INFO cluster.YarnClientSchedulerBackend: SchedulerBackend
> is ready for scheduling beginning after reached minRegisteredResourcesRatio:
> 0.8
> 17/08/03 15:18:42 INFO storage.BlockManagerMasterEndpoint: Registering block
> manager 192.168.0.7:51230 with 912.3 MB RAM, BlockManagerId(2, 192.168.0.7,
> 51230, None)
> 17/08/03 15:18:42 INFO Configuration.deprecation: dfs.data.dir is deprecated.
> Instead, use dfs.datanode.data.dir
> 17/08/03 15:18:42 INFO Configuration.deprecation: dfs.name.dir is deprecated.
> Instead, use dfs.namenode.name.dir
> 17/08/03 15:18:42 INFO spark.SparkRunner$Evaluator: Entering
> directly-translatable composite transform:
> 'WordCount.CountWords/Count.PerElement/Combine.perKey(Count)'
> 17/08/03 15:18:42 INFO spark.SparkRunner$Evaluator: Entering
> directly-translatable composite transform:
> 'WriteCounts/WriteFiles/View.AsIterable'
> 17/08/03 15:18:42 INFO spark.SparkRunner$Evaluator: Entering
> directly-translatable composite transform:
> 'WriteCounts/WriteFiles/Create.Values'
> 17/08/03 15:18:42 INFO metrics.MetricsAccumulator: Instantiated metrics
> accumulator:
> org.apache.beam.runners.core.metrics.MetricsContainerStepMap@2b1a901d
> 17/08/03 15:18:42 INFO aggregators.AggregatorsAccumulator: Instantiated
> aggregators accumulator:
> 17/08/03 15:18:42 INFO spark.SparkRunner$Evaluator: Evaluating
> Read(CompressedSource)
> 17/08/03 15:18:42 INFO spark.SparkRunner$Evaluator: Evaluating
> ParMultiDo(ExtractWords)
> {code}
> *YARN Job jog*
> {code}
> 17/08/03 13:00:33 INFO client.RMProxy: Connecting to ResourceManager at
> /0.0.0.0:8030
> 17/08/03 13:00:33 INFO yarn.YarnRMClient: Registering the ApplicationMaster
> 17/08/03 13:00:34 INFO yarn.YarnAllocator: Will request 2 executor
> container(s), each with 1 core(s) and 2432 MB memory (including 384 MB of
> overhead)
> 17/08/03 13:00:34 INFO yarn.YarnAllocator: Submitted 2 unlocalized container
> requests.
> 17/08/03 13:00:34 INFO yarn.ApplicationMaster: Started progress reporter
> thread with (heartbeat : 3000, initial allocation : 200) intervals
> 17/08/03 13:00:35 INFO impl.AMRMClientImpl: Received new token for :
> 192.168.0.7:50173
> 17/08/03 13:00:35 INFO yarn.YarnAllocator: Launching container
> container_1501744514957_0003_01_000002 on host 192.168.0.7
> 17/08/03 13:00:35 INFO yarn.YarnAllocator: Received 1 containers from YARN,
> launching executors on 1 of them.
> 17/08/03 13:00:35 INFO impl.ContainerManagementProtocolProxy:
> yarn.client.max-cached-nodemanagers-proxies : 0
> 17/08/03 13:00:35 INFO impl.ContainerManagementProtocolProxy: Opening proxy :
> 192.168.0.7:50173
> 17/08/03 13:00:37 INFO yarn.YarnAllocator: Launching container
> container_1501744514957_0003_01_000003 on host 192.168.0.7
> 17/08/03 13:00:37 INFO yarn.YarnAllocator: Received 1 containers from YARN,
> launching executors on 1 of them.
> 17/08/03 13:00:37 INFO impl.ContainerManagementProtocolProxy:
> yarn.client.max-cached-nodemanagers-proxies : 0
> 17/08/03 13:00:37 INFO impl.ContainerManagementProtocolProxy: Opening proxy :
> 192.168.0.7:50173
> {code}
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)