Please check the Yarn AM log to see why AM is failed to start. That's the reason why using `sc` will get such complaint.
On Fri, Dec 18, 2015 at 4:25 AM, Eran Witkon <eranwit...@gmail.com> wrote: > Hi, > I am trying to install spark 1.5.2 on Apache hadoop 2.6 and Hive and yarn > > spark-env.sh > export HADOOP_CONF_DIR=/usr/local/hadoop/etc/hadoop > > bash_profile > #HADOOP VARIABLES START > export JAVA_HOME=/usr/lib/jvm/java-8-oracle/ > export HADOOP_INSTALL=/usr/local/hadoop > export PATH=$PATH:$HADOOP_INSTALL/bin > export PATH=$PATH:$HADOOP_INSTALL/sbin > export HADOOP_MAPRED_HOME=$HADOOP_INSTALL > export HADOOP_COMMON_HOME=$HADOOP_INSTALL > export HADOOP_HDFS_HOME=$HADOOP_INSTALL > export YARN_HOME=$HADOOP_INSTALL > export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_INSTALL/lib/native > export HADOOP_OPTS="-Djava.library.path=$HADOOP_INSTALL/lib" > export HADOOP_USER_CLASSPATH_FIRST=true > export HADOOP_CONF_DIR=/usr/local/hadoop/etc/hadoop > export YARN_CONF_DIR=/usr/local/hadoop/etc/hadoop > #HADOOP VARIABLES END > > export SPARK_HOME=/usr/local/spark > export HIVE_HOME=/usr/local/hive > export PATH=$PATH:$HIVE_HOME/bin > > > When I run spark-shell > ./bin/spark-shell --master yarn-client > > Output: > 15/12/17 22:22:07 WARN util.NativeCodeLoader: Unable to load native-hadoop > library for your platform... using builtin-java classes where applicable > 15/12/17 22:22:07 INFO spark.SecurityManager: Changing view acls to: hduser > 15/12/17 22:22:07 INFO spark.SecurityManager: Changing modify acls to: > hduser > 15/12/17 22:22:07 INFO spark.SecurityManager: SecurityManager: > authentication disabled; ui acls disabled; users with view permissions: > Set(hduser); users with modify permissions: Set(hduser) > 15/12/17 22:22:07 INFO spark.HttpServer: Starting HTTP Server > 15/12/17 22:22:07 INFO server.Server: jetty-8.y.z-SNAPSHOT > 15/12/17 22:22:08 INFO server.AbstractConnector: Started > SocketConnector@0.0.0.0:38389 > 15/12/17 22:22:08 INFO util.Utils: Successfully started service 'HTTP > class server' on port 38389. > Welcome to > ____ __ > / __/__ ___ _____/ /__ > _\ \/ _ \/ _ `/ __/ '_/ > /___/ .__/\_,_/_/ /_/\_\ version 1.5.2 > /_/ > > Using Scala version 2.10.4 (Java HotSpot(TM) 64-Bit Server VM, Java > 1.8.0_66) > Type in expressions to have them evaluated. > Type :help for more information. > 15/12/17 22:22:11 WARN util.Utils: Your hostname, eranw-Lenovo-Yoga-2-Pro > resolves to a loopback address: 127.0.1.1; using 10.0.0.1 instead (on > interface wlp1s0) > 15/12/17 22:22:11 WARN util.Utils: Set SPARK_LOCAL_IP if you need to bind > to another address > 15/12/17 22:22:11 INFO spark.SparkContext: Running Spark version 1.5.2 > 15/12/17 22:22:11 INFO spark.SecurityManager: Changing view acls to: hduser > 15/12/17 22:22:11 INFO spark.SecurityManager: Changing modify acls to: > hduser > 15/12/17 22:22:11 INFO spark.SecurityManager: SecurityManager: > authentication disabled; ui acls disabled; users with view permissions: > Set(hduser); users with modify permissions: Set(hduser) > 15/12/17 22:22:11 INFO slf4j.Slf4jLogger: Slf4jLogger started > 15/12/17 22:22:11 INFO Remoting: Starting remoting > 15/12/17 22:22:12 INFO util.Utils: Successfully started service > 'sparkDriver' on port 36381. > 15/12/17 22:22:12 INFO Remoting: Remoting started; listening on addresses > :[akka.tcp://sparkDriver@10.0.0.1:36381] > 15/12/17 22:22:12 INFO spark.SparkEnv: Registering MapOutputTracker > 15/12/17 22:22:12 INFO spark.SparkEnv: Registering BlockManagerMaster > 15/12/17 22:22:12 INFO storage.DiskBlockManager: Created local directory > at /tmp/blockmgr-139fac31-5f21-4c61-9575-3110d5205f7d > 15/12/17 22:22:12 INFO storage.MemoryStore: MemoryStore started with > capacity 530.0 MB > 15/12/17 22:22:12 INFO spark.HttpFileServer: HTTP File server directory is > /tmp/spark-955ef002-a802-49c6-b440-0656861f737c/httpd-2127cbe1-97d7-40a5-a96f-75216f115f00 > 15/12/17 22:22:12 INFO spark.HttpServer: Starting HTTP Server > 15/12/17 22:22:12 INFO server.Server: jetty-8.y.z-SNAPSHOT > 15/12/17 22:22:12 INFO server.AbstractConnector: Started > SocketConnector@0.0.0.0:36760 > 15/12/17 22:22:12 INFO util.Utils: Successfully started service 'HTTP file > server' on port 36760. > 15/12/17 22:22:12 INFO spark.SparkEnv: Registering OutputCommitCoordinator > 15/12/17 22:22:12 INFO server.Server: jetty-8.y.z-SNAPSHOT > 15/12/17 22:22:12 INFO server.AbstractConnector: Started > SelectChannelConnector@0.0.0.0:4040 > 15/12/17 22:22:12 INFO util.Utils: Successfully started service 'SparkUI' > on port 4040. > 15/12/17 22:22:12 INFO ui.SparkUI: Started SparkUI at http://10.0.0.1:4040 > 15/12/17 22:22:12 WARN metrics.MetricsSystem: Using default name > DAGScheduler for source because spark.app.id is not set. > 15/12/17 22:22:12 INFO client.RMProxy: Connecting to ResourceManager at / > 0.0.0.0:8032 > 15/12/17 22:22:12 INFO yarn.Client: Requesting a new application from > cluster with 1 NodeManagers > 15/12/17 22:22:12 INFO yarn.Client: Verifying our application has not > requested more than the maximum memory capability of the cluster (8192 MB > per container) > 15/12/17 22:22:12 INFO yarn.Client: Will allocate AM container, with 896 > MB memory including 384 MB overhead > 15/12/17 22:22:12 INFO yarn.Client: Setting up container launch context > for our AM > 15/12/17 22:22:12 INFO yarn.Client: Setting up the launch environment for > our AM container > 15/12/17 22:22:12 INFO yarn.Client: Preparing resources for our AM > container > 15/12/17 22:22:13 INFO yarn.Client: Uploading resource > file:/usr/local/spark/lib/spark-assembly-1.5.2-hadoop2.6.0.jar -> > hdfs://localhost:54310/user/hduser/.sparkStaging/application_1450383657924_0003/spark-assembly-1.5.2-hadoop2.6.0.jar > 15/12/17 22:22:15 INFO yarn.Client: Uploading resource > file:/tmp/spark-955ef002-a802-49c6-b440-0656861f737c/__spark_conf__3605678380745697769.zip > -> > hdfs://localhost:54310/user/hduser/.sparkStaging/application_1450383657924_0003/__spark_conf__3605678380745697769.zip > 15/12/17 22:22:15 INFO spark.SecurityManager: Changing view acls to: hduser > 15/12/17 22:22:15 INFO spark.SecurityManager: Changing modify acls to: > hduser > 15/12/17 22:22:15 INFO spark.SecurityManager: SecurityManager: > authentication disabled; ui acls disabled; users with view permissions: > Set(hduser); users with modify permissions: Set(hduser) > 15/12/17 22:22:15 INFO yarn.Client: Submitting application 3 to > ResourceManager > 15/12/17 22:22:15 INFO impl.YarnClientImpl: Submitted application > application_1450383657924_0003 > 15/12/17 22:22:16 INFO yarn.Client: Application report for > application_1450383657924_0003 (state: ACCEPTED) > 15/12/17 22:22:16 INFO yarn.Client: > client token: N/A > diagnostics: N/A > ApplicationMaster host: N/A > ApplicationMaster RPC port: -1 > queue: default > start time: 1450383735512 > final status: UNDEFINED > tracking URL: > http://eranw-Lenovo-Yoga-2-Pro:8088/proxy/application_1450383657924_0003/ > user: hduser > 15/12/17 22:22:17 INFO yarn.Client: Application report for > application_1450383657924_0003 (state: ACCEPTED) > 15/12/17 22:22:18 INFO yarn.Client: Application report for > application_1450383657924_0003 (state: ACCEPTED) > 15/12/17 22:22:19 INFO yarn.Client: Application report for > application_1450383657924_0003 (state: ACCEPTED) > 15/12/17 22:22:20 INFO yarn.Client: Application report for > application_1450383657924_0003 (state: ACCEPTED) > 15/12/17 22:22:21 INFO cluster.YarnSchedulerBackend$YarnSchedulerEndpoint: > ApplicationMaster registered as AkkaRpcEndpointRef(Actor[akka.tcp:// > sparkYarnAM@10.0.0.1:34719/user/YarnAM#-313490512]) > 15/12/17 22:22:21 INFO cluster.YarnClientSchedulerBackend: Add WebUI > Filter. org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter, > Map(PROXY_HOSTS -> eranw-Lenovo-Yoga-2-Pro, PROXY_URI_BASES -> > http://eranw-Lenovo-Yoga-2-Pro:8088/proxy/application_1450383657924_0003), > /proxy/application_1450383657924_0003 > 15/12/17 22:22:21 INFO ui.JettyUtils: Adding filter: > org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter > 15/12/17 22:22:21 INFO yarn.Client: Application report for > application_1450383657924_0003 (state: RUNNING) > 15/12/17 22:22:21 INFO yarn.Client: > client token: N/A > diagnostics: N/A > ApplicationMaster host: 10.0.0.1 > ApplicationMaster RPC port: 0 > queue: default > start time: 1450383735512 > final status: UNDEFINED > tracking URL: > http://eranw-Lenovo-Yoga-2-Pro:8088/proxy/application_1450383657924_0003/ > user: hduser > 15/12/17 22:22:21 INFO cluster.YarnClientSchedulerBackend: Application > application_1450383657924_0003 has started running. > 15/12/17 22:22:21 INFO util.Utils: Successfully started service > 'org.apache.spark.network.netty.NettyBlockTransferService' on port 33155. > 15/12/17 22:22:21 INFO netty.NettyBlockTransferService: Server created on > 33155 > 15/12/17 22:22:21 INFO storage.BlockManagerMaster: Trying to register > BlockManager > 15/12/17 22:22:21 INFO storage.BlockManagerMasterEndpoint: Registering > block manager 10.0.0.1:33155 with 530.0 MB RAM, BlockManagerId(driver, > 10.0.0.1, 33155) > 15/12/17 22:22:21 INFO storage.BlockManagerMaster: Registered BlockManager > 15/12/17 22:22:22 WARN cluster.YarnSchedulerBackend$YarnSchedulerEndpoint: > ApplicationMaster has disassociated: 10.0.0.1:34719 > 15/12/17 22:22:22 WARN remote.ReliableDeliverySupervisor: Association with > remote system [akka.tcp://sparkYarnAM@10.0.0.1:34719] has failed, address > is now gated for [5000] ms. Reason: [Disassociated] > 15/12/17 22:22:22 INFO scheduler.EventLoggingListener: Logging events to > hdfs://localhost:54310/spark-log/application_1450383657924_0003 > 15/12/17 22:22:22 WARN cluster.YarnSchedulerBackend$YarnSchedulerEndpoint: > ApplicationMaster has disassociated: 10.0.0.1:34719 > 15/12/17 22:22:25 INFO cluster.YarnSchedulerBackend$YarnSchedulerEndpoint: > ApplicationMaster registered as AkkaRpcEndpointRef(Actor[akka.tcp:// > sparkYarnAM@10.0.0.1:39353/user/YarnAM#1861872417]) > 15/12/17 22:22:25 INFO cluster.YarnClientSchedulerBackend: Add WebUI > Filter. org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter, > Map(PROXY_HOSTS -> eranw-Lenovo-Yoga-2-Pro, PROXY_URI_BASES -> > http://eranw-Lenovo-Yoga-2-Pro:8088/proxy/application_1450383657924_0003), > /proxy/application_1450383657924_0003 > 15/12/17 22:22:25 INFO ui.JettyUtils: Adding filter: > org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter > 15/12/17 22:22:30 WARN cluster.YarnSchedulerBackend$YarnSchedulerEndpoint: > ApplicationMaster has disassociated: 10.0.0.1:39353 > 15/12/17 22:22:30 WARN remote.ReliableDeliverySupervisor: Association with > remote system [akka.tcp://sparkYarnAM@10.0.0.1:39353] has failed, address > is now gated for [5000] ms. Reason: [Disassociated] > 15/12/17 22:22:30 WARN cluster.YarnSchedulerBackend$YarnSchedulerEndpoint: > ApplicationMaster has disassociated: 10.0.0.1:39353 > 15/12/17 22:22:30 ERROR cluster.YarnClientSchedulerBackend: Yarn > application has already exited with state FINISHED! > 15/12/17 22:22:30 INFO handler.ContextHandler: stopped > o.s.j.s.ServletContextHandler{/metrics/json,null} > 15/12/17 22:22:30 INFO handler.ContextHandler: stopped > o.s.j.s.ServletContextHandler{/stages/stage/kill,null} > 15/12/17 22:22:30 INFO handler.ContextHandler: stopped > o.s.j.s.ServletContextHandler{/api,null} > 15/12/17 22:22:30 INFO handler.ContextHandler: stopped > o.s.j.s.ServletContextHandler{/,null} > 15/12/17 22:22:30 INFO handler.ContextHandler: stopped > o.s.j.s.ServletContextHandler{/static,null} > 15/12/17 22:22:30 INFO handler.ContextHandler: stopped > o.s.j.s.ServletContextHandler{/executors/threadDump/json,null} > 15/12/17 22:22:30 INFO handler.ContextHandler: stopped > o.s.j.s.ServletContextHandler{/executors/threadDump,null} > 15/12/17 22:22:30 INFO handler.ContextHandler: stopped > o.s.j.s.ServletContextHandler{/executors/json,null} > 15/12/17 22:22:30 INFO handler.ContextHandler: stopped > o.s.j.s.ServletContextHandler{/executors,null} > 15/12/17 22:22:30 INFO handler.ContextHandler: stopped > o.s.j.s.ServletContextHandler{/environment/json,null} > 15/12/17 22:22:30 INFO handler.ContextHandler: stopped > o.s.j.s.ServletContextHandler{/environment,null} > 15/12/17 22:22:30 INFO handler.ContextHandler: stopped > o.s.j.s.ServletContextHandler{/storage/rdd/json,null} > 15/12/17 22:22:30 INFO handler.ContextHandler: stopped > o.s.j.s.ServletContextHandler{/storage/rdd,null} > 15/12/17 22:22:30 INFO handler.ContextHandler: stopped > o.s.j.s.ServletContextHandler{/storage/json,null} > 15/12/17 22:22:30 INFO handler.ContextHandler: stopped > o.s.j.s.ServletContextHandler{/storage,null} > 15/12/17 22:22:30 INFO handler.ContextHandler: stopped > o.s.j.s.ServletContextHandler{/stages/pool/json,null} > 15/12/17 22:22:30 INFO handler.ContextHandler: stopped > o.s.j.s.ServletContextHandler{/stages/pool,null} > 15/12/17 22:22:30 INFO handler.ContextHandler: stopped > o.s.j.s.ServletContextHandler{/stages/stage/json,null} > 15/12/17 22:22:30 INFO handler.ContextHandler: stopped > o.s.j.s.ServletContextHandler{/stages/stage,null} > 15/12/17 22:22:30 INFO handler.ContextHandler: stopped > o.s.j.s.ServletContextHandler{/stages/json,null} > 15/12/17 22:22:30 INFO handler.ContextHandler: stopped > o.s.j.s.ServletContextHandler{/stages,null} > 15/12/17 22:22:30 INFO handler.ContextHandler: stopped > o.s.j.s.ServletContextHandler{/jobs/job/json,null} > 15/12/17 22:22:30 INFO handler.ContextHandler: stopped > o.s.j.s.ServletContextHandler{/jobs/job,null} > 15/12/17 22:22:30 INFO handler.ContextHandler: stopped > o.s.j.s.ServletContextHandler{/jobs/json,null} > 15/12/17 22:22:30 INFO handler.ContextHandler: stopped > o.s.j.s.ServletContextHandler{/jobs,null} > 15/12/17 22:22:31 INFO ui.SparkUI: Stopped Spark web UI at > http://10.0.0.1:4040 > 15/12/17 22:22:31 INFO scheduler.DAGScheduler: Stopping DAGScheduler > 15/12/17 22:22:31 INFO cluster.YarnClientSchedulerBackend: Shutting down > all executors > 15/12/17 22:22:31 INFO cluster.YarnClientSchedulerBackend: Asking each > executor to shut down > 15/12/17 22:22:31 INFO cluster.YarnClientSchedulerBackend: Stopped > 15/12/17 22:22:36 INFO spark.MapOutputTrackerMasterEndpoint: > MapOutputTrackerMasterEndpoint stopped! > 15/12/17 22:22:36 INFO storage.MemoryStore: MemoryStore cleared > 15/12/17 22:22:36 INFO storage.BlockManager: BlockManager stopped > 15/12/17 22:22:36 INFO storage.BlockManagerMaster: BlockManagerMaster > stopped > 15/12/17 22:22:36 INFO spark.SparkContext: Successfully stopped > SparkContext > 15/12/17 22:22:36 INFO > scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: > OutputCommitCoordinator stopped! > 15/12/17 22:22:36 INFO remote.RemoteActorRefProvider$RemotingTerminator: > Shutting down remote daemon. > 15/12/17 22:22:36 INFO remote.RemoteActorRefProvider$RemotingTerminator: > Remote daemon shut down; proceeding with flushing remote transports. > 15/12/17 22:22:36 INFO remote.RemoteActorRefProvider$RemotingTerminator: > Remoting shut down. > 15/12/17 22:22:42 INFO cluster.YarnClientSchedulerBackend: > SchedulerBackend is ready for scheduling beginning after waiting > maxRegisteredResourcesWaitingTime: 30000(ms) > 15/12/17 22:22:42 INFO repl.SparkILoop: Created spark context.. > Spark context available as sc. > 15/12/17 22:22:49 INFO hive.HiveContext: Initializing execution hive, > version 1.2.1 > 15/12/17 22:22:49 INFO client.ClientWrapper: Inspected Hadoop version: > 2.6.0 > 15/12/17 22:22:49 INFO client.ClientWrapper: Loaded > org.apache.hadoop.hive.shims.Hadoop23Shims for Hadoop version 2.6.0 > 15/12/17 22:22:49 INFO metastore.HiveMetaStore: 0: Opening raw store with > implemenation class:org.apache.hadoop.hive.metastore.ObjectStore > 15/12/17 22:22:49 INFO metastore.ObjectStore: ObjectStore, initialize > called > 15/12/17 22:22:50 INFO DataNucleus.Persistence: Property > hive.metastore.integral.jdo.pushdown unknown - will be ignored > 15/12/17 22:22:50 INFO DataNucleus.Persistence: Property > datanucleus.cache.level2 unknown - will be ignored > 15/12/17 22:22:50 WARN DataNucleus.Connection: BoneCP specified but not > present in CLASSPATH (or one of dependencies) > 15/12/17 22:22:50 WARN DataNucleus.Connection: BoneCP specified but not > present in CLASSPATH (or one of dependencies) > 15/12/17 22:22:59 INFO metastore.ObjectStore: Setting MetaStore object pin > classes with > hive.metastore.cache.pinobjtypes="Table,StorageDescriptor,SerDeInfo,Partition,Database,Type,FieldSchema,Order" > 15/12/17 22:23:00 INFO DataNucleus.Datastore: The class > "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as > "embedded-only" so does not have its own datastore table. > 15/12/17 22:23:00 INFO DataNucleus.Datastore: The class > "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as > "embedded-only" so does not have its own datastore table. > 15/12/17 22:23:01 INFO DataNucleus.Datastore: The class > "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as > "embedded-only" so does not have its own datastore table. > 15/12/17 22:23:01 INFO DataNucleus.Datastore: The class > "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as > "embedded-only" so does not have its own datastore table. > 15/12/17 22:23:02 INFO metastore.MetaStoreDirectSql: Using direct SQL, > underlying DB is DERBY > 15/12/17 22:23:02 INFO metastore.ObjectStore: Initialized ObjectStore > 15/12/17 22:23:02 WARN metastore.ObjectStore: Version information not > found in metastore. hive.metastore.schema.verification is not enabled so > recording the schema version 1.2.0 > 15/12/17 22:23:02 WARN metastore.ObjectStore: Failed to get database > default, returning NoSuchObjectException > 15/12/17 22:23:02 INFO metastore.HiveMetaStore: Added admin role in > metastore > 15/12/17 22:23:02 INFO metastore.HiveMetaStore: Added public role in > metastore > 15/12/17 22:23:03 INFO metastore.HiveMetaStore: No user is added in admin > role, since config is empty > 15/12/17 22:23:03 INFO metastore.HiveMetaStore: 0: get_all_databases > 15/12/17 22:23:03 INFO HiveMetaStore.audit: ugi=hduser ip=unknown-ip-addr > cmd=get_all_databases > 15/12/17 22:23:03 INFO metastore.HiveMetaStore: 0: get_functions: > db=default pat=* > 15/12/17 22:23:03 INFO HiveMetaStore.audit: ugi=hduser ip=unknown-ip-addr > cmd=get_functions: > db=default pat=* > 15/12/17 22:23:03 INFO DataNucleus.Datastore: The class > "org.apache.hadoop.hive.metastore.model.MResourceUri" is tagged as > "embedded-only" so does not have its own datastore table. > 15/12/17 22:23:09 INFO session.SessionState: Created local directory: > /tmp/eae419f8-7dad-417b-bc3e-77c048c88337_resources > 15/12/17 22:23:09 INFO session.SessionState: Created HDFS directory: > /tmp/hive/hduser/eae419f8-7dad-417b-bc3e-77c048c88337 > 15/12/17 22:23:09 INFO session.SessionState: Created local directory: > /tmp/hduser/eae419f8-7dad-417b-bc3e-77c048c88337 > 15/12/17 22:23:09 INFO session.SessionState: Created HDFS directory: > /tmp/hive/hduser/eae419f8-7dad-417b-bc3e-77c048c88337/_tmp_space.db > 15/12/17 22:23:09 INFO hive.HiveContext: default warehouse location is > /user/hive/warehouse > 15/12/17 22:23:09 INFO hive.HiveContext: Initializing > HiveMetastoreConnection version 1.2.1 using Spark classes. > 15/12/17 22:23:09 INFO client.ClientWrapper: Inspected Hadoop version: > 2.6.0 > 15/12/17 22:23:09 INFO client.ClientWrapper: Loaded > org.apache.hadoop.hive.shims.Hadoop23Shims for Hadoop version 2.6.0 > 15/12/17 22:23:10 WARN util.NativeCodeLoader: Unable to load native-hadoop > library for your platform... using builtin-java classes where applicable > 15/12/17 22:23:10 INFO metastore.HiveMetaStore: 0: Opening raw store with > implemenation class:org.apache.hadoop.hive.metastore.ObjectStore > 15/12/17 22:23:10 INFO metastore.ObjectStore: ObjectStore, initialize > called > 15/12/17 22:23:10 INFO DataNucleus.Persistence: Property > hive.metastore.integral.jdo.pushdown unknown - will be ignored > 15/12/17 22:23:10 INFO DataNucleus.Persistence: Property > datanucleus.cache.level2 unknown - will be ignored > 15/12/17 22:23:10 WARN DataNucleus.Connection: BoneCP specified but not > present in CLASSPATH (or one of dependencies) > 15/12/17 22:23:10 WARN DataNucleus.Connection: BoneCP specified but not > present in CLASSPATH (or one of dependencies) > 15/12/17 22:23:11 INFO metastore.ObjectStore: Setting MetaStore object pin > classes with > hive.metastore.cache.pinobjtypes="Table,StorageDescriptor,SerDeInfo,Partition,Database,Type,FieldSchema,Order" > 15/12/17 22:23:12 INFO DataNucleus.Datastore: The class > "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as > "embedded-only" so does not have its own datastore table. > 15/12/17 22:23:12 INFO DataNucleus.Datastore: The class > "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as > "embedded-only" so does not have its own datastore table. > 15/12/17 22:23:12 INFO DataNucleus.Datastore: The class > "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as > "embedded-only" so does not have its own datastore table. > 15/12/17 22:23:12 INFO DataNucleus.Datastore: The class > "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as > "embedded-only" so does not have its own datastore table. > 15/12/17 22:23:12 INFO DataNucleus.Query: Reading in results for query > "org.datanucleus.store.rdbms.query.SQLQuery@0" since the connection used > is closing > 15/12/17 22:23:12 INFO metastore.MetaStoreDirectSql: Using direct SQL, > underlying DB is DERBY > 15/12/17 22:23:12 INFO metastore.ObjectStore: Initialized ObjectStore > 15/12/17 22:23:13 INFO metastore.HiveMetaStore: Added admin role in > metastore > 15/12/17 22:23:13 INFO metastore.HiveMetaStore: Added public role in > metastore > 15/12/17 22:23:13 INFO metastore.HiveMetaStore: No user is added in admin > role, since config is empty > 15/12/17 22:23:13 INFO metastore.HiveMetaStore: 0: get_all_databases > 15/12/17 22:23:13 INFO HiveMetaStore.audit: ugi=hduser ip=unknown-ip-addr > cmd=get_all_databases > 15/12/17 22:23:13 INFO metastore.HiveMetaStore: 0: get_functions: > db=default pat=* > 15/12/17 22:23:13 INFO HiveMetaStore.audit: ugi=hduser ip=unknown-ip-addr > cmd=get_functions: > db=default pat=* > 15/12/17 22:23:13 INFO DataNucleus.Datastore: The class > "org.apache.hadoop.hive.metastore.model.MResourceUri" is tagged as > "embedded-only" so does not have its own datastore table. > 15/12/17 22:23:18 INFO session.SessionState: Created local directory: > /tmp/7f892b9d-38ef-4fab-bc44-c1afaf07f50f_resources > 15/12/17 22:23:18 INFO session.SessionState: Created HDFS directory: > /tmp/hive/hduser/7f892b9d-38ef-4fab-bc44-c1afaf07f50f > 15/12/17 22:23:18 INFO session.SessionState: Created local directory: > /tmp/hduser/7f892b9d-38ef-4fab-bc44-c1afaf07f50f > 15/12/17 22:23:18 INFO session.SessionState: Created HDFS directory: > /tmp/hive/hduser/7f892b9d-38ef-4fab-bc44-c1afaf07f50f/_tmp_space.db > 15/12/17 22:23:18 INFO repl.SparkILoop: Created sql context (with Hive > support).. > SQL context available as sqlContext. > > scala> > > When I am running the following code: > val data = sc.wholeTextFiles("hdfs://localhost:54310/spark-log/*") > > I am getting exception: > java.lang.IllegalStateException: Cannot call methods on a stopped > SparkContext > at org.apache.spark.SparkContext.org > $apache$spark$SparkContext$$assertNotStopped(SparkContext.scala:104) > at > org.apache.spark.SparkContext.defaultParallelism(SparkContext.scala:2063) > at > org.apache.spark.SparkContext.defaultMinPartitions(SparkContext.scala:2076) > at > org.apache.spark.SparkContext.wholeTextFiles$default$2(SparkContext.scala:864) > at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:21) > at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:26) > at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:28) > at $iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:30) > at $iwC$$iwC$$iwC$$iwC.<init>(<console>:32) > at $iwC$$iwC$$iwC.<init>(<console>:34) > at $iwC$$iwC.<init>(<console>:36) > at $iwC.<init>(<console>:38) > at <init>(<console>:40) > at .<init>(<console>:44) > at .<clinit>(<console>) > at .<init>(<console>:7) > at .<clinit>(<console>) > at $print(<console>) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:497) > at > org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:1065) > at > org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkIMain.scala:1340) > at org.apache.spark.repl.SparkIMain.loadAndRunReq$1(SparkIMain.scala:840) > at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:871) > at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:819) > at org.apache.spark.repl.SparkILoop.reallyInterpret$1(SparkILoop.scala:857) > at > org.apache.spark.repl.SparkILoop.interpretStartingWith(SparkILoop.scala:902) > at org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:814) > at org.apache.spark.repl.SparkILoop.processLine$1(SparkILoop.scala:657) > at org.apache.spark.repl.SparkILoop.innerLoop$1(SparkILoop.scala:665) > at org.apache.spark.repl.SparkILoop.org > $apache$spark$repl$SparkILoop$$loop(SparkILoop.scala:670) > at > org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply$mcZ$sp(SparkILoop.scala:997) > at > org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945) > at > org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945) > at > scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:135) > at org.apache.spark.repl.SparkILoop.org > $apache$spark$repl$SparkILoop$$process(SparkILoop.scala:945) > at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:1059) > at org.apache.spark.repl.Main$.main(Main.scala:31) > at org.apache.spark.repl.Main.main(Main.scala) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:497) > at > org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:674) > at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:180) > at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:205) > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:120) > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > > *What is wrong here? how should I configure spark to use yarn and hive > correctly?* > > Eran > >