And you can easily make a local repo and then share it via httpd. In this
case, it can be served to anywhere.

Cos

On Fri, Feb 19, 2016 at 10:15AM, Evans Ye wrote:
> You can make a local apt repo anytime. That should do the trick for 1 node
> cluster if you just want to test the feature.
> 
> 2016-02-19 2:59 GMT+08:00 MrAsanjar . <[email protected]>:
> 
> >  Evans, we couldn't  use BigTop Provisioner since  ppc artifacts are not in
> > the repository yet. That is why we had to made the script. Thanks
> >
> >
> > On Thu, Feb 18, 2016 at 1:40 PM, Konstantin Boudnik <[email protected]>
> > wrote:
> >
> > > On Fri, Feb 19, 2016 at 02:07AM, Evans Ye wrote:
> > > > Have you tried Bigtop Provisioner?
> > > >
> > >
> > https://cwiki.apache.org/confluence/display/BIGTOP/Bigtop+Provisioner+User+Guide
> > > >
> > > > -----
> > > > $ cd bigtop/bigtop-deploy/vm/vagrant-puppet-vm
> > > > $ cat vagrantconfig.yaml  # You just need to update repo and add spark
> > > into
> > > > component list
> > > > memory_size: 4096
> > > > number_cpus: 1
> > > > box: "puppetlabs/centos-7.0-64-nocm"
> > > > repo: "
> > > http://bigtop-repos.s3.amazonaws.com/releases/1.1.0/centos/7/x86_64";
> > > > num_instances: 1
> > > > distro: centos
> > > > components: [hadoop, yarn, spark]
> > > > enable_local_repo: false
> > > > run_smoke_tests: false
> > > > smoke_test_components: [mapreduce, pig]
> > > > jdk: "java-1.7.0-openjdk-devel.x86_64"
> > > >
> > > > $ ./docker-hadoop.sh -c 1   # wait for 5mins
> > > > $ vagrant ssh bigtop1
> > > > $ cd /usr/lib/spark
> > > > $ spark-submit --class org.apache.spark.examples.SparkPi --master
> > > > yarn-client /usr/lib/spark/lib/spark-examples-1.5.1-hadoop2.7.1.jar 10
> > > > ...
> > > > Pi is roughly 3.144728
> > >
> > > Good enough for 1.1.0 release ;)
> > >
> > > But seriously: we have a real good tool to provision clusters quickly and
> > > painlessly. And with Puppet to guarantee the consistency.
> > >
> > > Cos
> > >
> > > > -----
> > > >
> > > > Is this what you want?
> > > > I can run this either on Docker or on a CentOS 7 VM.
> > > > Sorry I don't have PPC machine to test.
> > > >
> > > >
> > > > 2016-02-18 11:46 GMT+08:00 MrAsanjar . <[email protected]>:
> > > >
> > > > > I have build a single node hadoop/spark sandbox based on the latest
> > > > > Apachue Bigtop 1.1.0 build. Spark in standalone mode + HDFS functions
> > > > > perfectly, however, fails if yarn-client/yarn-master mode is used as
> > > > > follow:
> > > > >
> > > > > *>>spark-submit --class org.apache.spark.examples.SparkPi --master
> > > > > yarn-client  /usr/lib/spark/lib/spark-examples-1.5.1-hadoop2.7.1.jar
> > > 10*
> > > > > 16/02/17 05:19:52 ERROR YarnClientSchedulerBackend: Yarn application
> > > has
> > > > > already exited with state FINISHED!
> > > > > Exception in thread "main" java.lang.IllegalStateException: Cannot
> > call
> > > > > methods on a stopped SparkContext
> > > > >     at org.apache.spark.SparkContext.org
> > > > > $apache$spark$SparkContext$$assertNotStopped(SparkContext.scala:104)
> > > > >     ......
> > > > >
> > > > >
> > > > > Looking at yarn Application log file, there is a *RECEIVED SIGNAL 15:
> > > > > SIGTERM *termination signal from the the yarn container.
> > > > > >>*yarn logs -applicationId application_1455683261278_0001*
> > > > >
> > > > > YARN executor launch context:
> > > > >   env:
> > > > >     CLASSPATH ->
> > > > >
> > > > >
> > >
> > {{PWD}}<CPS>{{PWD}}/__spark__.jar<CPS>$HADOOP_CONF_DIR<CPS>$HADOOP_COMMON_HOME/*<CPS>$HADOOP_COMMON_HOME/lib/*<CPS>$HADOOP_HDFS_HOME/*<CPS>$HADOOP_HDFS_HOME/lib/*<CPS>$HADOOP_MAPRED_HOME/*<CPS>$HADOOP_MAPRED_HOME/lib/*<CPS>$HADOOP_YARN_HOME/*<CPS>$HADOOP_YARN_HOME/lib/*<CPS>$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/*<CPS>$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/lib/*
> > > > >     SPARK_LOG_URL_STDERR ->
> > > http://2657cd5dc2f6:8042/node/containerlogs
> > > > >
> > > > >
> > >
> > ===============================================================================
> > > > >
> > > > > 16/02/17 04:29:18 INFO impl.ContainerManagementProtocolProxy: Opening
> > > proxy
> > > > > : 2657cd5dc2f6:33785
> > > > > 16/02/17 04:29:18 ERROR yarn.ApplicationMaster: *RECEIVED SIGNAL 15:
> > > > > SIGTERM*
> > > > > 16/02/17 04:29:18 INFO yarn.ApplicationMaster: Final app status:
> > > UNDEFINED,
> > > > > exitCode: 0, (reason: Shutdown hook called before final status was
> > > > > reported.)
> > > > > 16/02/17 04:29:18 INFO yarn.ApplicationMaster: Unregistering
> > > > > ApplicationMaster with UNDEFINED (diag message: Shutdown hook called
> > > before
> > > > > final status was reported.)
> > > > > 16/02/17 04:29:18 INFO impl.AMRMClientImpl: Waiting for application
> > to
> > > be
> > > > > successfully unregistered.
> > > > > 16/02/17 04:29:18 INFO yarn.ApplicationMaster: Deleting staging
> > > directory
> > > > > .sparkStaging/application_1455683261278_0001
> > > > > 16/02/17 04:29:18 INFO util.ShutdownHookManager: Shutdown hook called
> > > > > End of LogType:stderr
> > > > >
> > > > >
> > > > > BTW, I have successfully tested hadoop yarn by running
> > Teragen/Terasort
> > > > > mapreduce job.
> > > > > Before i start debugging, has anyone tested spark in yarn-client
> > mode?
> > > > >
> > >
> >

Attachment: signature.asc
Description: Digital signature

Reply via email to