Same thing happens. On 11 Aug 2014, at 21:05, Yan Fang <[email protected]> wrote:
> Cool, we are almost there. Could you remove > > <property> > <name>fs.hdfs.impl</name> > <value>org.apache.hadoop.hdfs. > DistributedFileSystem</value> > <description>The FileSystem for hdfs: uris.</description> > </property> > > To see how it works? > > > Fang, Yan > [email protected] > +1 (206) 849-4108 > > > On Mon, Aug 11, 2014 at 5:03 PM, Telles Nobrega <[email protected]> > wrote: > >> You may forget this last email, I was really stupid and put the files in a >> different folder. Now it could find the file but it’s not there yet… >> another error came up >> >> Exception in thread "main" java.util.ServiceConfigurationError: >> org.apache.hadoop.fs.FileSystem: Provider >> org.apache.hadoop.hdfs.DistributedFileSystem could not be instantiated >> at java.util.ServiceLoader.fail(ServiceLoader.java:224) >> at java.util.ServiceLoader.access$100(ServiceLoader.java:181) >> at >> java.util.ServiceLoader$LazyIterator.next(ServiceLoader.java:377) >> at java.util.ServiceLoader$1.next(ServiceLoader.java:445) >> at >> org.apache.hadoop.fs.FileSystem.loadFileSystems(FileSystem.java:2400) >> at >> org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2411) >> at >> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2428) >> at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88) >> at >> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2467) >> at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449) >> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367) >> at org.apache.hadoop.fs.Path.getFileSystem(Path.java:287) >> at >> org.apache.samza.job.yarn.ClientHelper.submitApplication(ClientHelper.scala:111) >> at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:55) >> at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:48) >> at org.apache.samza.job.JobRunner.run(JobRunner.scala:62) >> at org.apache.samza.job.JobRunner$.main(JobRunner.scala:37) >> at org.apache.samza.job.JobRunner.main(JobRunner.scala) >> Caused by: java.lang.NoClassDefFoundError: >> org/apache/hadoop/conf/Configuration$DeprecationDelta >> at >> org.apache.hadoop.hdfs.HdfsConfiguration.addDeprecatedKeys(HdfsConfiguration.java:66) >> at >> org.apache.hadoop.hdfs.HdfsConfiguration.<clinit>(HdfsConfiguration.java:31) >> at >> org.apache.hadoop.hdfs.DistributedFileSystem.<clinit>(DistributedFileSystem.java:106) >> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native >> Method) >> at >> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) >> at >> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) >> at java.lang.reflect.Constructor.newInstance(Constructor.java:526) >> at java.lang.Class.newInstance(Class.java:374) >> at >> java.util.ServiceLoader$LazyIterator.next(ServiceLoader.java:373) >> ... 15 more >> Caused by: java.lang.ClassNotFoundException: >> org.apache.hadoop.conf.Configuration$DeprecationDelta >> at java.net.URLClassLoader$1.run(URLClassLoader.java:366) >> at java.net.URLClassLoader$1.run(URLClassLoader.java:355) >> at java.security.AccessController.doPrivileged(Native Method) >> at java.net.URLClassLoader.findClass(URLClassLoader.java:354) >> at java.lang.ClassLoader.loadClass(ClassLoader.java:425) >> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) >> at java.lang.ClassLoader.loadClass(ClassLoader.java:358) >> ... 24 more >> >> On 11 Aug 2014, at 20:45, Telles Nobrega <[email protected]> wrote: >> >>> Hi, I copied hadoop-hdfs-2.3.0 to my-job/lib and it changed the error >> which is good but the error is back to >>> >>> Exception in thread "main" java.lang.RuntimeException: >> java.lang.ClassNotFoundException: Class >> org.apache.hadoop.hdfs.DistributedFileSystem not found >>> at >> org.apache.hadoop.conf.Configuration.getClass(Configuration.java:1720) >>> at >> org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2415) >>> at >> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2428) >>> at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88) >>> at >> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2467) >>> at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449) >>> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367) >>> at org.apache.hadoop.fs.Path.getFileSystem(Path.java:287) >>> at >> org.apache.samza.job.yarn.ClientHelper.submitApplication(ClientHelper.scala:111) >>> at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:55) >>> at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:48) >>> at org.apache.samza.job.JobRunner.run(JobRunner.scala:62) >>> at org.apache.samza.job.JobRunner$.main(JobRunner.scala:37) >>> at org.apache.samza.job.JobRunner.main(JobRunner.scala) >>> Caused by: java.lang.ClassNotFoundException: Class >> org.apache.hadoop.hdfs.DistributedFileSystem not found >>> at >> org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:1626) >>> at >> org.apache.hadoop.conf.Configuration.getClass(Configuration.java:1718) >>> ... 13 more >>> >>> Do I need to have this lib in all nodes at the job folder or just to >> submit? >>> >>> On 11 Aug 2014, at 20:11, Yan Fang <[email protected]> wrote: >>> >>>> Hi Telles, >>>> >>>> I replayed your problem and think I figured out why CLASSPATH does not >>>> work. Because in our script bin/run-class.sh, we have the line >>>> "CLASSPATH=$HADOOP_CONF_DIR", which actually ingores your setting. >>>> >>>> So a simple solution is to copy the hadoop-hdfs.jar to your samza lib >>>> directory. Then run bin/run-job ----config-factory=... >> --config-path=... . >>>> Let me know how it goes. Thank you. >>>> >>>> Cheers, >>>> >>>> Fang, Yan >>>> [email protected] >>>> +1 (206) 849-4108 >>>> >>>> >>>> On Mon, Aug 11, 2014 at 4:07 PM, Telles Nobrega < >> [email protected]> >>>> wrote: >>>> >>>>> Sure, thanks. >>>>> >>>>> >>>>> On Mon, Aug 11, 2014 at 6:22 PM, Yan Fang <[email protected]> >> wrote: >>>>> >>>>>> Hi Telles, >>>>>> >>>>>> I am not sure whether exporting the CLASSPATH works. (sometimes it >> does >>>>> not >>>>>> work for me...) My suggestion is to include the hdfs jar explicitly in >>>>> the >>>>>> package that you upload to hdfs. Also , remember to put the jar into >> your >>>>>> local samza (which is deploy/samza/lib if you go with the hello-samza >>>>>> tutorial) Let me know if that works. >>>>>> >>>>>> Cheers, >>>>>> >>>>>> Fang, Yan >>>>>> [email protected] >>>>>> +1 (206) 849-4108 >>>>>> >>>>>> >>>>>> On Mon, Aug 11, 2014 at 2:04 PM, Chris Riccomini < >>>>>> [email protected]> wrote: >>>>>> >>>>>>> Hey Telles, >>>>>>> >>>>>>> Hmm. I'm out of ideas. If Zhijie is around, he'd probably be of use, >>>>> but >>>>>> I >>>>>>> haven't heard from him in a while. >>>>>>> >>>>>>> I'm afraid your best bet is probably to email the YARN dev mailing >>>>> list, >>>>>>> since this is a YARN config issue. >>>>>>> >>>>>>> Cheers, >>>>>>> Chris >>>>>>> >>>>>>> On 8/11/14 1:58 PM, "Telles Nobrega" <[email protected]> >> wrote: >>>>>>> >>>>>>>> I exported export >>>>>>> >>>>>> >>>>>> >> CLASSPATH=$CLASSPATH:hadoop-2.3.0/share/hadoop/hdfs/hadoop-hdfs-2.3.0.jar >>>>>>>> and still happened the same problem. >>>>>>>> >>>>>>>> >>>>>>>> On Mon, Aug 11, 2014 at 5:35 PM, Chris Riccomini < >>>>>>>> [email protected]> wrote: >>>>>>>> >>>>>>>>> Hey Telles, >>>>>>>>> >>>>>>>>> It sounds like either the HDFS jar is missing from the classpath, >> or >>>>>> the >>>>>>>>> hdfs file system needs to be configured: >>>>>>>>> >>>>>>>>> <property> >>>>>>>>> <name>fs.hdfs.impl</name> >>>>>>>>> <value>org.apache.hadoop.hdfs.DistributedFileSystem</value> >>>>>>>>> <description>The FileSystem for hdfs: uris.</description> >>>>>>>>> </property> >>>>>>>>> >>>>>>>>> >>>>>>>>> (from >>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>> >> https://groups.google.com/a/cloudera.org/forum/#!topic/scm-users/lyho8ptA >>>>>>>>> zE >>>>>>>>> 0) >>>>>>>>> >>>>>>>>> I believe this will need to be configured for your NM. >>>>>>>>> >>>>>>>>> Cheers, >>>>>>>>> Chris >>>>>>>>> >>>>>>>>> On 8/11/14 1:31 PM, "Telles Nobrega" <[email protected]> >>>>> wrote: >>>>>>>>> >>>>>>>>>> Yes, it is like this: >>>>>>>>>> >>>>>>>>>> <configuration> >>>>>>>>>> <property> >>>>>>>>>> <name>dfs.datanode.data.dir</name> >>>>>>>>>> <value>file:///home/ubuntu/hadoop-2.3.0/hdfs/datanode</value> >>>>>>>>>> <description>Comma separated list of paths on the local >>>>>> filesystem >>>>>>>>> of >>>>>>>>>> a >>>>>>>>>> DataNode where it should store its blocks.</description> >>>>>>>>>> </property> >>>>>>>>>> >>>>>>>>>> <property> >>>>>>>>>> <name>dfs.namenode.name.dir</name> >>>>>>>>>> <value>file:///home/ubuntu/hadoop-2.3.0/hdfs/namenode</value> >>>>>>>>>> <description>Path on the local filesystem where the NameNode >>>>>> stores >>>>>>>>>> the >>>>>>>>>> namespace and transaction logs persistently.</description> >>>>>>>>>> </property> >>>>>>>>>> </configuration> >>>>>>>>>> ~ >>>>>>>>>> >>>>>>>>>> I saw some report that this may be a classpath problem. Does this >>>>>>>>> sounds >>>>>>>>>> right to you? >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Mon, Aug 11, 2014 at 5:25 PM, Yan Fang <[email protected]> >>>>>>> wrote: >>>>>>>>>> >>>>>>>>>>> Hi Telles, >>>>>>>>>>> >>>>>>>>>>> It looks correct. Did you put the hdfs-site.xml into your >>>>>>>>>>> HADOOP_CONF_DIR >>>>>>>>>>> ?(such as ~/.samza/conf) >>>>>>>>>>> >>>>>>>>>>> Fang, Yan >>>>>>>>>>> [email protected] >>>>>>>>>>> +1 (206) 849-4108 >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Mon, Aug 11, 2014 at 1:02 PM, Telles Nobrega >>>>>>>>>>> <[email protected]> >>>>>>>>>>> wrote: >>>>>>>>>>> >>>>>>>>>>>> Hi Yan Fang, >>>>>>>>>>>> >>>>>>>>>>>> I was able to deploy the file to hdfs, I can see them in all my >>>>>>>>> nodes >>>>>>>>>>> but >>>>>>>>>>>> when I tried running I got this error: >>>>>>>>>>>> >>>>>>>>>>>> Exception in thread "main" java.io.IOException: No FileSystem >>>>> for >>>>>>>>>>> scheme: >>>>>>>>>>>> hdfs >>>>>>>>>>>> at >>>>>>>>>>> >>>>>>> >>>>>> >>>>>>> >> org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2421) >>>>>>>>>>>> at >>>>>>>>>>> >>>>>>> >>>>>>> >> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2428) >>>>>>>>>>>> at >>>>> org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88) >>>>>>>>>>>> at >>>>>>>>>>> >>>>>>> >>>>>>> >> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2467) >>>>>>>>>>>> at >>>>>> org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449) >>>>>>>>>>>> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367) >>>>>>>>>>>> at org.apache.hadoop.fs.Path.getFileSystem(Path.java:287) >>>>>>>>>>>> at >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>> >> org.apache.samza.job.yarn.ClientHelper.submitApplication(ClientHelper.s >>>>>>>>>>> ca >>>>>>>>>>> la:111) >>>>>>>>>>>> at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:55) >>>>>>>>>>>> at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:48) >>>>>>>>>>>> at org.apache.samza.job.JobRunner.run(JobRunner.scala:62) >>>>>>>>>>>> at org.apache.samza.job.JobRunner$.main(JobRunner.scala:37) >>>>>>>>>>>> at org.apache.samza.job.JobRunner.main(JobRunner.scala) >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> This is my yarn.package.path config: >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>> >> yarn.package.path=hdfs://telles-master-samza:50070/samza-job-package-0 >>>>>>>>>>> .7 >>>>>>>>>>> .0-dist.tar.gz >>>>>>>>>>>> >>>>>>>>>>>> Thanks in advance >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Mon, Aug 11, 2014 at 3:00 PM, Yan Fang < >>>>> [email protected]> >>>>>>>>>>> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> Hi Telles, >>>>>>>>>>>>> >>>>>>>>>>>>> In terms of "*I tried pushing the tar file to HDFS but I got >>>>> an >>>>>>>>>>> error >>>>>>>>>>>> from >>>>>>>>>>>>> hadoop saying that it couldn’t find core-site.xml file*.", I >>>>>>>>> guess >>>>>>>>>>> you >>>>>>>>>>>> set >>>>>>>>>>>>> the HADOOP_CONF_DIR variable and made it point to >>>>>> ~/.samza/conf. >>>>>>>>> You >>>>>>>>>>> can >>>>>>>>>>>> do >>>>>>>>>>>>> 1) make the HADOOP_CONF_DIR point to the directory where your >>>>>>>>> conf >>>>>>>>>>> files >>>>>>>>>>>>> are, such as /etc/hadoop/conf. Or 2) copy the config files to >>>>>>>>>>>>> ~/.samza/conf. Thank you, >>>>>>>>>>>>> >>>>>>>>>>>>> Cheer, >>>>>>>>>>>>> >>>>>>>>>>>>> Fang, Yan >>>>>>>>>>>>> [email protected] >>>>>>>>>>>>> +1 (206) 849-4108 >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> On Mon, Aug 11, 2014 at 7:40 AM, Chris Riccomini < >>>>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> Hey Telles, >>>>>>>>>>>>>> >>>>>>>>>>>>>> To get YARN working with the HTTP file system, you need to >>>>>>>>> follow >>>>>>>>>>> the >>>>>>>>>>>>>> instructions on: >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>> >> http://samza.incubator.apache.org/learn/tutorials/0.7.0/run-in-multi-node >>>>>>>>>>> -y >>>>>>>>>>>>>> arn.html >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> In the "Set Up Http Filesystem for YARN" section. >>>>>>>>>>>>>> >>>>>>>>>>>>>> You shouldn't need to compile anything (no Gradle, which is >>>>>>>>> what >>>>>>>>>>> your >>>>>>>>>>>>>> stack trace is showing). This setup should be done for all >>>>> of >>>>>>>>> the >>>>>>>>>>> NMs, >>>>>>>>>>>>>> since they will be the ones downloading your job's package >>>>>>>>> (from >>>>>>>>>>>>>> yarn.package.path). >>>>>>>>>>>>>> >>>>>>>>>>>>>> Cheers, >>>>>>>>>>>>>> Chris >>>>>>>>>>>>>> >>>>>>>>>>>>>> On 8/9/14 9:44 PM, "Telles Nobrega" < >>>>> [email protected] >>>>>>> >>>>>>>>>>> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> Hi again, I tried installing the scala libs but the Http >>>>>>>>> problem >>>>>>>>>>> still >>>>>>>>>>>>>>> occurs. I realised that I need to compile incubator samza >>>>> in >>>>>>>>> the >>>>>>>>>>>>> machines >>>>>>>>>>>>>>> that I¹m going to run the jobs, but the compilation fails >>>>>> with >>>>>>>>>>> this >>>>>>>>>>>> huge >>>>>>>>>>>>>>> message: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> # >>>>>>>>>>>>>>> # There is insufficient memory for the Java Runtime >>>>>>>>> Environment >>>>>>>>>>> to >>>>>>>>>>>>>>> continue. >>>>>>>>>>>>>>> # Native memory allocation (malloc) failed to allocate >>>>>>>>> 3946053632 >>>>>>>>>>>> bytes >>>>>>>>>>>>>>> for committing reserved memory. >>>>>>>>>>>>>>> # An error report file with more information is saved as: >>>>>>>>>>>>>>> # >>>>>> /home/ubuntu/incubator-samza/samza-kafka/hs_err_pid2506.log >>>>>>>>>>>>>>> Could not write standard input into: Gradle Worker 13. >>>>>>>>>>>>>>> java.io.IOException: Broken pipe >>>>>>>>>>>>>>> at java.io.FileOutputStream.writeBytes(Native >>>>> Method) >>>>>>>>>>>>>>> at >>>>>>>>>>> java.io.FileOutputStream.write(FileOutputStream.java:345) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>> >>>>>>> >>>>>>> >> java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>> java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.process.internal.streams.ExecOutputHandleRunner.run(ExecOut >>>>>>>>>>>> pu >>>>>>>>>>>> tH >>>>>>>>>>>>>>> andleRunner.java:53) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecuto >>>>>>>>>>>> rI >>>>>>>>>>>> mp >>>>>>>>>>>>>>> l$1.run(DefaultExecutorFactory.java:66) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.j >>>>>>>>>>>> av >>>>>>>>>>>> a: >>>>>>>>>>>>>>> 1145) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor. >>>>>>>>>>>> ja >>>>>>>>>>>> va >>>>>>>>>>>>>>> :615) >>>>>>>>>>>>>>> at java.lang.Thread.run(Thread.java:744) >>>>>>>>>>>>>>> Process 'Gradle Worker 13' finished with non-zero exit >>>>>> value 1 >>>>>>>>>>>>>>> org.gradle.process.internal.ExecException: Process 'Gradle >>>>>>>>> Worker >>>>>>>>>>> 13' >>>>>>>>>>>>>>> finished with non-zero exit value 1 >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.process.internal.DefaultExecHandle$ExecResultImpl.assertNor >>>>>>>>>>>> ma >>>>>>>>>>>> lE >>>>>>>>>>>>>>> xitValue(DefaultExecHandle.java:362) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.process.internal.DefaultWorkerProcess.onProcessStop(Default >>>>>>>>>>>> Wo >>>>>>>>>>>> rk >>>>>>>>>>>>>>> erProcess.java:89) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.process.internal.DefaultWorkerProcess.access$000(DefaultWor >>>>>>>>>>>> ke >>>>>>>>>>>> rP >>>>>>>>>>>>>>> rocess.java:33) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.process.internal.DefaultWorkerProcess$1.executionFinished(D >>>>>>>>>>>> ef >>>>>>>>>>>> au >>>>>>>>>>>>>>> ltWorkerProcess.java:55) >>>>>>>>>>>>>>> at >>>>>> sun.reflect.NativeMethodAccessorImpl.invoke0(Native >>>>>>>>>>> Method) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j >>>>>>>>>>>> av >>>>>>>>>>>> a: >>>>>>>>>>>>>>> 57) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccess >>>>>>>>>>>> or >>>>>>>>>>>> Im >>>>>>>>>>>>>>> pl.java:43) >>>>>>>>>>>>>>> at java.lang.reflect.Method.invoke(Method.java:606) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDi >>>>>>>>>>>> sp >>>>>>>>>>>> at >>>>>>>>>>>>>>> ch.java:35) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDi >>>>>>>>>>>> sp >>>>>>>>>>>> at >>>>>>>>>>>>>>> ch.java:24) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java: >>>>>>>>>>>> 81 >>>>>>>>>>>> ) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java: >>>>>>>>>>>> 30 >>>>>>>>>>>> ) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.messaging.dispatch.ProxyDispatchAdapter$DispatchingInvocati >>>>>>>>>>>> on >>>>>>>>>>>> Ha >>>>>>>>>>>>>>> ndler.invoke(ProxyDispatchAdapter.java:93) >>>>>>>>>>>>>>> at com.sun.proxy.$Proxy46.executionFinished(Unknown >>>>>>>>>>> Source) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.process.internal.DefaultExecHandle.setEndStateInfo(DefaultE >>>>>>>>>>>> xe >>>>>>>>>>>> cH >>>>>>>>>>>>>>> andle.java:212) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.process.internal.DefaultExecHandle.finished(DefaultExecHand >>>>>>>>>>>> le >>>>>>>>>>>> .j >>>>>>>>>>>>>>> ava:309) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.process.internal.ExecHandleRunner.completed(ExecHandleRunne >>>>>>>>>>>> r. >>>>>>>>>>>> ja >>>>>>>>>>>>>>> va:108) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.process.internal.ExecHandleRunner.run(ExecHandleRunner.java >>>>>>>>>>>> :8 >>>>>>>>>>>> 8) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecuto >>>>>>>>>>>> rI >>>>>>>>>>>> mp >>>>>>>>>>>>>>> l$1.run(DefaultExecutorFactory.java:66) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.j >>>>>>>>>>>> av >>>>>>>>>>>> a: >>>>>>>>>>>>>>> 1145) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor. >>>>>>>>>>>> ja >>>>>>>>>>>> va >>>>>>>>>>>>>>> :615) >>>>>>>>>>>>>>> at java.lang.Thread.run(Thread.java:744) >>>>>>>>>>>>>>> OpenJDK 64-Bit Server VM warning: INFO: >>>>>>>>>>>>>>> os::commit_memory(0x000000070a6c0000, 3946053632, 0) >>>>> failed; >>>>>>>>>>>>>>> error='Cannot allocate memory' (errno=12) >>>>>>>>>>>>>>> # >>>>>>>>>>>>>>> # There is insufficient memory for the Java Runtime >>>>>>>>> Environment >>>>>>>>>>> to >>>>>>>>>>>>>>> continue. >>>>>>>>>>>>>>> # Native memory allocation (malloc) failed to allocate >>>>>>>>> 3946053632 >>>>>>>>>>>> bytes >>>>>>>>>>>>>>> for committing reserved memory. >>>>>>>>>>>>>>> # An error report file with more information is saved as: >>>>>>>>>>>>>>> # >>>>>> /home/ubuntu/incubator-samza/samza-kafka/hs_err_pid2518.log >>>>>>>>>>>>>>> Could not write standard input into: Gradle Worker 14. >>>>>>>>>>>>>>> java.io.IOException: Broken pipe >>>>>>>>>>>>>>> at java.io.FileOutputStream.writeBytes(Native >>>>> Method) >>>>>>>>>>>>>>> at >>>>>>>>>>> java.io.FileOutputStream.write(FileOutputStream.java:345) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>> >>>>>>> >>>>>>> >> java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>> java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.process.internal.streams.ExecOutputHandleRunner.run(ExecOut >>>>>>>>>>>> pu >>>>>>>>>>>> tH >>>>>>>>>>>>>>> andleRunner.java:53) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecuto >>>>>>>>>>>> rI >>>>>>>>>>>> mp >>>>>>>>>>>>>>> l$1.run(DefaultExecutorFactory.java:66) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.j >>>>>>>>>>>> av >>>>>>>>>>>> a: >>>>>>>>>>>>>>> 1145) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor. >>>>>>>>>>>> ja >>>>>>>>>>>> va >>>>>>>>>>>>>>> :615) >>>>>>>>>>>>>>> at java.lang.Thread.run(Thread.java:744) >>>>>>>>>>>>>>> Process 'Gradle Worker 14' finished with non-zero exit >>>>>> value 1 >>>>>>>>>>>>>>> org.gradle.process.internal.ExecException: Process 'Gradle >>>>>>>>> Worker >>>>>>>>>>> 14' >>>>>>>>>>>>>>> finished with non-zero exit value 1 >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.process.internal.DefaultExecHandle$ExecResultImpl.assertNor >>>>>>>>>>>> ma >>>>>>>>>>>> lE >>>>>>>>>>>>>>> xitValue(DefaultExecHandle.java:362) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.process.internal.DefaultWorkerProcess.onProcessStop(Default >>>>>>>>>>>> Wo >>>>>>>>>>>> rk >>>>>>>>>>>>>>> erProcess.java:89) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.process.internal.DefaultWorkerProcess.access$000(DefaultWor >>>>>>>>>>>> ke >>>>>>>>>>>> rP >>>>>>>>>>>>>>> rocess.java:33) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.process.internal.DefaultWorkerProcess$1.executionFinished(D >>>>>>>>>>>> ef >>>>>>>>>>>> au >>>>>>>>>>>>>>> ltWorkerProcess.java:55) >>>>>>>>>>>>>>> at >>>>>> sun.reflect.NativeMethodAccessorImpl.invoke0(Native >>>>>>>>>>> Method) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j >>>>>>>>>>>> av >>>>>>>>>>>> a: >>>>>>>>>>>>>>> 57) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccess >>>>>>>>>>>> or >>>>>>>>>>>> Im >>>>>>>>>>>>>>> pl.java:43) >>>>>>>>>>>>>>> at java.lang.reflect.Method.invoke(Method.java:606) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDi >>>>>>>>>>>> sp >>>>>>>>>>>> at >>>>>>>>>>>>>>> ch.java:35) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDi >>>>>>>>>>>> sp >>>>>>>>>>>> at >>>>>>>>>>>>>>> ch.java:24) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java: >>>>>>>>>>>> 81 >>>>>>>>>>>> ) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java: >>>>>>>>>>>> 30 >>>>>>>>>>>> ) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.messaging.dispatch.ProxyDispatchAdapter$DispatchingInvocati >>>>>>>>>>>> on >>>>>>>>>>>> Ha >>>>>>>>>>>>>>> ndler.invoke(ProxyDispatchAdapter.java:93) >>>>>>>>>>>>>>> at com.sun.proxy.$Proxy46.executionFinished(Unknown >>>>>>>>>>> Source) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.process.internal.DefaultExecHandle.setEndStateInfo(DefaultE >>>>>>>>>>>> xe >>>>>>>>>>>> cH >>>>>>>>>>>>>>> andle.java:212) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.process.internal.DefaultExecHandle.finished(DefaultExecHand >>>>>>>>>>>> le >>>>>>>>>>>> .j >>>>>>>>>>>>>>> ava:309) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.process.internal.ExecHandleRunner.completed(ExecHandleRunne >>>>>>>>>>>> r. >>>>>>>>>>>> ja >>>>>>>>>>>>>>> va:108) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.process.internal.ExecHandleRunner.run(ExecHandleRunner.java >>>>>>>>>>>> :8 >>>>>>>>>>>> 8) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecuto >>>>>>>>>>>> rI >>>>>>>>>>>> mp >>>>>>>>>>>>>>> l$1.run(DefaultExecutorFactory.java:66) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.j >>>>>>>>>>>> av >>>>>>>>>>>> a: >>>>>>>>>>>>>>> 1145) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>> >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor. >>>>>>>>>>>> ja >>>>>>>>>>>> va >>>>>>>>>>>>>>> :615) >>>>>>>>>>>>>>> at java.lang.Thread.r >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Do I need more memory for my machines? Each already has >>>>>> 4GB. I >>>>>>>>>>> really >>>>>>>>>>>>>>> need to have this running. I¹m not sure which way is best >>>>>>>>> http or >>>>>>>>>>> hdfs >>>>>>>>>>>>>>> which one you suggest and how can i solve my problem for >>>>>> each >>>>>>>>>>> case. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Thanks in advance and sorry for bothering this much. >>>>>>>>>>>>>>> On 10 Aug 2014, at 00:20, Telles Nobrega >>>>>>>>>>> <[email protected]> >>>>>>>>>>>>> wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Hi Chris, now I have the tar file in my RM machine, and >>>>>> the >>>>>>>>>>> yarn >>>>>>>>>>>> path >>>>>>>>>>>>>>>> points to it. I changed the core-site.xml to use >>>>>>>>> HttpFileSystem >>>>>>>>>>>> instead >>>>>>>>>>>>>>>> of HDFS now it is failing with >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Application application_1407640485281_0001 failed 2 >>>>> times >>>>>>>>> due >>>>>>>>>>> to >>>>>>>>>>> AM >>>>>>>>>>>>>>>> Container for appattempt_1407640485281_0001_000002 exited >>>>>>>>> with >>>>>>>>>>>>>>>> exitCode:-1000 due to: java.lang.ClassNotFoundException: >>>>>>>>> Class >>>>>>>>>>>>>>>> org.apache.samza.util.hadoop.HttpFileSystem not found >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> I think I can solve this just installing scala files >>>>> from >>>>>>>>> the >>>>>>>>>>> samza >>>>>>>>>>>>>>>> tutorial, can you confirm that? >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> On 09 Aug 2014, at 08:34, Telles Nobrega >>>>>>>>>>> <[email protected] >>>>>>>>>>>> >>>>>>>>>>>>>>>> wrote: >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Hi Chris, >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> I think the problem is that I forgot to update the >>>>>>>>>>>> yarn.job.package. >>>>>>>>>>>>>>>>> I will try again to see if it works now. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> I have one more question, how can I stop (command line) >>>>>> the >>>>>>>>>>> jobs >>>>>>>>>>>>>>>>> running in my topology, for the experiment that I will >>>>>> run, >>>>>>>>> I >>>>>>>>>>> need >>>>>>>>>>>> to >>>>>>>>>>>>>>>>> run the same job in 4 minutes intervals. So I need to >>>>> kill >>>>>>>>> it, >>>>>>>>>>> clean >>>>>>>>>>>>>>>>> the kafka topics and rerun. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Thanks in advance. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> On 08 Aug 2014, at 12:41, Chris Riccomini >>>>>>>>>>>>>>>>> <[email protected]> wrote: >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Hey Telles, >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Do I need to have the job folder on each machine in >>>>> my >>>>>>>>>>> cluster? >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> No, you should not need to do this. There are two ways >>>>>> to >>>>>>>>>>> deploy >>>>>>>>>>>>> your >>>>>>>>>>>>>>>>>> tarball to the YARN grid. One is to put it in HDFS, >>>>> and >>>>>>>>> the >>>>>>>>>>> other >>>>>>>>>>>> is >>>>>>>>>>>>>>>>>> to >>>>>>>>>>>>>>>>>> put it on an HTTP server. The link to running a Samza >>>>>> job >>>>>>>>> in >>>>>>>>>>> a >>>>>>>>>>>>>>>>>> multi-node >>>>>>>>>>>>>>>>>> YARN cluster describes how to do both (either HTTP >>>>>> server >>>>>>>>> or >>>>>>>>>>>> HDFS). >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> In both cases, once the tarball is put in on the >>>>>> HTTP/HDFS >>>>>>>>>>>>> server(s), >>>>>>>>>>>>>>>>>> you >>>>>>>>>>>>>>>>>> must update yarn.package.path to point to it. From >>>>>> there, >>>>>>>>> the >>>>>>>>>>> YARN >>>>>>>>>>>>> NM >>>>>>>>>>>>>>>>>> should download it for you automatically when you >>>>> start >>>>>>>>> your >>>>>>>>>>> job. >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> * Can you send along a paste of your job config? >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Cheers, >>>>>>>>>>>>>>>>>> Chris >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> On 8/8/14 8:04 AM, "Claudio Martins" >>>>>>>>>>> <[email protected]> >>>>>>>>>>>>> wrote: >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Hi Telles, it looks to me that you forgot to update >>>>> the >>>>>>>>>>>>>>>>>>> "yarn.package.path" >>>>>>>>>>>>>>>>>>> attribute in your config file for the task. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> - Claudio Martins >>>>>>>>>>>>>>>>>>> Head of Engineering >>>>>>>>>>>>>>>>>>> MobileAware USA Inc. / www.mobileaware.com >>>>>>>>>>>>>>>>>>> office: +1 617 986 5060 / mobile: +1 617 480 5288 >>>>>>>>>>>>>>>>>>> linkedin: www.linkedin.com/in/martinsclaudio >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> On Fri, Aug 8, 2014 at 10:55 AM, Telles Nobrega >>>>>>>>>>>>>>>>>>> <[email protected]> >>>>>>>>>>>>>>>>>>> wrote: >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> this is my first time trying to run a job on a >>>>>> multinode >>>>>>>>>>>>>>>>>>>> environment. I >>>>>>>>>>>>>>>>>>>> have the cluster set up, I can see in the GUI that >>>>> all >>>>>>>>>>> nodes >>>>>>>>>>> are >>>>>>>>>>>>>>>>>>>> working. >>>>>>>>>>>>>>>>>>>> Do I need to have the job folder on each machine in >>>>> my >>>>>>>>>>> cluster? >>>>>>>>>>>>>>>>>>>> - The first time I tried running with the job on the >>>>>>>>>>> namenode >>>>>>>>>>>>>>>>>>>> machine >>>>>>>>>>>>>>>>>>>> and >>>>>>>>>>>>>>>>>>>> it failed saying: >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Application application_1407509228798_0001 failed 2 >>>>>>>>> times >>>>>>>>>>> due >>>>>>>>>>> to >>>>>>>>>>>>> AM >>>>>>>>>>>>>>>>>>>> Container for appattempt_1407509228798_0001_000002 >>>>>>>>> exited >>>>>>>>>>> with >>>>>>>>>>>>>>>>>>>> exitCode: >>>>>>>>>>>>>>>>>>>> -1000 due to: File >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>>>>>>> >> file:/home/ubuntu/alarm-samza/samza-job-package/target/samza-job- >>>>>>>>>>>>>>>>> pa >>>>>>>>>>>>>>>>> ck >>>>>>>>>>>>>>>>>>>> age- >>>>>>>>>>>>>>>>>>>> 0.7.0-dist.tar.gz >>>>>>>>>>>>>>>>>>>> does not exist >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> So I copied the folder to each machine in my cluster >>>>>> and >>>>>>>>>>> got >>>>>>>>>>>> this >>>>>>>>>>>>>>>>>>>> error: >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Application application_1407509228798_0002 failed 2 >>>>>>>>> times >>>>>>>>>>> due >>>>>>>>>>> to >>>>>>>>>>>>> AM >>>>>>>>>>>>>>>>>>>> Container for appattempt_1407509228798_0002_000002 >>>>>>>>> exited >>>>>>>>>>> with >>>>>>>>>>>>>>>>>>>> exitCode: >>>>>>>>>>>>>>>>>>>> -1000 due to: Resource >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>>>>>>>>>>>> >> file:/home/ubuntu/alarm-samza/samza-job-package/target/samza-job- >>>>>>>>>>>>>>>>> pa >>>>>>>>>>>>>>>>> ck >>>>>>>>>>>>>>>>>>>> age- >>>>>>>>>>>>>>>>>>>> 0.7.0-dist.tar.gz >>>>>>>>>>>>>>>>>>>> changed on src filesystem (expected 1407509168000, >>>>> was >>>>>>>>>>>>> 1407509434000 >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> What am I missing? >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> p.s.: I followed this >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> < >>>>>>>>>>>>>> >>>>>>>>>>> >>>>>> https://github.com/yahoo/samoa/wiki/Executing-SAMOA-with-Apache-Samz >>>>>>>>>>>>>>>>>>>> a> >>>>>>>>>>>>>>>>>>>> tutorial >>>>>>>>>>>>>>>>>>>> and this >>>>>>>>>>>>>>>>>>>> < >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>> >>>>>>> >> http://samza.incubator.apache.org/learn/tutorials/0.7.0/run-in-multi- >>>>>>>>>>>>>>>>>>>> node >>>>>>>>>>>>>>>>>>>> -yarn.html >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> to >>>>>>>>>>>>>>>>>>>> set up the cluster. >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Help is much appreciated. >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Thanks in advance. >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>> ------------------------------------------ >>>>>>>>>>>>>>>>>>>> Telles Mota Vidal Nobrega >>>>>>>>>>>>>>>>>>>> M.sc. Candidate at UFCG >>>>>>>>>>>>>>>>>>>> B.sc. in Computer Science at UFCG >>>>>>>>>>>>>>>>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> ------------------------------------------ >>>>>>>>>>>> Telles Mota Vidal Nobrega >>>>>>>>>>>> M.sc. Candidate at UFCG >>>>>>>>>>>> B.sc. in Computer Science at UFCG >>>>>>>>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> ------------------------------------------ >>>>>>>>>> Telles Mota Vidal Nobrega >>>>>>>>>> M.sc. Candidate at UFCG >>>>>>>>>> B.sc. in Computer Science at UFCG >>>>>>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> ------------------------------------------ >>>>>>>> Telles Mota Vidal Nobrega >>>>>>>> M.sc. Candidate at UFCG >>>>>>>> B.sc. in Computer Science at UFCG >>>>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG >>>>>>> >>>>>>> >>>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> ------------------------------------------ >>>>> Telles Mota Vidal Nobrega >>>>> M.sc. Candidate at UFCG >>>>> B.sc. in Computer Science at UFCG >>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG >>>>> >>> >> >>
