You may forget this last email, I was really stupid and put the files in a
different folder. Now it could find the file but it’s not there yet… another
error came up
Exception in thread "main" java.util.ServiceConfigurationError:
org.apache.hadoop.fs.FileSystem: Provider
org.apache.hadoop.hdfs.DistributedFileSystem could not be instantiated
at java.util.ServiceLoader.fail(ServiceLoader.java:224)
at java.util.ServiceLoader.access$100(ServiceLoader.java:181)
at java.util.ServiceLoader$LazyIterator.next(ServiceLoader.java:377)
at java.util.ServiceLoader$1.next(ServiceLoader.java:445)
at org.apache.hadoop.fs.FileSystem.loadFileSystems(FileSystem.java:2400)
at
org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2411)
at
org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2428)
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88)
at
org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2467)
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449)
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367)
at org.apache.hadoop.fs.Path.getFileSystem(Path.java:287)
at
org.apache.samza.job.yarn.ClientHelper.submitApplication(ClientHelper.scala:111)
at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:55)
at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:48)
at org.apache.samza.job.JobRunner.run(JobRunner.scala:62)
at org.apache.samza.job.JobRunner$.main(JobRunner.scala:37)
at org.apache.samza.job.JobRunner.main(JobRunner.scala)
Caused by: java.lang.NoClassDefFoundError:
org/apache/hadoop/conf/Configuration$DeprecationDelta
at
org.apache.hadoop.hdfs.HdfsConfiguration.addDeprecatedKeys(HdfsConfiguration.java:66)
at
org.apache.hadoop.hdfs.HdfsConfiguration.<clinit>(HdfsConfiguration.java:31)
at
org.apache.hadoop.hdfs.DistributedFileSystem.<clinit>(DistributedFileSystem.java:106)
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
at java.lang.Class.newInstance(Class.java:374)
at java.util.ServiceLoader$LazyIterator.next(ServiceLoader.java:373)
... 15 more
Caused by: java.lang.ClassNotFoundException:
org.apache.hadoop.conf.Configuration$DeprecationDelta
at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
... 24 more
On 11 Aug 2014, at 20:45, Telles Nobrega <[email protected]> wrote:
> Hi, I copied hadoop-hdfs-2.3.0 to my-job/lib and it changed the error which
> is good but the error is back to
>
> Exception in thread "main" java.lang.RuntimeException:
> java.lang.ClassNotFoundException: Class
> org.apache.hadoop.hdfs.DistributedFileSystem not found
> at
> org.apache.hadoop.conf.Configuration.getClass(Configuration.java:1720)
> at
> org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2415)
> at
> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2428)
> at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88)
> at
> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2467)
> at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449)
> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367)
> at org.apache.hadoop.fs.Path.getFileSystem(Path.java:287)
> at
> org.apache.samza.job.yarn.ClientHelper.submitApplication(ClientHelper.scala:111)
> at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:55)
> at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:48)
> at org.apache.samza.job.JobRunner.run(JobRunner.scala:62)
> at org.apache.samza.job.JobRunner$.main(JobRunner.scala:37)
> at org.apache.samza.job.JobRunner.main(JobRunner.scala)
> Caused by: java.lang.ClassNotFoundException: Class
> org.apache.hadoop.hdfs.DistributedFileSystem not found
> at
> org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:1626)
> at
> org.apache.hadoop.conf.Configuration.getClass(Configuration.java:1718)
> ... 13 more
>
> Do I need to have this lib in all nodes at the job folder or just to submit?
>
> On 11 Aug 2014, at 20:11, Yan Fang <[email protected]> wrote:
>
>> Hi Telles,
>>
>> I replayed your problem and think I figured out why CLASSPATH does not
>> work. Because in our script bin/run-class.sh, we have the line
>> "CLASSPATH=$HADOOP_CONF_DIR", which actually ingores your setting.
>>
>> So a simple solution is to copy the hadoop-hdfs.jar to your samza lib
>> directory. Then run bin/run-job ----config-factory=... --config-path=... .
>> Let me know how it goes. Thank you.
>>
>> Cheers,
>>
>> Fang, Yan
>> [email protected]
>> +1 (206) 849-4108
>>
>>
>> On Mon, Aug 11, 2014 at 4:07 PM, Telles Nobrega <[email protected]>
>> wrote:
>>
>>> Sure, thanks.
>>>
>>>
>>> On Mon, Aug 11, 2014 at 6:22 PM, Yan Fang <[email protected]> wrote:
>>>
>>>> Hi Telles,
>>>>
>>>> I am not sure whether exporting the CLASSPATH works. (sometimes it does
>>> not
>>>> work for me...) My suggestion is to include the hdfs jar explicitly in
>>> the
>>>> package that you upload to hdfs. Also , remember to put the jar into your
>>>> local samza (which is deploy/samza/lib if you go with the hello-samza
>>>> tutorial) Let me know if that works.
>>>>
>>>> Cheers,
>>>>
>>>> Fang, Yan
>>>> [email protected]
>>>> +1 (206) 849-4108
>>>>
>>>>
>>>> On Mon, Aug 11, 2014 at 2:04 PM, Chris Riccomini <
>>>> [email protected]> wrote:
>>>>
>>>>> Hey Telles,
>>>>>
>>>>> Hmm. I'm out of ideas. If Zhijie is around, he'd probably be of use,
>>> but
>>>> I
>>>>> haven't heard from him in a while.
>>>>>
>>>>> I'm afraid your best bet is probably to email the YARN dev mailing
>>> list,
>>>>> since this is a YARN config issue.
>>>>>
>>>>> Cheers,
>>>>> Chris
>>>>>
>>>>> On 8/11/14 1:58 PM, "Telles Nobrega" <[email protected]> wrote:
>>>>>
>>>>>> I exported export
>>>>>
>>>>
>>>> CLASSPATH=$CLASSPATH:hadoop-2.3.0/share/hadoop/hdfs/hadoop-hdfs-2.3.0.jar
>>>>>> and still happened the same problem.
>>>>>>
>>>>>>
>>>>>> On Mon, Aug 11, 2014 at 5:35 PM, Chris Riccomini <
>>>>>> [email protected]> wrote:
>>>>>>
>>>>>>> Hey Telles,
>>>>>>>
>>>>>>> It sounds like either the HDFS jar is missing from the classpath, or
>>>> the
>>>>>>> hdfs file system needs to be configured:
>>>>>>>
>>>>>>> <property>
>>>>>>> <name>fs.hdfs.impl</name>
>>>>>>> <value>org.apache.hadoop.hdfs.DistributedFileSystem</value>
>>>>>>> <description>The FileSystem for hdfs: uris.</description>
>>>>>>> </property>
>>>>>>>
>>>>>>>
>>>>>>> (from
>>>>>>>
>>>>>>>
>>>>>
>>>>
>>> https://groups.google.com/a/cloudera.org/forum/#!topic/scm-users/lyho8ptA
>>>>>>> zE
>>>>>>> 0)
>>>>>>>
>>>>>>> I believe this will need to be configured for your NM.
>>>>>>>
>>>>>>> Cheers,
>>>>>>> Chris
>>>>>>>
>>>>>>> On 8/11/14 1:31 PM, "Telles Nobrega" <[email protected]>
>>> wrote:
>>>>>>>
>>>>>>>> Yes, it is like this:
>>>>>>>>
>>>>>>>> <configuration>
>>>>>>>> <property>
>>>>>>>> <name>dfs.datanode.data.dir</name>
>>>>>>>> <value>file:///home/ubuntu/hadoop-2.3.0/hdfs/datanode</value>
>>>>>>>> <description>Comma separated list of paths on the local
>>>> filesystem
>>>>>>> of
>>>>>>>> a
>>>>>>>> DataNode where it should store its blocks.</description>
>>>>>>>> </property>
>>>>>>>>
>>>>>>>> <property>
>>>>>>>> <name>dfs.namenode.name.dir</name>
>>>>>>>> <value>file:///home/ubuntu/hadoop-2.3.0/hdfs/namenode</value>
>>>>>>>> <description>Path on the local filesystem where the NameNode
>>>> stores
>>>>>>>> the
>>>>>>>> namespace and transaction logs persistently.</description>
>>>>>>>> </property>
>>>>>>>> </configuration>
>>>>>>>> ~
>>>>>>>>
>>>>>>>> I saw some report that this may be a classpath problem. Does this
>>>>>>> sounds
>>>>>>>> right to you?
>>>>>>>>
>>>>>>>>
>>>>>>>> On Mon, Aug 11, 2014 at 5:25 PM, Yan Fang <[email protected]>
>>>>> wrote:
>>>>>>>>
>>>>>>>>> Hi Telles,
>>>>>>>>>
>>>>>>>>> It looks correct. Did you put the hdfs-site.xml into your
>>>>>>>>> HADOOP_CONF_DIR
>>>>>>>>> ?(such as ~/.samza/conf)
>>>>>>>>>
>>>>>>>>> Fang, Yan
>>>>>>>>> [email protected]
>>>>>>>>> +1 (206) 849-4108
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Mon, Aug 11, 2014 at 1:02 PM, Telles Nobrega
>>>>>>>>> <[email protected]>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>> Hi Yan Fang,
>>>>>>>>>>
>>>>>>>>>> I was able to deploy the file to hdfs, I can see them in all my
>>>>>>> nodes
>>>>>>>>> but
>>>>>>>>>> when I tried running I got this error:
>>>>>>>>>>
>>>>>>>>>> Exception in thread "main" java.io.IOException: No FileSystem
>>> for
>>>>>>>>> scheme:
>>>>>>>>>> hdfs
>>>>>>>>>> at
>>>>>>>>>
>>>>>
>>>>
>>>>> org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2421)
>>>>>>>>>> at
>>>>>>>>>
>>>>>
>>>>> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2428)
>>>>>>>>>> at
>>> org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88)
>>>>>>>>>> at
>>>>>>>>>
>>>>>
>>>>> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2467)
>>>>>>>>>> at
>>>> org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449)
>>>>>>>>>> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367)
>>>>>>>>>> at org.apache.hadoop.fs.Path.getFileSystem(Path.java:287)
>>>>>>>>>> at
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>> org.apache.samza.job.yarn.ClientHelper.submitApplication(ClientHelper.s
>>>>>>>>> ca
>>>>>>>>> la:111)
>>>>>>>>>> at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:55)
>>>>>>>>>> at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:48)
>>>>>>>>>> at org.apache.samza.job.JobRunner.run(JobRunner.scala:62)
>>>>>>>>>> at org.apache.samza.job.JobRunner$.main(JobRunner.scala:37)
>>>>>>>>>> at org.apache.samza.job.JobRunner.main(JobRunner.scala)
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> This is my yarn.package.path config:
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>> yarn.package.path=hdfs://telles-master-samza:50070/samza-job-package-0
>>>>>>>>> .7
>>>>>>>>> .0-dist.tar.gz
>>>>>>>>>>
>>>>>>>>>> Thanks in advance
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Mon, Aug 11, 2014 at 3:00 PM, Yan Fang <
>>> [email protected]>
>>>>>>>>> wrote:
>>>>>>>>>>
>>>>>>>>>>> Hi Telles,
>>>>>>>>>>>
>>>>>>>>>>> In terms of "*I tried pushing the tar file to HDFS but I got
>>> an
>>>>>>>>> error
>>>>>>>>>> from
>>>>>>>>>>> hadoop saying that it couldn’t find core-site.xml file*.", I
>>>>>>> guess
>>>>>>>>> you
>>>>>>>>>> set
>>>>>>>>>>> the HADOOP_CONF_DIR variable and made it point to
>>>> ~/.samza/conf.
>>>>>>> You
>>>>>>>>> can
>>>>>>>>>> do
>>>>>>>>>>> 1) make the HADOOP_CONF_DIR point to the directory where your
>>>>>>> conf
>>>>>>>>> files
>>>>>>>>>>> are, such as /etc/hadoop/conf. Or 2) copy the config files to
>>>>>>>>>>> ~/.samza/conf. Thank you,
>>>>>>>>>>>
>>>>>>>>>>> Cheer,
>>>>>>>>>>>
>>>>>>>>>>> Fang, Yan
>>>>>>>>>>> [email protected]
>>>>>>>>>>> +1 (206) 849-4108
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Mon, Aug 11, 2014 at 7:40 AM, Chris Riccomini <
>>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> Hey Telles,
>>>>>>>>>>>>
>>>>>>>>>>>> To get YARN working with the HTTP file system, you need to
>>>>>>> follow
>>>>>>>>> the
>>>>>>>>>>>> instructions on:
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>
>>>>
>>> http://samza.incubator.apache.org/learn/tutorials/0.7.0/run-in-multi-node
>>>>>>>>> -y
>>>>>>>>>>>> arn.html
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> In the "Set Up Http Filesystem for YARN" section.
>>>>>>>>>>>>
>>>>>>>>>>>> You shouldn't need to compile anything (no Gradle, which is
>>>>>>> what
>>>>>>>>> your
>>>>>>>>>>>> stack trace is showing). This setup should be done for all
>>> of
>>>>>>> the
>>>>>>>>> NMs,
>>>>>>>>>>>> since they will be the ones downloading your job's package
>>>>>>> (from
>>>>>>>>>>>> yarn.package.path).
>>>>>>>>>>>>
>>>>>>>>>>>> Cheers,
>>>>>>>>>>>> Chris
>>>>>>>>>>>>
>>>>>>>>>>>> On 8/9/14 9:44 PM, "Telles Nobrega" <
>>> [email protected]
>>>>>
>>>>>>>>> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> Hi again, I tried installing the scala libs but the Http
>>>>>>> problem
>>>>>>>>> still
>>>>>>>>>>>>> occurs. I realised that I need to compile incubator samza
>>> in
>>>>>>> the
>>>>>>>>>>> machines
>>>>>>>>>>>>> that I¹m going to run the jobs, but the compilation fails
>>>> with
>>>>>>>>> this
>>>>>>>>>> huge
>>>>>>>>>>>>> message:
>>>>>>>>>>>>>
>>>>>>>>>>>>> #
>>>>>>>>>>>>> # There is insufficient memory for the Java Runtime
>>>>>>> Environment
>>>>>>>>> to
>>>>>>>>>>>>> continue.
>>>>>>>>>>>>> # Native memory allocation (malloc) failed to allocate
>>>>>>> 3946053632
>>>>>>>>>> bytes
>>>>>>>>>>>>> for committing reserved memory.
>>>>>>>>>>>>> # An error report file with more information is saved as:
>>>>>>>>>>>>> #
>>>> /home/ubuntu/incubator-samza/samza-kafka/hs_err_pid2506.log
>>>>>>>>>>>>> Could not write standard input into: Gradle Worker 13.
>>>>>>>>>>>>> java.io.IOException: Broken pipe
>>>>>>>>>>>>> at java.io.FileOutputStream.writeBytes(Native
>>> Method)
>>>>>>>>>>>>> at
>>>>>>>>> java.io.FileOutputStream.write(FileOutputStream.java:345)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>
>>>>>
>>>>> java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>> java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.process.internal.streams.ExecOutputHandleRunner.run(ExecOut
>>>>>>>>>> pu
>>>>>>>>>> tH
>>>>>>>>>>>>> andleRunner.java:53)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecuto
>>>>>>>>>> rI
>>>>>>>>>> mp
>>>>>>>>>>>>> l$1.run(DefaultExecutorFactory.java:66)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.j
>>>>>>>>>> av
>>>>>>>>>> a:
>>>>>>>>>>>>> 1145)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.
>>>>>>>>>> ja
>>>>>>>>>> va
>>>>>>>>>>>>> :615)
>>>>>>>>>>>>> at java.lang.Thread.run(Thread.java:744)
>>>>>>>>>>>>> Process 'Gradle Worker 13' finished with non-zero exit
>>>> value 1
>>>>>>>>>>>>> org.gradle.process.internal.ExecException: Process 'Gradle
>>>>>>> Worker
>>>>>>>>> 13'
>>>>>>>>>>>>> finished with non-zero exit value 1
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.process.internal.DefaultExecHandle$ExecResultImpl.assertNor
>>>>>>>>>> ma
>>>>>>>>>> lE
>>>>>>>>>>>>> xitValue(DefaultExecHandle.java:362)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.process.internal.DefaultWorkerProcess.onProcessStop(Default
>>>>>>>>>> Wo
>>>>>>>>>> rk
>>>>>>>>>>>>> erProcess.java:89)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.process.internal.DefaultWorkerProcess.access$000(DefaultWor
>>>>>>>>>> ke
>>>>>>>>>> rP
>>>>>>>>>>>>> rocess.java:33)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.process.internal.DefaultWorkerProcess$1.executionFinished(D
>>>>>>>>>> ef
>>>>>>>>>> au
>>>>>>>>>>>>> ltWorkerProcess.java:55)
>>>>>>>>>>>>> at
>>>> sun.reflect.NativeMethodAccessorImpl.invoke0(Native
>>>>>>>>> Method)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j
>>>>>>>>>> av
>>>>>>>>>> a:
>>>>>>>>>>>>> 57)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccess
>>>>>>>>>> or
>>>>>>>>>> Im
>>>>>>>>>>>>> pl.java:43)
>>>>>>>>>>>>> at java.lang.reflect.Method.invoke(Method.java:606)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDi
>>>>>>>>>> sp
>>>>>>>>>> at
>>>>>>>>>>>>> ch.java:35)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDi
>>>>>>>>>> sp
>>>>>>>>>> at
>>>>>>>>>>>>> ch.java:24)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java:
>>>>>>>>>> 81
>>>>>>>>>> )
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java:
>>>>>>>>>> 30
>>>>>>>>>> )
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.messaging.dispatch.ProxyDispatchAdapter$DispatchingInvocati
>>>>>>>>>> on
>>>>>>>>>> Ha
>>>>>>>>>>>>> ndler.invoke(ProxyDispatchAdapter.java:93)
>>>>>>>>>>>>> at com.sun.proxy.$Proxy46.executionFinished(Unknown
>>>>>>>>> Source)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.process.internal.DefaultExecHandle.setEndStateInfo(DefaultE
>>>>>>>>>> xe
>>>>>>>>>> cH
>>>>>>>>>>>>> andle.java:212)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.process.internal.DefaultExecHandle.finished(DefaultExecHand
>>>>>>>>>> le
>>>>>>>>>> .j
>>>>>>>>>>>>> ava:309)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.process.internal.ExecHandleRunner.completed(ExecHandleRunne
>>>>>>>>>> r.
>>>>>>>>>> ja
>>>>>>>>>>>>> va:108)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.process.internal.ExecHandleRunner.run(ExecHandleRunner.java
>>>>>>>>>> :8
>>>>>>>>>> 8)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecuto
>>>>>>>>>> rI
>>>>>>>>>> mp
>>>>>>>>>>>>> l$1.run(DefaultExecutorFactory.java:66)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.j
>>>>>>>>>> av
>>>>>>>>>> a:
>>>>>>>>>>>>> 1145)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.
>>>>>>>>>> ja
>>>>>>>>>> va
>>>>>>>>>>>>> :615)
>>>>>>>>>>>>> at java.lang.Thread.run(Thread.java:744)
>>>>>>>>>>>>> OpenJDK 64-Bit Server VM warning: INFO:
>>>>>>>>>>>>> os::commit_memory(0x000000070a6c0000, 3946053632, 0)
>>> failed;
>>>>>>>>>>>>> error='Cannot allocate memory' (errno=12)
>>>>>>>>>>>>> #
>>>>>>>>>>>>> # There is insufficient memory for the Java Runtime
>>>>>>> Environment
>>>>>>>>> to
>>>>>>>>>>>>> continue.
>>>>>>>>>>>>> # Native memory allocation (malloc) failed to allocate
>>>>>>> 3946053632
>>>>>>>>>> bytes
>>>>>>>>>>>>> for committing reserved memory.
>>>>>>>>>>>>> # An error report file with more information is saved as:
>>>>>>>>>>>>> #
>>>> /home/ubuntu/incubator-samza/samza-kafka/hs_err_pid2518.log
>>>>>>>>>>>>> Could not write standard input into: Gradle Worker 14.
>>>>>>>>>>>>> java.io.IOException: Broken pipe
>>>>>>>>>>>>> at java.io.FileOutputStream.writeBytes(Native
>>> Method)
>>>>>>>>>>>>> at
>>>>>>>>> java.io.FileOutputStream.write(FileOutputStream.java:345)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>
>>>>>
>>>>> java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>> java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.process.internal.streams.ExecOutputHandleRunner.run(ExecOut
>>>>>>>>>> pu
>>>>>>>>>> tH
>>>>>>>>>>>>> andleRunner.java:53)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecuto
>>>>>>>>>> rI
>>>>>>>>>> mp
>>>>>>>>>>>>> l$1.run(DefaultExecutorFactory.java:66)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.j
>>>>>>>>>> av
>>>>>>>>>> a:
>>>>>>>>>>>>> 1145)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.
>>>>>>>>>> ja
>>>>>>>>>> va
>>>>>>>>>>>>> :615)
>>>>>>>>>>>>> at java.lang.Thread.run(Thread.java:744)
>>>>>>>>>>>>> Process 'Gradle Worker 14' finished with non-zero exit
>>>> value 1
>>>>>>>>>>>>> org.gradle.process.internal.ExecException: Process 'Gradle
>>>>>>> Worker
>>>>>>>>> 14'
>>>>>>>>>>>>> finished with non-zero exit value 1
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.process.internal.DefaultExecHandle$ExecResultImpl.assertNor
>>>>>>>>>> ma
>>>>>>>>>> lE
>>>>>>>>>>>>> xitValue(DefaultExecHandle.java:362)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.process.internal.DefaultWorkerProcess.onProcessStop(Default
>>>>>>>>>> Wo
>>>>>>>>>> rk
>>>>>>>>>>>>> erProcess.java:89)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.process.internal.DefaultWorkerProcess.access$000(DefaultWor
>>>>>>>>>> ke
>>>>>>>>>> rP
>>>>>>>>>>>>> rocess.java:33)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.process.internal.DefaultWorkerProcess$1.executionFinished(D
>>>>>>>>>> ef
>>>>>>>>>> au
>>>>>>>>>>>>> ltWorkerProcess.java:55)
>>>>>>>>>>>>> at
>>>> sun.reflect.NativeMethodAccessorImpl.invoke0(Native
>>>>>>>>> Method)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j
>>>>>>>>>> av
>>>>>>>>>> a:
>>>>>>>>>>>>> 57)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccess
>>>>>>>>>> or
>>>>>>>>>> Im
>>>>>>>>>>>>> pl.java:43)
>>>>>>>>>>>>> at java.lang.reflect.Method.invoke(Method.java:606)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDi
>>>>>>>>>> sp
>>>>>>>>>> at
>>>>>>>>>>>>> ch.java:35)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDi
>>>>>>>>>> sp
>>>>>>>>>> at
>>>>>>>>>>>>> ch.java:24)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java:
>>>>>>>>>> 81
>>>>>>>>>> )
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java:
>>>>>>>>>> 30
>>>>>>>>>> )
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.messaging.dispatch.ProxyDispatchAdapter$DispatchingInvocati
>>>>>>>>>> on
>>>>>>>>>> Ha
>>>>>>>>>>>>> ndler.invoke(ProxyDispatchAdapter.java:93)
>>>>>>>>>>>>> at com.sun.proxy.$Proxy46.executionFinished(Unknown
>>>>>>>>> Source)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.process.internal.DefaultExecHandle.setEndStateInfo(DefaultE
>>>>>>>>>> xe
>>>>>>>>>> cH
>>>>>>>>>>>>> andle.java:212)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.process.internal.DefaultExecHandle.finished(DefaultExecHand
>>>>>>>>>> le
>>>>>>>>>> .j
>>>>>>>>>>>>> ava:309)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.process.internal.ExecHandleRunner.completed(ExecHandleRunne
>>>>>>>>>> r.
>>>>>>>>>> ja
>>>>>>>>>>>>> va:108)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.process.internal.ExecHandleRunner.run(ExecHandleRunner.java
>>>>>>>>>> :8
>>>>>>>>>> 8)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecuto
>>>>>>>>>> rI
>>>>>>>>>> mp
>>>>>>>>>>>>> l$1.run(DefaultExecutorFactory.java:66)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.j
>>>>>>>>>> av
>>>>>>>>>> a:
>>>>>>>>>>>>> 1145)
>>>>>>>>>>>>> at
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.
>>>>>>>>>> ja
>>>>>>>>>> va
>>>>>>>>>>>>> :615)
>>>>>>>>>>>>> at java.lang.Thread.r
>>>>>>>>>>>>>
>>>>>>>>>>>>> Do I need more memory for my machines? Each already has
>>>> 4GB. I
>>>>>>>>> really
>>>>>>>>>>>>> need to have this running. I¹m not sure which way is best
>>>>>>> http or
>>>>>>>>> hdfs
>>>>>>>>>>>>> which one you suggest and how can i solve my problem for
>>>> each
>>>>>>>>> case.
>>>>>>>>>>>>>
>>>>>>>>>>>>> Thanks in advance and sorry for bothering this much.
>>>>>>>>>>>>> On 10 Aug 2014, at 00:20, Telles Nobrega
>>>>>>>>> <[email protected]>
>>>>>>>>>>> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> Hi Chris, now I have the tar file in my RM machine, and
>>>> the
>>>>>>>>> yarn
>>>>>>>>>> path
>>>>>>>>>>>>>> points to it. I changed the core-site.xml to use
>>>>>>> HttpFileSystem
>>>>>>>>>> instead
>>>>>>>>>>>>>> of HDFS now it is failing with
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Application application_1407640485281_0001 failed 2
>>> times
>>>>>>> due
>>>>>>>>> to
>>>>>>>>> AM
>>>>>>>>>>>>>> Container for appattempt_1407640485281_0001_000002 exited
>>>>>>> with
>>>>>>>>>>>>>> exitCode:-1000 due to: java.lang.ClassNotFoundException:
>>>>>>> Class
>>>>>>>>>>>>>> org.apache.samza.util.hadoop.HttpFileSystem not found
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> I think I can solve this just installing scala files
>>> from
>>>>>>> the
>>>>>>>>> samza
>>>>>>>>>>>>>> tutorial, can you confirm that?
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On 09 Aug 2014, at 08:34, Telles Nobrega
>>>>>>>>> <[email protected]
>>>>>>>>>>
>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Hi Chris,
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> I think the problem is that I forgot to update the
>>>>>>>>>> yarn.job.package.
>>>>>>>>>>>>>>> I will try again to see if it works now.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> I have one more question, how can I stop (command line)
>>>> the
>>>>>>>>> jobs
>>>>>>>>>>>>>>> running in my topology, for the experiment that I will
>>>> run,
>>>>>>> I
>>>>>>>>> need
>>>>>>>>>> to
>>>>>>>>>>>>>>> run the same job in 4 minutes intervals. So I need to
>>> kill
>>>>>>> it,
>>>>>>>>> clean
>>>>>>>>>>>>>>> the kafka topics and rerun.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Thanks in advance.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On 08 Aug 2014, at 12:41, Chris Riccomini
>>>>>>>>>>>>>>> <[email protected]> wrote:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Hey Telles,
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Do I need to have the job folder on each machine in
>>> my
>>>>>>>>> cluster?
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> No, you should not need to do this. There are two ways
>>>> to
>>>>>>>>> deploy
>>>>>>>>>>> your
>>>>>>>>>>>>>>>> tarball to the YARN grid. One is to put it in HDFS,
>>> and
>>>>>>> the
>>>>>>>>> other
>>>>>>>>>> is
>>>>>>>>>>>>>>>> to
>>>>>>>>>>>>>>>> put it on an HTTP server. The link to running a Samza
>>>> job
>>>>>>> in
>>>>>>>>> a
>>>>>>>>>>>>>>>> multi-node
>>>>>>>>>>>>>>>> YARN cluster describes how to do both (either HTTP
>>>> server
>>>>>>> or
>>>>>>>>>> HDFS).
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> In both cases, once the tarball is put in on the
>>>> HTTP/HDFS
>>>>>>>>>>> server(s),
>>>>>>>>>>>>>>>> you
>>>>>>>>>>>>>>>> must update yarn.package.path to point to it. From
>>>> there,
>>>>>>> the
>>>>>>>>> YARN
>>>>>>>>>>> NM
>>>>>>>>>>>>>>>> should download it for you automatically when you
>>> start
>>>>>>> your
>>>>>>>>> job.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> * Can you send along a paste of your job config?
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Cheers,
>>>>>>>>>>>>>>>> Chris
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On 8/8/14 8:04 AM, "Claudio Martins"
>>>>>>>>> <[email protected]>
>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Hi Telles, it looks to me that you forgot to update
>>> the
>>>>>>>>>>>>>>>>> "yarn.package.path"
>>>>>>>>>>>>>>>>> attribute in your config file for the task.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> - Claudio Martins
>>>>>>>>>>>>>>>>> Head of Engineering
>>>>>>>>>>>>>>>>> MobileAware USA Inc. / www.mobileaware.com
>>>>>>>>>>>>>>>>> office: +1 617 986 5060 / mobile: +1 617 480 5288
>>>>>>>>>>>>>>>>> linkedin: www.linkedin.com/in/martinsclaudio
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> On Fri, Aug 8, 2014 at 10:55 AM, Telles Nobrega
>>>>>>>>>>>>>>>>> <[email protected]>
>>>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Hi,
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> this is my first time trying to run a job on a
>>>> multinode
>>>>>>>>>>>>>>>>>> environment. I
>>>>>>>>>>>>>>>>>> have the cluster set up, I can see in the GUI that
>>> all
>>>>>>>>> nodes
>>>>>>>>> are
>>>>>>>>>>>>>>>>>> working.
>>>>>>>>>>>>>>>>>> Do I need to have the job folder on each machine in
>>> my
>>>>>>>>> cluster?
>>>>>>>>>>>>>>>>>> - The first time I tried running with the job on the
>>>>>>>>> namenode
>>>>>>>>>>>>>>>>>> machine
>>>>>>>>>>>>>>>>>> and
>>>>>>>>>>>>>>>>>> it failed saying:
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Application application_1407509228798_0001 failed 2
>>>>>>> times
>>>>>>>>> due
>>>>>>>>> to
>>>>>>>>>>> AM
>>>>>>>>>>>>>>>>>> Container for appattempt_1407509228798_0001_000002
>>>>>>> exited
>>>>>>>>> with
>>>>>>>>>>>>>>>>>> exitCode:
>>>>>>>>>>>>>>>>>> -1000 due to: File
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>>>>>>> file:/home/ubuntu/alarm-samza/samza-job-package/target/samza-job-
>>>>>>>>>>>>>>> pa
>>>>>>>>>>>>>>> ck
>>>>>>>>>>>>>>>>>> age-
>>>>>>>>>>>>>>>>>> 0.7.0-dist.tar.gz
>>>>>>>>>>>>>>>>>> does not exist
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> So I copied the folder to each machine in my cluster
>>>> and
>>>>>>>>> got
>>>>>>>>>> this
>>>>>>>>>>>>>>>>>> error:
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Application application_1407509228798_0002 failed 2
>>>>>>> times
>>>>>>>>> due
>>>>>>>>> to
>>>>>>>>>>> AM
>>>>>>>>>>>>>>>>>> Container for appattempt_1407509228798_0002_000002
>>>>>>> exited
>>>>>>>>> with
>>>>>>>>>>>>>>>>>> exitCode:
>>>>>>>>>>>>>>>>>> -1000 due to: Resource
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>>>>>>>>>>> file:/home/ubuntu/alarm-samza/samza-job-package/target/samza-job-
>>>>>>>>>>>>>>> pa
>>>>>>>>>>>>>>> ck
>>>>>>>>>>>>>>>>>> age-
>>>>>>>>>>>>>>>>>> 0.7.0-dist.tar.gz
>>>>>>>>>>>>>>>>>> changed on src filesystem (expected 1407509168000,
>>> was
>>>>>>>>>>> 1407509434000
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> What am I missing?
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> p.s.: I followed this
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> <
>>>>>>>>>>>>
>>>>>>>>>
>>>> https://github.com/yahoo/samoa/wiki/Executing-SAMOA-with-Apache-Samz
>>>>>>>>>>>>>>>>>> a>
>>>>>>>>>>>>>>>>>> tutorial
>>>>>>>>>>>>>>>>>> and this
>>>>>>>>>>>>>>>>>> <
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>
>>>>> http://samza.incubator.apache.org/learn/tutorials/0.7.0/run-in-multi-
>>>>>>>>>>>>>>>>>> node
>>>>>>>>>>>>>>>>>> -yarn.html
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> to
>>>>>>>>>>>>>>>>>> set up the cluster.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Help is much appreciated.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Thanks in advance.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>>> ------------------------------------------
>>>>>>>>>>>>>>>>>> Telles Mota Vidal Nobrega
>>>>>>>>>>>>>>>>>> M.sc. Candidate at UFCG
>>>>>>>>>>>>>>>>>> B.sc. in Computer Science at UFCG
>>>>>>>>>>>>>>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> ------------------------------------------
>>>>>>>>>> Telles Mota Vidal Nobrega
>>>>>>>>>> M.sc. Candidate at UFCG
>>>>>>>>>> B.sc. in Computer Science at UFCG
>>>>>>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> ------------------------------------------
>>>>>>>> Telles Mota Vidal Nobrega
>>>>>>>> M.sc. Candidate at UFCG
>>>>>>>> B.sc. in Computer Science at UFCG
>>>>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG
>>>>>>>
>>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>> ------------------------------------------
>>>>>> Telles Mota Vidal Nobrega
>>>>>> M.sc. Candidate at UFCG
>>>>>> B.sc. in Computer Science at UFCG
>>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG
>>>>>
>>>>>
>>>>
>>>
>>>
>>>
>>> --
>>> ------------------------------------------
>>> Telles Mota Vidal Nobrega
>>> M.sc. Candidate at UFCG
>>> B.sc. in Computer Science at UFCG
>>> Software Engineer at OpenStack Project - HP/LSD-UFCG
>>>
>