Hi again, I tried installing the scala libs but the Http problem still occurs. 
I realised that I need to compile incubator samza in the machines that I’m 
going to run the jobs, but the compilation fails with this huge message: 

#
# There is insufficient memory for the Java Runtime Environment to continue.
# Native memory allocation (malloc) failed to allocate 3946053632 bytes for 
committing reserved memory.
# An error report file with more information is saved as:
# /home/ubuntu/incubator-samza/samza-kafka/hs_err_pid2506.log
Could not write standard input into: Gradle Worker 13.
java.io.IOException: Broken pipe
        at java.io.FileOutputStream.writeBytes(Native Method)
        at java.io.FileOutputStream.write(FileOutputStream.java:345)
        at 
java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82)
        at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140)
        at 
org.gradle.process.internal.streams.ExecOutputHandleRunner.run(ExecOutputHandleRunner.java:53)
        at 
org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecutorImpl$1.run(DefaultExecutorFactory.java:66)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
        at java.lang.Thread.run(Thread.java:744)
Process 'Gradle Worker 13' finished with non-zero exit value 1
org.gradle.process.internal.ExecException: Process 'Gradle Worker 13' finished 
with non-zero exit value 1
        at 
org.gradle.process.internal.DefaultExecHandle$ExecResultImpl.assertNormalExitValue(DefaultExecHandle.java:362)
        at 
org.gradle.process.internal.DefaultWorkerProcess.onProcessStop(DefaultWorkerProcess.java:89)
        at 
org.gradle.process.internal.DefaultWorkerProcess.access$000(DefaultWorkerProcess.java:33)
        at 
org.gradle.process.internal.DefaultWorkerProcess$1.executionFinished(DefaultWorkerProcess.java:55)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at 
org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:35)
        at 
org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24)
        at 
org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java:81)
        at 
org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java:30)
        at 
org.gradle.messaging.dispatch.ProxyDispatchAdapter$DispatchingInvocationHandler.invoke(ProxyDispatchAdapter.java:93)
        at com.sun.proxy.$Proxy46.executionFinished(Unknown Source)
        at 
org.gradle.process.internal.DefaultExecHandle.setEndStateInfo(DefaultExecHandle.java:212)
        at 
org.gradle.process.internal.DefaultExecHandle.finished(DefaultExecHandle.java:309)
        at 
org.gradle.process.internal.ExecHandleRunner.completed(ExecHandleRunner.java:108)
        at 
org.gradle.process.internal.ExecHandleRunner.run(ExecHandleRunner.java:88)
        at 
org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecutorImpl$1.run(DefaultExecutorFactory.java:66)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
        at java.lang.Thread.run(Thread.java:744)
OpenJDK 64-Bit Server VM warning: INFO: os::commit_memory(0x000000070a6c0000, 
3946053632, 0) failed; error='Cannot allocate memory' (errno=12)
#
# There is insufficient memory for the Java Runtime Environment to continue.
# Native memory allocation (malloc) failed to allocate 3946053632 bytes for 
committing reserved memory.
# An error report file with more information is saved as:
# /home/ubuntu/incubator-samza/samza-kafka/hs_err_pid2518.log
Could not write standard input into: Gradle Worker 14.
java.io.IOException: Broken pipe
        at java.io.FileOutputStream.writeBytes(Native Method)
        at java.io.FileOutputStream.write(FileOutputStream.java:345)
        at 
java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82)
        at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140)
        at 
org.gradle.process.internal.streams.ExecOutputHandleRunner.run(ExecOutputHandleRunner.java:53)
        at 
org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecutorImpl$1.run(DefaultExecutorFactory.java:66)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
        at java.lang.Thread.run(Thread.java:744)
Process 'Gradle Worker 14' finished with non-zero exit value 1
org.gradle.process.internal.ExecException: Process 'Gradle Worker 14' finished 
with non-zero exit value 1
        at 
org.gradle.process.internal.DefaultExecHandle$ExecResultImpl.assertNormalExitValue(DefaultExecHandle.java:362)
        at 
org.gradle.process.internal.DefaultWorkerProcess.onProcessStop(DefaultWorkerProcess.java:89)
        at 
org.gradle.process.internal.DefaultWorkerProcess.access$000(DefaultWorkerProcess.java:33)
        at 
org.gradle.process.internal.DefaultWorkerProcess$1.executionFinished(DefaultWorkerProcess.java:55)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at 
org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:35)
        at 
org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24)
        at 
org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java:81)
        at 
org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java:30)
        at 
org.gradle.messaging.dispatch.ProxyDispatchAdapter$DispatchingInvocationHandler.invoke(ProxyDispatchAdapter.java:93)
        at com.sun.proxy.$Proxy46.executionFinished(Unknown Source)
        at 
org.gradle.process.internal.DefaultExecHandle.setEndStateInfo(DefaultExecHandle.java:212)
        at 
org.gradle.process.internal.DefaultExecHandle.finished(DefaultExecHandle.java:309)
        at 
org.gradle.process.internal.ExecHandleRunner.completed(ExecHandleRunner.java:108)
        at 
org.gradle.process.internal.ExecHandleRunner.run(ExecHandleRunner.java:88)
        at 
org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecutorImpl$1.run(DefaultExecutorFactory.java:66)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
        at java.lang.Thread.r

Do I need more memory for my machines? Each already has 4GB. I really need to 
have this running. I’m not sure which way is best http or hdfs which one you 
suggest and how can i solve my problem for each case.

Thanks in advance and sorry for bothering this much.
On 10 Aug 2014, at 00:20, Telles Nobrega <[email protected]> wrote:

> Hi Chris, now I have the tar file in my RM machine, and the yarn path points 
> to it. I changed the core-site.xml to use HttpFileSystem instead of HDFS now 
> it is failing with
> 
> Application application_1407640485281_0001 failed 2 times due to AM Container 
> for appattempt_1407640485281_0001_000002 exited with exitCode:-1000 due to: 
> java.lang.ClassNotFoundException: Class 
> org.apache.samza.util.hadoop.HttpFileSystem not found
> 
> I think I can solve this just installing scala files from the samza tutorial, 
> can you confirm that?
> 
> On 09 Aug 2014, at 08:34, Telles Nobrega <[email protected]> wrote:
> 
>> Hi Chris,
>> 
>> I think the problem is that I forgot to update the yarn.job.package.
>> I will try again to see if it works now.
>> 
>> I have one more question, how can I stop (command line) the jobs running in 
>> my topology, for the experiment that I will run, I need to run the same job 
>> in 4 minutes intervals. So I need to kill it, clean the kafka topics and 
>> rerun.
>> 
>> Thanks in advance.
>> 
>> On 08 Aug 2014, at 12:41, Chris Riccomini <[email protected]> 
>> wrote:
>> 
>>> Hey Telles,
>>> 
>>>>> Do I need to have the job folder on each machine in my cluster?
>>> 
>>> No, you should not need to do this. There are two ways to deploy your
>>> tarball to the YARN grid. One is to put it in HDFS, and the other is to
>>> put it on an HTTP server. The link to running a Samza job in a multi-node
>>> YARN cluster describes how to do both (either HTTP server or HDFS).
>>> 
>>> In both cases, once the tarball is put in on the HTTP/HDFS server(s), you
>>> must update yarn.package.path to point to it. From there, the YARN NM
>>> should download it for you automatically when you start your job.
>>> 
>>> * Can you send along a paste of your job config?
>>> 
>>> Cheers,
>>> Chris
>>> 
>>> On 8/8/14 8:04 AM, "Claudio Martins" <[email protected]> wrote:
>>> 
>>>> Hi Telles, it looks to me that you forgot to update the
>>>> "yarn.package.path"
>>>> attribute in your config file for the task.
>>>> 
>>>> - Claudio Martins
>>>> Head of Engineering
>>>> MobileAware USA Inc. / www.mobileaware.com
>>>> office: +1  617 986 5060 / mobile: +1 617 480 5288
>>>> linkedin: www.linkedin.com/in/martinsclaudio
>>>> 
>>>> 
>>>> On Fri, Aug 8, 2014 at 10:55 AM, Telles Nobrega <[email protected]>
>>>> wrote:
>>>> 
>>>>> Hi,
>>>>> 
>>>>> this is my first time trying to run a job on a multinode environment. I
>>>>> have the cluster set up, I can see in the GUI that all nodes are
>>>>> working.
>>>>> Do I need to have the job folder on each machine in my cluster?
>>>>> - The first time I tried running with the job on the namenode machine
>>>>> and
>>>>> it failed saying:
>>>>> 
>>>>> Application application_1407509228798_0001 failed 2 times due to AM
>>>>> Container for appattempt_1407509228798_0001_000002 exited with exitCode:
>>>>> -1000 due to: File
>>>>> 
>>>>> 
>>>>> file:/home/ubuntu/alarm-samza/samza-job-package/target/samza-job-package-
>>>>> 0.7.0-dist.tar.gz
>>>>> does not exist
>>>>> 
>>>>> So I copied the folder to each machine in my cluster and got this error:
>>>>> 
>>>>> Application application_1407509228798_0002 failed 2 times due to AM
>>>>> Container for appattempt_1407509228798_0002_000002 exited with exitCode:
>>>>> -1000 due to: Resource
>>>>> 
>>>>> 
>>>>> file:/home/ubuntu/alarm-samza/samza-job-package/target/samza-job-package-
>>>>> 0.7.0-dist.tar.gz
>>>>> changed on src filesystem (expected 1407509168000, was 1407509434000
>>>>> 
>>>>> What am I missing?
>>>>> 
>>>>> p.s.: I followed this
>>>>> <https://github.com/yahoo/samoa/wiki/Executing-SAMOA-with-Apache-Samza>
>>>>> tutorial
>>>>> and this
>>>>> <
>>>>> 
>>>>> http://samza.incubator.apache.org/learn/tutorials/0.7.0/run-in-multi-node
>>>>> -yarn.html
>>>>>> 
>>>>> to
>>>>> set up the cluster.
>>>>> 
>>>>> Help is much appreciated.
>>>>> 
>>>>> Thanks in advance.
>>>>> 
>>>>> --
>>>>> ------------------------------------------
>>>>> Telles Mota Vidal Nobrega
>>>>> M.sc. Candidate at UFCG
>>>>> B.sc. in Computer Science at UFCG
>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG
>>>>> 
>>> 
>> 
> 

Reply via email to