Re: Scheduling in YARN according to available resources
Hi Tariq, Glad to see that your issue is resolved, thank you. This re-affirms the compatibility issue with openJDK. Thanks Regards, Ravi On Sat, Feb 21, 2015 at 1:40 PM, tesm...@gmail.com wrote: > Dear Nair, > > Your tip in your first email saved my day. Tahnks once again. I am happy > with Oracle JDK. > > Regards, > Tariq > > On Sat, Feb 21, 2015 at 4:05 PM, R Nair > wrote: > >> one of it is in the forum, if you search in google you will get more. I >> am not saying it may not work, but you will have to select and apply some >> patches. One of my friends also had the same problem and with too much >> difficulty, he got this into work. So better avoid :) >> >> https://github.com/elasticsearch/elasticsearch-hadoop/issues/197 >> >> Thanks and regards, >> Nair >> >> On Sat, Feb 21, 2015 at 8:20 AM, tesm...@gmail.com >> wrote: >> >>> Thanks Nair. >>> >>> Managed installing Oracle JDK and it is working great. Thanks for the >>> tip. >>> >>> Any idea why OpenJDK is crashing and Oracle JDK works? >>> >>> Regards, >>> Tariq >>> >>> >>> >>> >>> On Sat, Feb 21, 2015 at 7:14 AM, tesm...@gmail.com >>> wrote: >>> Thanks for your answer Nair, Is installing Oracle JDK on Ubuntu is that complicated as described in this link http://askubuntu.com/questions/56104/how-can-i-install-sun-oracles-proprietary-java-jdk-6-7-8-or-jre Is there an alternate? Regards On Sat, Feb 21, 2015 at 6:50 AM, R Nair wrote: > I had an issue very similar, I changed and used Oracle JDK. There is > nothing I see wrong with your configuration in my first look, thanks > > Regards, > Nair > > On Sat, Feb 21, 2015 at 1:42 AM, tesm...@gmail.com > wrote: > >> I have 7 nodes in my Hadoop cluster [8GB RAM and 4VCPUs to each >> nodes], 1 Namenode + 6 datanodes. >> >> I followed the link from Hortonwroks [ >> http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.0.6.0/bk_installing_manually_book/content/rpm-chap1-11.html >> ] and made calculation according to the hardware configruation on my >> nodes. Added the update mapred-site and yarn-site.xml files in my >> question. >> Still my application is crashing with the same exection >> >> My mapreduce application has 34 input splits with a block size of >> 128MB. >> >> **mapred-site.xml** has the following properties: >> >> mapreduce.framework.name = yarn >> mapred.child.java.opts= -Xmx2048m >> mapreduce.map.memory.mb = 4096 >> mapreduce.map.java.opts = -Xmx2048m >> >> **yarn-site.xml** has the following properties: >> >> yarn.resourcemanager.hostname= hadoop-master >> yarn.nodemanager.aux-services= mapreduce_shuffle >> yarn.nodemanager.resource.memory-mb = 6144 >> yarn.scheduler.minimum-allocation-mb = 2048 >> yarn.scheduler.maximum-allocation-mb = 6144 >> >> >> Exception from container-launch: ExitCodeException exitCode=134: >> /bin/bash: line 1: 3876 Aborted (core dumped) >> /usr/lib/jvm/java-7-openjdk-amd64/bin/java >> -Djava.net.preferIPv4Stack=true >> -Dhadoop.metrics.log.level=WARN -Xmx8192m >> -Djava.io.tmpdir=/tmp/hadoop-ubuntu/nm-local-dir/usercache/ubuntu/appcache/application_1424264025191_0002/container_1424264025191_0002_01_11/tmp >> -Dlog4j.configuration=container-log4j.properties >> -Dyarn.app.container.log.dir=/home/ubuntu/hadoop/logs/userlogs/application_1424264025191_0002/container_1424264025191_0002_01_11 >> -Dyarn.app.container.log.filesize=0 >> -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild >> 192.168.0.12 50842 attempt_1424264025191_0002_m_05_0 11 > >> >> /home/ubuntu/hadoop/logs/userlogs/application_1424264025191_0002/container_1424264025191_0002_01_11/stdout >> 2> >> >> /home/ubuntu/hadoop/logs/userlogs/application_1424264025191_0002/container_1424264025191_0002_01_11/stderr >> >> >> How can avoid this?any help is appreciated >> >> It looks to me that YAN is trying to launch all the container >> simultaneously and anot according to the available resources. Is >> there an option to restrict number of containers on hadoop ndoes? >> >> Regards, >> Tariq >> >> > > > -- > Warmest Regards, > > Ravi Shankar > >>> >> >> >> -- >> Warmest Regards, >> >> Ravi Shankar >> > > -- Warmest Regards, Ravi Shankar
Re: Scheduling in YARN according to available resources
Thanks for your answer Nair, Is installing Oracle JDK on Ubuntu is that complicated as described in this link http://askubuntu.com/questions/56104/how-can-i-install-sun-oracles-proprietary-java-jdk-6-7-8-or-jre Is there an alternate? Regards On Sat, Feb 21, 2015 at 6:50 AM, R Nair wrote: > I had an issue very similar, I changed and used Oracle JDK. There is > nothing I see wrong with your configuration in my first look, thanks > > Regards, > Nair > > On Sat, Feb 21, 2015 at 1:42 AM, tesm...@gmail.com > wrote: > >> I have 7 nodes in my Hadoop cluster [8GB RAM and 4VCPUs to each nodes], 1 >> Namenode + 6 datanodes. >> >> I followed the link from Hortonwroks [ >> http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.0.6.0/bk_installing_manually_book/content/rpm-chap1-11.html >> ] and made calculation according to the hardware configruation on my >> nodes. Added the update mapred-site and yarn-site.xml files in my question. >> Still my application is crashing with the same exection >> >> My mapreduce application has 34 input splits with a block size of 128MB. >> >> **mapred-site.xml** has the following properties: >> >> mapreduce.framework.name = yarn >> mapred.child.java.opts= -Xmx2048m >> mapreduce.map.memory.mb = 4096 >> mapreduce.map.java.opts = -Xmx2048m >> >> **yarn-site.xml** has the following properties: >> >> yarn.resourcemanager.hostname= hadoop-master >> yarn.nodemanager.aux-services= mapreduce_shuffle >> yarn.nodemanager.resource.memory-mb = 6144 >> yarn.scheduler.minimum-allocation-mb = 2048 >> yarn.scheduler.maximum-allocation-mb = 6144 >> >> >> Exception from container-launch: ExitCodeException exitCode=134: >> /bin/bash: line 1: 3876 Aborted (core dumped) >> /usr/lib/jvm/java-7-openjdk-amd64/bin/java >> -Djava.net.preferIPv4Stack=true >> -Dhadoop.metrics.log.level=WARN -Xmx8192m >> -Djava.io.tmpdir=/tmp/hadoop-ubuntu/nm-local-dir/usercache/ubuntu/appcache/application_1424264025191_0002/container_1424264025191_0002_01_11/tmp >> -Dlog4j.configuration=container-log4j.properties >> -Dyarn.app.container.log.dir=/home/ubuntu/hadoop/logs/userlogs/application_1424264025191_0002/container_1424264025191_0002_01_11 >> -Dyarn.app.container.log.filesize=0 >> -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild >> 192.168.0.12 50842 attempt_1424264025191_0002_m_05_0 11 > >> >> /home/ubuntu/hadoop/logs/userlogs/application_1424264025191_0002/container_1424264025191_0002_01_11/stdout >> 2> >> >> /home/ubuntu/hadoop/logs/userlogs/application_1424264025191_0002/container_1424264025191_0002_01_11/stderr >> >> >> How can avoid this?any help is appreciated >> >> It looks to me that YAN is trying to launch all the container >> simultaneously and anot according to the available resources. Is there >> an option to restrict number of containers on hadoop ndoes? >> >> Regards, >> Tariq >> >> > > > -- > Warmest Regards, > > Ravi Shankar >
Re: Scheduling in YARN according to available resources
I had an issue very similar, I changed and used Oracle JDK. There is nothing I see wrong with your configuration in my first look, thanks Regards, Nair On Sat, Feb 21, 2015 at 1:42 AM, tesm...@gmail.com wrote: > I have 7 nodes in my Hadoop cluster [8GB RAM and 4VCPUs to each nodes], 1 > Namenode + 6 datanodes. > > I followed the link from Hortonwroks [ > http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.0.6.0/bk_installing_manually_book/content/rpm-chap1-11.html > ] and made calculation according to the hardware configruation on my > nodes. Added the update mapred-site and yarn-site.xml files in my question. > Still my application is crashing with the same exection > > My mapreduce application has 34 input splits with a block size of 128MB. > > **mapred-site.xml** has the following properties: > > mapreduce.framework.name = yarn > mapred.child.java.opts= -Xmx2048m > mapreduce.map.memory.mb = 4096 > mapreduce.map.java.opts = -Xmx2048m > > **yarn-site.xml** has the following properties: > > yarn.resourcemanager.hostname= hadoop-master > yarn.nodemanager.aux-services= mapreduce_shuffle > yarn.nodemanager.resource.memory-mb = 6144 > yarn.scheduler.minimum-allocation-mb = 2048 > yarn.scheduler.maximum-allocation-mb = 6144 > > > Exception from container-launch: ExitCodeException exitCode=134: > /bin/bash: line 1: 3876 Aborted (core dumped) > /usr/lib/jvm/java-7-openjdk-amd64/bin/java > -Djava.net.preferIPv4Stack=true > -Dhadoop.metrics.log.level=WARN -Xmx8192m > -Djava.io.tmpdir=/tmp/hadoop-ubuntu/nm-local-dir/usercache/ubuntu/appcache/application_1424264025191_0002/container_1424264025191_0002_01_11/tmp > -Dlog4j.configuration=container-log4j.properties > -Dyarn.app.container.log.dir=/home/ubuntu/hadoop/logs/userlogs/application_1424264025191_0002/container_1424264025191_0002_01_11 > -Dyarn.app.container.log.filesize=0 > -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild > 192.168.0.12 50842 attempt_1424264025191_0002_m_05_0 11 > > > /home/ubuntu/hadoop/logs/userlogs/application_1424264025191_0002/container_1424264025191_0002_01_11/stdout > 2> > > /home/ubuntu/hadoop/logs/userlogs/application_1424264025191_0002/container_1424264025191_0002_01_11/stderr > > > How can avoid this?any help is appreciated > > It looks to me that YAN is trying to launch all the container > simultaneously and anot according to the available resources. Is there an > option to restrict number of containers on hadoop ndoes? > > Regards, > Tariq > > -- Warmest Regards, Ravi Shankar