[ 
https://issues.apache.org/jira/browse/MAPREDUCE-3436?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13189435#comment-13189435
 ] 

Siddharth Seth commented on MAPREDUCE-3436:
-------------------------------------------

@Bruno - "the link was not 0.0.0.0" - that's with and without the patch, 
correct ? With the patch - it goes to the correct history server (second node). 
Without the patch, the history url is the same as the ResourceManager host ?

bq.
There are links to the history for each application. None of them can be 
reached since they all point to the ip 0.0.0.0. For instance:
http://0.0.0.0:8088/proxy/application_1321658790349_0002/jobhistory/job/job_1321658790349_2_2
@Ahmed, that's from the jira description - and I'm assuming that's what is 
being fixed. This exists in a single node cluster.
On the 2 node deployment that you're using - could you try not setting 
"yarn.resourcemanager.address", and "yarn.web-proxy.address" to see if the 
tracker url is something other than 0.0.0.0.
For the scenario Bruno has described in his previous comment - rm address set, 
history address set, history webapp address not set - the patch is good.
                
> jobhistory link may be broken depending on the interface it is listening on
> ---------------------------------------------------------------------------
>
>                 Key: MAPREDUCE-3436
>                 URL: https://issues.apache.org/jira/browse/MAPREDUCE-3436
>             Project: Hadoop Map/Reduce
>          Issue Type: Bug
>          Components: mrv2, webapps
>    Affects Versions: 0.23.0, 0.23.1
>            Reporter: Bruno Mahé
>            Assignee: Ahmed Radwan
>              Labels: bigtop
>         Attachments: MAPREDUCE-3436.patch, MAPREDUCE-3436_rev2.patch
>
>
> On the following page : http://<RESOURCE_MANAGER>:8088/cluster/apps
> There are links to the history for each application. None of them can be 
> reached since they all point to the ip 0.0.0.0. For instance:
> http://0.0.0.0:8088/proxy/application_1321658790349_0002/jobhistory/job/job_1321658790349_2_2
> Am I missing something?
> [root@bigtop-fedora-15 ~]# jps
> 9968 ResourceManager
> 1495 NameNode
> 1645 DataNode
> 12935 Jps
> 11140 -- process information unavailable
> 5309 JobHistoryServer
> 10237 NodeManager
> [root@bigtop-fedora-15 ~]# netstat -tlpn | grep 8088
> tcp        0      0 :::8088                     :::*                        
> LISTEN      9968/java    
> For reference, here is my configuration:
> root@bigtop-fedora-15 ~]# cat /etc/yarn/conf/yarn-site.xml 
> <?xml version="1.0"?>
> <configuration>
> <!-- Site specific YARN configuration properties -->
>    <property>
>       <name>yarn.nodemanager.aux-services</name>
>       <value>mapreduce.shuffle</value>
>     </property>
>     <property>
>       <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
>       <value>org.apache.hadoop.mapred.ShuffleHandler</value>
>     </property>
>     <property>
>       <name>mapreduce.admin.user.env</name>
>       
> <value>CLASSPATH=/etc/hadoop/conf/*:/usr/lib/hadoop/*:/usr/lib/hadoop/lib/*</value>
>     </property>
> </configuration>
> [root@bigtop-fedora-15 ~]# cat /etc/hadoop/conf/hdfs-site.xml 
> <?xml version="1.0"?>
> <configuration>
>   <property>
>     <name>dfs.replication</name>
>     <value>1</value>
>   </property>
>   <property>
>      <name>dfs.permissions</name>
>      <value>false</value>
>   </property>
>   <property>
>      <!-- specify this so that running 'hadoop namenode -format' formats the 
> right dir -->
>      <name>dfs.name.dir</name>
>      <value>/var/lib/hadoop/cache/hadoop/dfs/name</value>
>   </property>
> </configuration>
> [root@bigtop-fedora-15 ~]# cat /etc/hadoop/conf/core-site.xml 
> <?xml version="1.0"?>
> <configuration>
>   <property>
>     <name>fs.default.name</name>
>     <value>hdfs://localhost:8020</value>
>   </property>
>   <property>
>      <name>hadoop.tmp.dir</name>
>      <value>/var/lib/hadoop/cache/${user.name}</value>
>   </property>
>   <!-- OOZIE proxy user setting -->
>   <property>
>     <name>hadoop.proxyuser.oozie.hosts</name>
>     <value>*</value>
>   </property>
>   <property>
>     <name>hadoop.proxyuser.oozie.groups</name>
>     <value>*</value>
>   </property>
> </configuration>

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira


Reply via email to