[ 
https://issues.apache.org/jira/browse/HADOOP-2270?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Karam Singh updated HADOOP-2270:
--------------------------------

    Labels:   (was: newbie)

> Title: DFS submit client params overrides final params on cluster 
> ------------------------------------------------------------------
>
>                 Key: HADOOP-2270
>                 URL: https://issues.apache.org/jira/browse/HADOOP-2270
>             Project: Hadoop Common
>          Issue Type: Bug
>          Components: conf
>    Affects Versions: 0.15.1
>            Reporter: Karam Singh
>
> hdfs client params over-rides the params set as final on hdfs cluster nodes. 
> default valuesv of cleint side hadoop-site.xml values override the final 
> prameters of hdfs hadoop-site.xml .
> oberved the following cases -:
> 1. dfs.trash.root=/recycle, dfs.trash.interval=10 and dfs.replication=2 
> marked final under hadoop-site.xml on hdfs cluster.
>    When fsShel command "hadoop dfs -put local_dir dest" fired from submission 
> host
>    Files will still get replicated 3 times (default) instead of final 
> dfs.replication=2.
>    Similarly when "hadoop dfs -rmr dfs_dir OR hadoop dfs -rm file_path " 
> fired from submit client the file/driectory diectly got deleted without being 
> moved to /recycle.
>    Here hadoop-site.xml on submit client does not specify dfs.trash.root, 
> dfs.trash.interval and dfs.replication.
>   
>    Same is the case when we submit mapred JOB from client and job.xml 
> dispalys default values which overrides the lsuter values.
> 2. dfs.trash.root=/recycle, dfs.trash.interval=10 and dfs.replication=2 
> marked final under hadoop-site.xml on hdfs cluster.
>    And 
>    dfs.trash.root=/rubbish, dfs.trash.interval=2 and dfs.replication=5 under 
> hadoop-site.xml on submit client.
>    When fsShel command "hadoop dfs -put local_dir dest" fired from submit 
> client
>    Files will  get replicated 5 times instead of final dfs.replication=2.
>    Similarly when "hadoop dfs -rmr dfs_dir OR hadoop dfs -rm file_path " 
> fired from submit client the file/driectory diectly will be moved to /rubbish 
> instead of /recycle.
>   
>    Same is the case when we submit mapred job from client, job.xml displays 
> following values -:
>    dfs.trash.root=/rubbish, dfs.trash.interval=2 and dfs.replication=5



--
This message was sent by Atlassian JIRA
(v6.2#6252)

Reply via email to