hadoop configure effect without restart

2015-12-22 Thread yaoxiaohua
Hi,

 

https://issues.apache.org/jira/browse/HADOOP-7001

 

I also have this kind of need, my environment is hadoop2.3cdh5.0.2, 

This feature is included in this version hadoop?

Or I need patch this?

How to use this feature? Is there a demo?

thanks

 

 

 

Best Regards,

Evan



fs.s3a.endpoint not working

2015-12-22 Thread Phillips, Caleb
Hi All,

New to this list. Looking for a bit of help:

I'm having trouble connecting Hadoop to a S3-compatable (non AWS) object store.

This issue was discussed, but left unresolved, in this thread:

https://mail-archives.apache.org/mod_mbox/spark-user/201507.mbox/%3cca+0w_au5es_flugzmgwkkga3jya1asi3u+isjcuymfntvnk...@mail.gmail.com%3E

And here, on Cloudera's forums (the second post is mine):

https://community.cloudera.com/t5/Data-Ingestion-Integration/fs-s3a-endpoint-ignored-in-hdfs-site-xml/m-p/33694#M1180

I'm running Hadoop 2.6.3 with Java 1.8 (65) on a Linux host. Using Hadoop, I'm 
able to connect to S3 on AWS, and e.g., list/put/get files.

However, when I point the fs.s3a.endpoint configuration directive at my non-AWS 
S3-Compatable object storage, it appears to still point at (and authenticate 
against) AWS.

I've checked and double-checked my credentials and configuration using both 
Python's boto library and the s3cmd tool, both of which connect to this non-AWS 
data store just fine.

Any help would be much appreciated. Thanks!

--
Caleb Phillips, Ph.D.
Data Scientist | Computational Science Center

National Renewable Energy Laboratory (NREL)
15013 Denver West Parkway | Golden, CO 80401
303-275-4297 | caleb.phill...@nrel.gov

-
To unsubscribe, e-mail: user-unsubscr...@hadoop.apache.org
For additional commands, e-mail: user-h...@hadoop.apache.org



Re: diagnosing the difference between dfs 'du' and 'df'

2015-12-22 Thread Anu Engineer
Just  a guess, but could you please check what is your dfs.replication set to ? 

You should be able to find that setting in hdfs-site.xml or in core-site.xml

Thanks
Anu
 

On 12/21/15, 6:21 PM, "Martin Serrano"  wrote:

>Hi,
>
>I have an application that is writing data rapidly directly to HDFS
>(creates and appends) as well as to HBase (10-15 tables).  The disk free
>for the filesystem will report that a large percentage of the system is
>in use:
>
>$ hdfs dfs -df -h /
>Filesystem Size Used  Available  Use%
>hdfs://ha   882.6 G  472.6 G409.9 G   54%
>
>Yet when I try to figure out where the disk space is being used using
>dfs -du reports:
>
>$ hdfs dfs -du -h /
>0/app-logs
>7.6 G/apps
>382.2 M  /hdp
>0/mapred
>0/mr-history
>8.5 K/tmp
>3.8 G/user
>
>A dfsadmin -report during the same time frame is below.  I'm trying to
>figure out where all of this space is going to.  When my application is
>killed or quiescent, the df and dfsadmin reports fall in line with what
>I would expect.  I'm running HDP 2.3 with a default configuration as set
>up by Ambari.  I'm looking for hints or suggestions on how I can
>investigate this issue.  It seems crazy that ingesting 12g or so of data
>can temporarily consume (reserve?) ~300g of HDFS.
>
>Thanks,
>Martin
>
>Configured Capacity: 947644268544 (882.56 GB)
>Present Capacity: 947064596261 (882.02 GB)
>DFS Remaining: 490046627240 (456.39 GB)
>DFS Used: 457017969021 (425.63 GB)
>DFS Used%: 48.26%
>Under replicated blocks: 0
>Blocks with corrupt replicas: 0
>Missing blocks: 0
>Missing blocks (with replication factor 1): 0
>
>-
>Live datanodes (3):
>
>Name: *.*.*.*:50010 (**.com)
>Hostname: **.com
>Decommission Status : Normal
>Configured Capacity: 315881422848 (294.19 GB)
>DFS Used: 218955099179 (203.92 GB)
>Non DFS Used: 168255175 (160.46 MB)
>DFS Remaining: 96758068494 (90.11 GB)
>DFS Used%: 69.32%
>DFS Remaining%: 30.63%
>Configured Cache Capacity: 0 (0 B)
>Cache Used: 0 (0 B)
>Cache Remaining: 0 (0 B)
>Cache Used%: 100.00%
>Cache Remaining%: 0.00%
>Xceivers: 15
>Last contact: Mon Dec 21 17:17:38 EST 2015
>
>
>Name: *.*.*.*:50010 (**.com)
>Hostname: **.com
>Decommission Status : Normal
>Configured Capacity: 315881422848 (294.19 GB)
>DFS Used: 218873337575 (203.84 GB)
>Non DFS Used: 151608508 (144.59 MB)
>DFS Remaining: 96856476765 (90.20 GB)
>DFS Used%: 69.29%
>DFS Remaining%: 30.66%
>Configured Cache Capacity: 0 (0 B)
>Cache Used: 0 (0 B)
>Cache Remaining: 0 (0 B)
>Cache Used%: 100.00%
>Cache Remaining%: 0.00%
>Xceivers: 16
>Last contact: Mon Dec 21 17:17:38 EST 2015
>
>
>Name: *.*.*.*:50010 (*.com)
>Hostname: ***.com
>Decommission Status : Normal
>Configured Capacity: 315881422848 (294.19 GB)
>DFS Used: 19189532267 (17.87 GB)
>Non DFS Used: 259808600 (247.77 MB)
>DFS Remaining: 296432081981 (276.07 GB)
>DFS Used%: 6.07%
>DFS Remaining%: 93.84%
>Configured Cache Capacity: 0 (0 B)
>Cache Used: 0 (0 B)
>Cache Remaining: 0 (0 B)
>Cache Used%: 100.00%
>Cache Remaining%: 0.00%
>Xceivers: 16
>Last contact: Mon Dec 21 17:17:39 EST 2015
>
>
>
>-
>To unsubscribe, e-mail: user-unsubscr...@hadoop.apache.org
>For additional commands, e-mail: user-h...@hadoop.apache.org
>
>


RE: [Classpath Issue]NoClassFoundException occurs when depending on the 3rd jar

2015-12-22 Thread Frank Luo
Make sure you call job.setJarByClass with right parameters. 
http://stackoverflow.com/questions/3912267/hadoop-query-regarding-setjarbyclass-method-of-job-class

Other than that, try to do 2 and 3 together just to test it out. There is no 
reason it doesn’t work.

From: Todd [mailto:bit1...@163.com]
Sent: Tuesday, December 22, 2015 12:01 AM
To: user@hadoop.apache.org
Subject: [Classpath Issue]NoClassFoundException occurs when depending on the 
3rd jar

Hi,
I have two jars, A and B. A contains the class that has the main method, B 
contains the mapper and reducer. A will collect B's mapper and reducer through 
reflection.
I am using the following commands to submit the job,but B's mapper class not 
found exception is thrown.
1.
$HADOOP_CLASSPATH=B.jar:other jars
hadoop jar A.jar

2. HADOOP_CLASSPATH=B.jar:other jars;hadoop jar A.jar
3. hadoop jar A.jar -libjars B.jar

All the above three ways don't work. Can someone help me on this? Thanks!!

This email and any attachments transmitted with it are intended for use by the 
intended recipient(s) only. If you have received this email in error, please 
notify the sender immediately and then delete it. If you are not the intended 
recipient, you must not keep, use, disclose, copy or distribute this email 
without the author’s prior permission. We take precautions to minimize the risk 
of transmitting software viruses, but we advise you to perform your own virus 
checks on any attachment to this message. We cannot accept liability for any 
loss or damage caused by software viruses. The information contained in this 
communication may be confidential and may be subject to the attorney-client 
privilege.


Re: [Classpath Issue]NoClassFoundException occurs when depending on the 3rd jar

2015-12-22 Thread Brian Jeltema
try:
   
   export HADOOP_CLASSPATH=B.jar
   hadoop jar A.jar -libjars B.jar


> On Dec 22, 2015, at 1:01 AM, Todd  wrote:
> 
> Hi,
> I have two jars, A and B. A contains the class that has the main method, B 
> contains the mapper and reducer. A will collect B's mapper and reducer 
> through reflection.
> I am using the following commands to submit the job,but B's mapper class not 
> found exception is thrown.
> 1. 
> $HADOOP_CLASSPATH=B.jar:other jars
> hadoop jar A.jar
> 
> 2. HADOOP_CLASSPATH=B.jar:other jars;hadoop jar A.jar
> 3. hadoop jar A.jar -libjars B.jar
> 
> All the above three ways don't work. Can someone help me on this? Thanks!!
>