Greetings all, I get the following error at seemingly irregular intervals when I'm trying to do the following...
hadoop fs -put /scratch1/tdp/data/* input (The data is a few hundred files of wikistats data, about 75GB in total). 11/03/04 15:55:05 WARN hdfs.DFSClient: DataStreamer Exception: org.apache.hadoop .ipc.RemoteException: java.io.IOException: File /user/pedersen/input/pagecounts- 20110129-020001 could only be replicated to 0 nodes, instead of 1 I've searched around on the error message, and have actually found a lot of postings, but they seem to be as irregular as the error itself (both in terms of explanations and fixes). http://www.mail-archive.com/[email protected]/msg00407.html http://wiki.apache.org/hadoop/HowToSetupYourDevelopmentEnvironment http://www.phacai.com/hadoop-error-could-only-be-replicated-to-0-nodes-instead-of-1 http://permalink.gmane.org/gmane.comp.jakarta.lucene.hadoop.user/20198 Is there a "best" currently understood explanation for this error, and the preferred way to resolve it? We are running in fully distributed mode here... Thanks! Ted -- Ted Pedersen http://www.d.umn.edu/~tpederse
