A down grade! I wouldn't have guessed. Thanks Do you know if anything happened to the class MultipleOutputs?
I just tried running some of my old MR code against 1.0 and it seems MultipleOutputs cannot be found in the new hadoop-core-1.0.0.jar. On 2 February 2012 10:45, Harsh J <ha...@cloudera.com> wrote: > Note that 0.21 to 1.0 is "sort-of" a downgrade in some ways, > considering 1.0 is a rename of the 0.20-series. You probably want to > review a lot of config params since those may not be present in 1.0. > > On Thu, Feb 2, 2012 at 11:47 PM, Geoffry Roberts > <geoffry.robe...@gmail.com> wrote: > > All, > > > > I seem to have solved my problem. > > > > In my hdfs.site.xml I had the following: > > > > <property> > > <name>dfs.name.dir</name> > > <value>file:///hdfs/name</ > >> > >> value> > >> </property> > >> <property> > >> <name>dfs.data.dir</name> > >> <value>file:///hdfs/data</value> > >> </property> > > > > The above worked on version 0.21.0, apparently not in 1.0. > > > > I changed them to > > /hdfs/name and /hdfs/data respecively and, well, at least my name node is > > running. > > > > > > On 2 February 2012 09:48, Geoffry Roberts <geoffry.robe...@gmail.com> > wrote: > >> > >> Thanks for the quick response. > >> > >> Here's a snippet from my hdfs.site.xml file. > >> > >> <name>dfs.http.address</name> > >> <value>qq000:50070</value> > >> > >> qq000 is my name node. Is this correct? > >> > >> I have also noticed that my name node is crashing. It says my hdfs is > in > >> a inconsistent state. I guess I'll have to (shudder) rebuild it. > >> > >> The complete contents of hdfs.site.xml is below. > >> > >> <configuration> > >> <property> > >> <name>dfs.replication</name> > >> <value>3</value> > >> <description>Default block replication. > >> The actual number of replications can be specified when the file is > >> created. > >> The default is used if replication is not specified in create time. > >> </description> > >> </property> > >> <property> > >> <name>dfs.name.dir</name> > >> <value>file:///hdfs/name</value> > >> </property> > >> <property> > >> <name>dfs.data.dir</name> > >> <value>file:///hdfs/data</value> > >> </property> > >> <property> > >> <name>dfs.hosts</name> > >> <value>includes</value> > >> <final>true</final> > >> </property> > >> <property> > >> <name>dfs.hosts.exclude</name> > >> <value>excludes</value> > >> <final>true</final> > >> </property> > >> > >> <property> > >> <name>dfs.webhdfs.enabled</name> > >> <value>true</value> > >> </property> > >> <property> > >> <name>dfs.http.address</name> > >> <value>qq000:50070</value> > >> <description>The name of the default file system. Either the > >> literal string "local" or a host:port for NDFS. > >> </description> > >> <final>true</final> > >> </property> > >> </configuration> > >> > >> > >> > >> On 2 February 2012 09:30, Harsh J <ha...@cloudera.com> wrote: > >>> > >>> Geoffry, > >>> > >>> What is your "dfs.http.address" set to? What's your NameNode's HTTP > >>> address, basically? Have you tried that one? > >>> > >>> On Thu, Feb 2, 2012 at 10:54 PM, Geoffry Roberts > >>> <geoffry.robe...@gmail.com> wrote: > >>> > All, > >>> > > >>> > I have been using hadoop 0.21.0 for sometime now. This past Monday I > >>> > installed hadoop 1.0. > >>> > > >>> > I've been reading about WebHDFS and it sounds like something I could > >>> > use but > >>> > I can't seem to get it working. I could definately use some > guidance. > >>> > I can > >>> > find little in the way of documentation. > >>> > > >>> > I added the following property to hdfs_site.xml and bounced hadoop, > but > >>> > nothing seems to be listening on port 50070, which so far a I can > glean > >>> > is > >>> > where WebHDFS should be listening. > >>> > > >>> > <property> > >>> > <name>dfs.webhdfs.enabled</name> > >>> > <value>true</value> > >>> > </property> > >>> > > >>> > Am I on the correct port? Is there anything else? > >>> > > >>> > Thanks > >>> > > >>> > -- > >>> > Geoffry Roberts > >>> > > >>> > >>> > >>> > >>> -- > >>> Harsh J > >>> Customer Ops. Engineer > >>> Cloudera | http://tiny.cloudera.com/about > >> > >> > >> > >> > >> -- > >> Geoffry Roberts > >> > > > > > > > > -- > > Geoffry Roberts > > > > > > -- > Harsh J > Customer Ops. Engineer > Cloudera | http://tiny.cloudera.com/about > -- Geoffry Roberts