Hi Jay, Are you able to look at the logs or the web interface? Can you find out why it's getting killed?
Also, can you verify that these ports are open and a process is connected to them (maybe with netstat)? http://www.cloudera.com/blog/2009/08/hadoop-default-ports-quick-reference/ Thanks, Tom On Fri, Oct 28, 2011 at 7:57 PM, Jay Vyas <[email protected]> wrote: > Thanks tom : Thats interesting.... > > First, I tried, and it complained that the input directory didnt exist, so I > ran > $> hadoop fs -mkdir /user/cloudera/input > > Then, I tried to do this : > > $> hadoop jar /usr/lib/hadoop-0.20/hadoop-examples.jar grep input output2 > 'dfs[a-z.]+' > > And it seemed to start working ...... But then it abruptly printed "killed" > somehow at the end of the job [scroll down] ? > > Maybe this is related to why i cant connect ..... ?! > > 1) the hadoop jar 11/10/14 21:34:43 WARN util.NativeCodeLoader: Unable to > load native-hadoop library for your platform... using builtin-java classes > where applicable > 11/10/14 21:34:43 WARN snappy.LoadSnappy: Snappy native library not loaded > 11/10/14 21:34:43 INFO mapred.FileInputFormat: Total input paths to process > : 0 > 11/10/14 21:34:44 INFO mapred.JobClient: Running job: job_201110142010_0009 > 11/10/14 21:34:45 INFO mapred.JobClient: map 0% reduce 0% > 11/10/14 21:34:55 INFO mapred.JobClient: map 0% reduce 100% > 11/10/14 21:34:57 INFO mapred.JobClient: Job complete: job_201110142010_0009 > 11/10/14 21:34:57 INFO mapred.JobClient: Counters: 14 > 11/10/14 21:34:57 INFO mapred.JobClient: Job Counters > 11/10/14 21:34:57 INFO mapred.JobClient: Launched reduce tasks=1 > 11/10/14 21:34:57 INFO mapred.JobClient: SLOTS_MILLIS_MAPS=5627 > 11/10/14 21:34:57 INFO mapred.JobClient: Total time spent by all reduces > waiting after reserving slots (ms)=0 > 11/10/14 21:34:57 INFO mapred.JobClient: Total time spent by all maps > waiting after reserving slots (ms)=0 > 11/10/14 21:34:57 INFO mapred.JobClient: SLOTS_MILLIS_REDUCES=5050 > 11/10/14 21:34:57 INFO mapred.JobClient: FileSystemCounters > 11/10/14 21:34:57 INFO mapred.JobClient: FILE_BYTES_WRITTEN=53452 > 11/10/14 21:34:57 INFO mapred.JobClient: HDFS_BYTES_WRITTEN=86 > 11/10/14 21:34:57 INFO mapred.JobClient: Map-Reduce Framework > 11/10/14 21:34:57 INFO mapred.JobClient: Reduce input groups=0 > 11/10/14 21:34:57 INFO mapred.JobClient: Combine output records=0 > 11/10/14 21:34:57 INFO mapred.JobClient: Reduce shuffle bytes=0 > 11/10/14 21:34:57 INFO mapred.JobClient: Reduce output records=0 > 11/10/14 21:34:57 INFO mapred.JobClient: Spilled Records=0 > 11/10/14 21:34:57 INFO mapred.JobClient: Combine input records=0 > 11/10/14 21:34:57 INFO mapred.JobClient: Reduce input records=0 > 11/10/14 21:34:57 WARN mapred.JobClient: Use GenericOptionsParser for > parsing the arguments. Applications should implement Tool for the same. > 11/10/14 21:34:58 INFO mapred.FileInputFormat: Total input paths to process > : 1 > 11/10/14 21:34:58 INFO mapred.JobClient: Running job: job_201110142010_0010 > 11/10/14 21:34:59 INFO mapred.JobClient: map 0% reduce 0% > Killed > > > On Fri, Oct 28, 2011 at 8:24 PM, Tom Melendez <[email protected]> wrote: > >> Hi Jay, >> >> Some questions for you: >> >> - Does the hadoop client itself work from that same machine? >> - Are you actually able to run the hadoop example jar (in other words, >> your setup is valid otherwise)? >> - Is port 8020 actually available? (you can telnet or nc to it?) >> - What does jps show on the namenode? >> >> Thanks, >> >> Tom >> >> On Fri, Oct 28, 2011 at 4:04 PM, Jay Vyas <[email protected]> wrote: >> > Hi guys : Made more progress debugging my hadoop connection, but still >> > haven't got it working...... It looks like my VM (cloudera hadoop) won't >> > let me in. I find that there is no issue connecting to the name node - >> that >> > is , using hftp and 50070...... >> > >> > via standard HFTP as in here : >> > >> > //This method works fine - connecting directly to hadoop's namenode and >> > querying the filesystem >> > public static void main1(String[] args) throws Exception >> > { >> > String uri = "hftp://155.37.101.76:50070/"; >> > >> > System.out.println( "uri: " + uri ); >> > Configuration conf = new Configuration(); >> > >> > FileSystem fs = FileSystem.get( URI.create( uri ), conf ); >> > fs.printStatistics(); >> > } >> > >> > >> > But unfortunately, I can't get into hdfs ..... Any thoughts on this ? I >> am >> > modifying the uri to access port 8020 >> > which is what is in my core-site.xml . >> > >> > // This fails, resulting (trys to connect over and over again, >> eventually >> > gives up printing "already tried to connect 20 times"....) >> > public static void main(String[] args) >> > { >> > try { >> > String uri = "hdfs://155.37.101.76:8020/"; >> > >> > System.out.println( "uri: " + uri ); >> > Configuration conf = new Configuration(); >> > >> > FileSystem fs = FileSystem.get( URI.create( uri ), conf ); >> > fs.printStatistics(); >> > } catch (Exception e) { >> > // TODO Auto-generated catch block >> > e.printStackTrace(); >> > } >> > } >> > >> > The error message is : >> > >> > 11/10/28 19:03:38 INFO ipc.Client: Retrying connect to server: / >> > 155.37.101.76:8020. Already tried 0 time(s). >> > 11/10/28 19:03:39 INFO ipc.Client: Retrying connect to server: / >> > 155.37.101.76:8020. Already tried 1 time(s). >> > 11/10/28 19:03:40 INFO ipc.Client: Retrying connect to server: / >> > 155.37.101.76:8020. Already tried 2 time(s). >> > 11/10/28 19:03:41 INFO ipc.Client: Retrying connect to server: / >> > 155.37.101.76:8020. Already tried 3 time(s). >> > >> > Any thoughts on this would be *really* be appreciated ... Thanks guys. >> > >> > > > > -- > Jay Vyas > MMSB/UCHC >
