Hi tom : which log will have info about why a process was Killed?

Sent from my iPad

On Oct 28, 2011, at 11:41 PM, Tom Melendez <[email protected]> wrote:

> Hi Jay,
> 
> Are you able to look at the logs or the web interface?  Can you find
> out why it's getting killed?
> 
> Also, can you verify that these ports are open and a process is
> connected to them (maybe with netstat)?
> 
> http://www.cloudera.com/blog/2009/08/hadoop-default-ports-quick-reference/
> 
> Thanks,
> 
> Tom
> 
> On Fri, Oct 28, 2011 at 7:57 PM, Jay Vyas <[email protected]> wrote:
>> Thanks tom : Thats interesting....
>> 
>> First, I tried, and it complained that the input directory didnt exist, so I
>> ran
>> $> hadoop fs -mkdir /user/cloudera/input
>> 
>> Then, I tried to do this :
>> 
>> $> hadoop jar /usr/lib/hadoop-0.20/hadoop-examples.jar grep input output2
>> 'dfs[a-z.]+'
>> 
>> And it seemed to start working ...... But then it abruptly printed "killed"
>> somehow at the end of the job [scroll down] ?
>> 
>> Maybe this is related to why i cant connect ..... ?!
>> 
>> 1) the hadoop jar 11/10/14 21:34:43 WARN util.NativeCodeLoader: Unable to
>> load native-hadoop library for your platform... using builtin-java classes
>> where applicable
>> 11/10/14 21:34:43 WARN snappy.LoadSnappy: Snappy native library not loaded
>> 11/10/14 21:34:43 INFO mapred.FileInputFormat: Total input paths to process
>> : 0
>> 11/10/14 21:34:44 INFO mapred.JobClient: Running job: job_201110142010_0009
>> 11/10/14 21:34:45 INFO mapred.JobClient:  map 0% reduce 0%
>> 11/10/14 21:34:55 INFO mapred.JobClient:  map 0% reduce 100%
>> 11/10/14 21:34:57 INFO mapred.JobClient: Job complete: job_201110142010_0009
>> 11/10/14 21:34:57 INFO mapred.JobClient: Counters: 14
>> 11/10/14 21:34:57 INFO mapred.JobClient:   Job Counters
>> 11/10/14 21:34:57 INFO mapred.JobClient:     Launched reduce tasks=1
>> 11/10/14 21:34:57 INFO mapred.JobClient:     SLOTS_MILLIS_MAPS=5627
>> 11/10/14 21:34:57 INFO mapred.JobClient:     Total time spent by all reduces
>> waiting after reserving slots (ms)=0
>> 11/10/14 21:34:57 INFO mapred.JobClient:     Total time spent by all maps
>> waiting after reserving slots (ms)=0
>> 11/10/14 21:34:57 INFO mapred.JobClient:     SLOTS_MILLIS_REDUCES=5050
>> 11/10/14 21:34:57 INFO mapred.JobClient:   FileSystemCounters
>> 11/10/14 21:34:57 INFO mapred.JobClient:     FILE_BYTES_WRITTEN=53452
>> 11/10/14 21:34:57 INFO mapred.JobClient:     HDFS_BYTES_WRITTEN=86
>> 11/10/14 21:34:57 INFO mapred.JobClient:   Map-Reduce Framework
>> 11/10/14 21:34:57 INFO mapred.JobClient:     Reduce input groups=0
>> 11/10/14 21:34:57 INFO mapred.JobClient:     Combine output records=0
>> 11/10/14 21:34:57 INFO mapred.JobClient:     Reduce shuffle bytes=0
>> 11/10/14 21:34:57 INFO mapred.JobClient:     Reduce output records=0
>> 11/10/14 21:34:57 INFO mapred.JobClient:     Spilled Records=0
>> 11/10/14 21:34:57 INFO mapred.JobClient:     Combine input records=0
>> 11/10/14 21:34:57 INFO mapred.JobClient:     Reduce input records=0
>> 11/10/14 21:34:57 WARN mapred.JobClient: Use GenericOptionsParser for
>> parsing the arguments. Applications should implement Tool for the same.
>> 11/10/14 21:34:58 INFO mapred.FileInputFormat: Total input paths to process
>> : 1
>> 11/10/14 21:34:58 INFO mapred.JobClient: Running job: job_201110142010_0010
>> 11/10/14 21:34:59 INFO mapred.JobClient:  map 0% reduce 0%
>> Killed
>> 
>> 
>> On Fri, Oct 28, 2011 at 8:24 PM, Tom Melendez <[email protected]> wrote:
>> 
>>> Hi Jay,
>>> 
>>> Some questions for you:
>>> 
>>> - Does the hadoop client itself work from that same machine?
>>> - Are you actually able to run the hadoop example jar (in other words,
>>> your setup is valid otherwise)?
>>> - Is port 8020 actually available?  (you can telnet or nc to it?)
>>> - What does jps show on the namenode?
>>> 
>>> Thanks,
>>> 
>>> Tom
>>> 
>>> On Fri, Oct 28, 2011 at 4:04 PM, Jay Vyas <[email protected]> wrote:
>>>> Hi guys : Made more progress debugging my hadoop connection, but still
>>>> haven't got it working......  It looks like my VM (cloudera hadoop) won't
>>>> let me in.  I find that there is no issue connecting to the name node -
>>> that
>>>> is , using hftp and 50070......
>>>> 
>>>> via standard HFTP as in here :
>>>> 
>>>> //This method works fine - connecting directly to hadoop's namenode and
>>>> querying the filesystem
>>>> public static void main1(String[] args) throws Exception
>>>>    {
>>>>        String uri = "hftp://155.37.101.76:50070/";;
>>>> 
>>>>        System.out.println( "uri: " + uri );
>>>>        Configuration conf = new Configuration();
>>>> 
>>>>        FileSystem fs = FileSystem.get( URI.create( uri ), conf );
>>>>        fs.printStatistics();
>>>>    }
>>>> 
>>>> 
>>>> But unfortunately, I can't get into hdfs ..... Any thoughts on this ?  I
>>> am
>>>> modifying the uri to access port 8020
>>>> which is what is in my core-site.xml .
>>>> 
>>>>   // This fails, resulting (trys to connect over and over again,
>>> eventually
>>>> gives up printing "already tried to connect 20 times"....)
>>>>    public static void main(String[] args)
>>>>    {
>>>>        try {
>>>>            String uri = "hdfs://155.37.101.76:8020/";
>>>> 
>>>>            System.out.println( "uri: " + uri );
>>>>            Configuration conf = new Configuration();
>>>> 
>>>>            FileSystem fs = FileSystem.get( URI.create( uri ), conf );
>>>>            fs.printStatistics();
>>>>        } catch (Exception e) {
>>>>            // TODO Auto-generated catch block
>>>>            e.printStackTrace();
>>>>        }
>>>>    }
>>>> 
>>>> The error message is :
>>>> 
>>>> 11/10/28 19:03:38 INFO ipc.Client: Retrying connect to server: /
>>>> 155.37.101.76:8020. Already tried 0 time(s).
>>>> 11/10/28 19:03:39 INFO ipc.Client: Retrying connect to server: /
>>>> 155.37.101.76:8020. Already tried 1 time(s).
>>>> 11/10/28 19:03:40 INFO ipc.Client: Retrying connect to server: /
>>>> 155.37.101.76:8020. Already tried 2 time(s).
>>>> 11/10/28 19:03:41 INFO ipc.Client: Retrying connect to server: /
>>>> 155.37.101.76:8020. Already tried 3 time(s).
>>>> 
>>>> Any thoughts on this would be *really* be appreciated  ... Thanks guys.
>>>> 
>>> 
>> 
>> 
>> 
>> --
>> Jay Vyas
>> MMSB/UCHC
>> 

Reply via email to