Can you edit your logging configurations to DEBUG, try again and check the logs? Also check hadoop mapper logs.
Thanks and Regards, Sonal On Thu, Feb 18, 2010 at 11:04 AM, prasenjit mukherjee <[email protected]>wrote: > Sorry for the delay. Here is from my /tmp/root/hive.log file. Any other > files I should be looking into. > > 2010-02-18 00:29:56,082 WARN mapred.JobClient > (JobClient.java:configureCommandLineOptions(580)) - Use GenericOptionsParser > for parsing the arguments. Applications should implement Tool for the same. > 2010-02-18 00:30:39,506 ERROR exec.ExecDriver > (SessionState.java:printError(279)) - Ended Job = job_201002171050_0011 with > errors > 2010-02-18 00:30:39,514 ERROR ql.Driver (SessionState.java:printError(279)) > - FAILED: Execution Error, return code 2 from > org.apache.hadoop.hive.ql.exec.ExecDriver > > > > On Wed, Feb 17, 2010 at 6:36 PM, Sonal Goyal <[email protected]>wrote: > >> Hi, >> >> What do your Hive logs say? You can also check the Hadoop mapper and >> reduce job logs. >> >> Thanks and Regards, >> Sonal >> >> >> >> On Wed, Feb 17, 2010 at 4:18 PM, prasenjit mukherjee < >> [email protected]> wrote: >> >>> >>> Here is my std-error : >>> hive> insert overwrite local directory '/tmp/mystuff' select transform(*) >>> using 'my.py' FROM myhivetable; >>> Total MapReduce jobs = 1 >>> Number of reduce tasks is set to 0 since there's no reduce operator >>> Starting Job = job_201002160457_0033, Tracking URL = >>> http://ec2-204-236-205-98.compute-1.amazonaws.com:50030/jobdetails.jsp?jobid=job_201002160457_0033 >>> Kill Command = /usr/lib/hadoop/bin/hadoop job -Dmapred.job.tracker= >>> ec2-204-236-205-98.compute-1.amazonaws.com:8021 -kill >>> job_201002160457_0033 >>> 2010-02-17 05:40:28,380 map = 0%, reduce =0% >>> 2010-02-17 05:41:12,469 map = 100%, reduce =100% >>> Ended Job = job_201002160457_0033 with errors >>> FAILED: Execution Error, return code 2 from >>> org.apache.hadoop.hive.ql.exec.ExecDriver >>> >>> >>> I am trying to use the following command : >>> >>> hive ql : >>> >>> add file /root/my.py >>> insert overwrite local directory '/tmp/mystuff' select transform(*) >>> using 'my.py' FROM myhivetable; >>> >>> and following is my my.py: >>> #!/usr/bin/python >>> import sys >>> for line in sys.stdin: >>> line = line.strip() >>> flds = line.split('\t') >>> (cl_id,cook_id)=flds[:2] >>> sub_id=cl_id >>> if cl_id.startswith('foo'): sub_id=cook_id; >>> print ','.join([sub_id,flds[2],flds[3]]) >>> >>> This works fine, as I tested it in commandline using : echo -e >>> 'aa\tbb\tcc\tdd' | /root/my.py >>> >>> Any pointers ? >>> >> >> >
