Add these file to hive lib folder

>>> hadoop-0.20-core.jar
>>> hive/lib/hive-exec-0.7.1.jar
>>> hive/lib/hive-jdbc-0.7.1.jar
>>> hive/lib/hive-metastore-0.7.1.jar
>>> hive/lib/hive-service-0.7.1.jar
>>> hive/lib/libfb303.jar
>>> lib/commons-logging-1.0.4.jar
>>> slf4j-api-1.6.1.jar
>>> slf4j-log4j12-1.6.1.jar


and then try


On Wed, May 30, 2012 at 2:23 PM, Peyton Peng <pengp...@yunyou.tv> wrote:

>   Hi,
>
> I build the hive table mapped with hbase table,
>
> CREATE TABLE http_access(key string, client_ip string, client_port int,
> request_method string, event_time timestamp)
> STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
> WITH SERDEPROPERTIES (
> "hbase.columns.mapping" =
> ":key,client:ip,client:port,request:method,request:event_time"
> );
>
> The data  is store with hbase client.
>
> I get an issue while query with hive for (hbase/hive integration), while I
> execute sql: select ip, port, request_method from http_access, it works
> well with no problem,
>
> but while I execute below sql: select ip, port, event_time from
> http_access, I got below exception.
>
> The only difference between two sqls is: the event_time is timestamp type, I
> can scan the corresponding hbase table and see the value of event_time is:
>  1338365792142                          column=request:event_time,
> timestamp=1338365739818, value=Wed May 30 16:15:06 CST 2012
>
>
> Anyone who know what the issue is? (Not sure if I made a wrong mapping or
> should I just store the timestamp value as long in hbase? currently I store
> the value as java.util.Date)
>
> Thank you very much....
>
> Regards,
> Peyton
>
> *Exception tracking:*
>
> *Total MapReduce jobs = 1*
> *Launching Job 1 out of 1*
> *Number of reduce tasks is set to 0 since there's no reduce operator*
> *****hdfs://Server:9000/user/hive/warehouse/http_access*
> *Starting Job = job_201205291421_0008, Tracking URL = **
> http://Server:50030/jobdetails.jsp?jobid=job_201205291421_0008*<http://SWHS2.SH.Server:50030/jobdetails.jsp?jobid=job_201205291421_0008>
> *Kill Command = /<here is the hadoop_home>/libexec/../bin/hadoop job
> -Dmapred.job.tracker=Server:9001 �Ckill job_201205291421_0008*
> *Hadoop job information for Stage-1: number of mappers: 1; number of
> reducers: 0*
> *2012-05-30 16:28:01,572 Stage-1 map = 0%,  reduce = 0%*
> *2012-05-30 16:28:34,707 Stage-1 map = 100%,  reduce = 100%*
> *Ended Job = job_201205291421_0008 with errors*
> *Error during job, obtaining debugging information...*
> *Examining task ID: task_201205291421_0008_m_000002 (and more) from job
> job_201205291421_0008*
> *Exception in thread "Thread-211" java.lang.RuntimeException: Error while
> reading from task log url*
> *        at
> org.apache.hadoop.hive.ql.exec.errors.TaskLogProcessor.getErrors(TaskLogProcessor.java:130)
> *
> *        at
> org.apache.hadoop.hive.ql.exec.JobDebugger.showJobFailDebugInfo(JobDebugger.java:211)
> *
> *        at
> org.apache.hadoop.hive.ql.exec.JobDebugger.run(JobDebugger.java:81)*
> *        at java.lang.Thread.run(Thread.java:619)*
> *Caused by: java.io.IOException: Server returned HTTP response code: 400
> for URL: **
> http://Server:50060/tasklog?taskid=attempt_201205291421_0008_m_000000_1&start=-8193
> *<http://SWHS2.SH.Server:50060/tasklog?taskid=attempt_201205291421_0008_m_000000_1&start=-8193>
> *        at
> sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1305)
> *
> *        at java.net.URL.openStream(URL.java:1009)*
> *        at
> org.apache.hadoop.hive.ql.exec.errors.TaskLogProcessor.getErrors(TaskLogProcessor.java:120)
> *
> *        ... 3 more*
> *FAILED: Execution Error, return code 2 from
> org.apache.hadoop.hive.ql.exec.MapRedTask*
> *MapReduce Jobs Launched: *
> *Job 0: Map: 1   HDFS Read: 0 HDFS Write: 0 FAIL*
> *Total MapReduce CPU Time Spent: 0 msec*
> **
>
>
>



-- 


∞
Shashwat Shriparv

Reply via email to