Can you please send the output of 'describe extended activity_test'. This will 
help us understand what's happening with all the create table parameters.

Also - as a sanity check - can you please check hadoop dfs -cat  /data/sample/* 
(to make sure data got loaded/moved into that dir)

-----Original Message-----
From: Josh Ferguson [mailto:[EMAIL PROTECTED] 
Sent: Tuesday, November 25, 2008 7:03 PM
To: [email protected]
Subject: Re: Trouble Loading Into External Table

hive> CREATE EXTERNAL TABLE activity_test
     > (occurred_at INT, actor_id INT, actee_id INT, properties  
MAP<STRING, STRING>)
     > ROW FORMAT DELIMITED
     > FIELDS TERMINATED BY '124'
     > COLLECTION ITEMS TERMINATED BY '44'
     > MAP KEYS TERMINATED BY '58'
     > LINES TERMINATED BY '10'
     > STORED AS TEXTFILE
     > LOCATION '/data/sample';
OK

hive> LOAD DATA LOCAL INPATH '/Users/josh/Hive/sample.tab' INTO TABLE  
activity_test;
Copying data from file:/Users/josh/Hive/sample.tab
Loading data to table activity_test
OK

$ hadoop fs -cat /data/sample/sample.tab
1227422134|2|1|paid:44519,tax:2120,value:42399

hive> FROM activity_test INSERT OVERWRITE DIRECTORY '/data/output2'  
SELECT activity_test.occurred_at, activity_test.actor_id,  
activity_test.actee_id, activity_test.properties;
Total MapReduce jobs = 1
Starting Job = job_200811250653_0022, Tracking URL = http://{clipped}: 
50030/jobdetails.jsp?jobid=job_200811250653_0022
Kill Command = /Users/josh/Hadoop/bin/hadoop job  - 
Dmapred.job.tracker={clipped}:54311 -kill job_200811250653_0022
  map = 0%,  reduce =0%
  map = 50%,  reduce =0%
  map = 100%,  reduce =0%
Ended Job = job_200811250653_0022
Moving data to: /data/output2
OK

$ hadoop fs -cat /data/output2/*
012{}

Still getting incorrect results, is there anything else I could try?

Josh Ferguson

On Nov 25, 2008, at 6:34 PM, Ashish Thusoo wrote:

> Can you try putting the ascii value within quotes, so for example  
> FIELDS TERMINATED BY '124' etc...
>
> You can also look at the following file in the source to see an  
> example of how this is done
>
> ql/src/test/queries/clientpositive/input_dynamicserde.q
>
> Ashish
>
> -----Original Message-----
> From: Josh Ferguson [mailto:[EMAIL PROTECTED]
> Sent: Tuesday, November 25, 2008 6:18 PM
> To: [email protected]
> Subject: Trouble Loading Into External Table
>
> Ok so I'm trying to create an external table and load a delimited  
> file into it, then just do a basic select out of it, here is a  
> description of my scenario along with steps and results I took.
> Hopefully someone can help me figure out what I'm doing wrong.
>
> # Sample.tab
>
> 1227422134|2|1|paid:44519,tax:2120,value:42399
>
> # CREATE TABLE
>
> hive> CREATE EXTERNAL TABLE activity_test
>> (occurred_at INT, actor_id INT, actee_id INT, properties  
>> MAP<STRING, STRING>)
>> ROW FORMAT DELIMITED
>> FIELDS TERMINATED BY "|"
>> COLLECTION ITEMS TERMINATED BY ","
>> MAP KEYS TERMINATED BY ":"
>> LOCATION '/data/sample';
> OK
>
> # LOAD DATA
>
> hive> LOAD DATA LOCAL INPATH '/Users/josh/Hive/sample.tab' INTO TABLE
> activity_test;
> Copying data from file:/Users/josh/Hive/sample.tab Loading data to  
> table activity_test OK
>
> # SELECT OVERWRITE DIRECTORY
>
> hive> FROM activity_test INSERT OVERWRITE DIRECTORY '/data/output'
> SELECT activity_test.occurred_at, activity_test.actor_id,  
> activity_test.actee_id, activity_test.properties; Total MapReduce  
> jobs = 1 Starting Job = job_200811250653_0018, Tracking URL =  
> http://{clipped}:
> 50030/jobdetails.jsp?jobid=job_200811250653_0018
> Kill Command = /Users/josh/Hadoop/bin/hadoop job  -
> Dmapred.job.tracker={clipped}:54311 -kill job_200811250653_0018
>   map = 0%,  reduce =0%
>   map = 50%,  reduce =0%
>   map = 100%,  reduce =0%
> Ended Job = job_200811250653_0018
> Moving data to: /data/output
> OK
> Time taken: 72.329 seconds
>
> $ hadoop fs -cat /data/output/*
> 012{}
>
> This obviously isn't the correct output, and are just some default  
> values for those columns, what am I doing wrong?
>
> Thanks
>
> Josh Ferguson

Reply via email to