Prasad, Isn't the fractional part of the TIMESTAMP type supposed to be optional, as per the error message :
Failed with exception java.io.IOException:java.lang.IllegalArgumentException: Timestamp format must be yyyy-mm-dd hh:mm:ss[.fffffffff] Shall we understand 9 digits for fractional part are mandatory ? Thanks 2013/3/6 Prasad Mujumdar <pras...@cloudera.com> > Dilip, > > Looks like you are using the data from the original schema for this new > table that has single timestamp column. When I tried with just the > timestamp from your data, the query runs fine. I guess the original issue > you hit on the data that didn't have fraction part (1969-12-31 19:00:00, no > .fffff). > > thanks > Prasad > > > On Tue, Mar 5, 2013 at 2:56 PM, Dileep Kumar <dileepkumar...@gmail.com>wrote: > >> --hdfs dfs -mkdir /hive/tpcds/date_ts >> >> create external table date_ts >> ( >> d_date timestamp >> ) >> row format delimited fields terminated by '|' >> location '/hive/tpcds/date_ts'; >> >> [cloudera@localhost tmp-work]$ hive -e "select * from date_ts" >> Logging initialized using configuration in >> file:/etc/hive/conf.dist/hive-log4j.properties >> Hive history >> file=/tmp/cloudera/hive_job_log_cloudera_201303052251_950655265.txt >> OK >> Failed with exception >> java.io.IOException:java.lang.IllegalArgumentException: Timestamp format >> must be yyyy-mm-dd hh:mm:ss[.fffffffff] >> Time taken: 3.556 seconds >> [cloudera@localhost tmp-work]$ hdfs dfs -cat /hive/tpcds/date_ts/* >> 2415022|AAAAAAAAOKJNECAA|1900-01-02 >> 02:00:21.000000000|0|1|1|1900|1|1|2|1|1900|1|1|Monday|1900Q1|N|N|Y|2415021|2415020|2414657|2414930|N|N|N|N|N| >> >> >> >> >> >> On Mon, Mar 4, 2013 at 6:00 PM, Dileep Kumar <dileepkumar...@gmail.com>wrote: >> >>> No. >>> Here are the errors: >>> Task with the most failures(4): >>> ----- >>> Task ID: >>> task_1361599885844_0013_m_000000 >>> >>> URL: >>> >>> http://localhost.localdomain:50030/taskdetails.jsp?jobid=job_1361599885844_0013&tipid=task_1361599885844_0013_m_000000 >>> ----- >>> Diagnostic Messages for this Task: >>> Error: java.lang.RuntimeException: >>> org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while >>> processing row >>> {"d_date_sk":2415022,"d_date_id":"AAAAAAAAOKJNECAA","d_date":"1969-12-31 >>> 19:00:00","d_month_seq":0,"d_week_seq":1,"d_quarter_seq":1,"d_year":1900,"d_dow":1,"d_moy":1,"d_dom":2,"d_qoy":1,"d_fy_year":1900,"d_fy_quarter_seq":1,"d_fy_week_seq":1,"d_day_name":"Monday","d_quarter_name":"1900Q1","d_holiday":"N","d_weekend":"N","d_following_holiday":"Y","d_first_dom":2415021,"d_last_dom":2415020,"d_same_day_ly":2414657,"d_same_day_lq":2414930,"d_current_day":"N","d_current_week":"N","d_current_month":"N","d_current_quarter":"N","d_current_year":"N"} >>> at >>> org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:161) >>> at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54) >>> at >>> org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:399) >>> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:334) >>> at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:152) >>> at java.security.AccessController.doPrivileged(Native Method) >>> at javax.security.auth.Subject.doAs(Subject.java:396) >>> at >>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1332) >>> at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:147) >>> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive >>> Runtime Error while processing row >>> {"d_date_sk":2415022,"d_date_id":"AAAAAAAAOKJNECAA","d_date":"1969-12-31 >>> 19:00:00","d_month_seq":0,"d_week_seq":1,"d_quarter_seq":1,"d_year":1900,"d_dow":1,"d_moy":1,"d_dom":2,"d_qoy":1,"d_fy_year":1900,"d_fy_quarter_seq":1,"d_fy_week_seq":1,"d_day_name":"Monday","d_quarter_name":"1900Q1","d_holiday":"N","d_weekend":"N","d_following_holiday":"Y","d_first_dom":2415021,"d_last_dom":2415020,"d_same_day_ly":2414657,"d_same_day_lq":2414930,"d_current_day":"N","d_current_week":"N","d_current_month":"N","d_current_quarter":"N","d_current_year":"N"} >>> at >>> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:548) >>> at >>> org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:143) >>> ... 8 more >>> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Error >>> evaluating d_date >>> at >>> org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:80) >>> at >>> org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:471) >>> at >>> org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:762) >>> at >>> org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:83) >>> at >>> org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:471) >>> at >>> org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:762) >>> at >>> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:529) >>> ... 9 more >>> Caused by: java.lang.IllegalArgumentException: Timestamp format must be >>> yyyy-mm-dd hh:mm:ss[.fffffffff] >>> at java.sql.Timestamp.valueOf(Timestamp.java:185) >>> at >>> org.apache.hadoop.hive.serde2.lazy.LazyTimestamp.init(LazyTimestamp.java:74) >>> at >>> org.apache.hadoop.hive.serde2.lazy.LazyStruct.uncheckedGetField(LazyStruct.java:219) >>> at >>> org.apache.hadoop.hive.serde2.lazy.LazyStruct.getField(LazyStruct.java:192) >>> at >>> org.apache.hadoop.hive.serde2.lazy.objectinspector.LazySimpleStructObjectInspector.getStructFieldData(LazySimpleStructObjectInspector.java:188) >>> at >>> org.apache.hadoop.hive.ql.exec.ExprNodeColumnEvaluator.evaluate(ExprNodeColumnEvaluator.java:98) >>> at >>> org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:76) >>> ... 15 more >>> >>> >>> FAILED: Execution Error, return code 2 from >>> org.apache.hadoop.hive.ql.exec.MapRedTask >>> MapReduce Jobs Launched: >>> Job 0: Map: 1 HDFS Read: 0 HDFS Write: 0 FAIL >>> Total MapReduce CPU Time Spent: 0 msec >>> >>> >>> >>> On Mon, Mar 4, 2013 at 5:51 PM, Mark Grover <grover.markgro...@gmail.com >>> > wrote: >>> >>>> Hi Dilip, >>>> Are you able to run this query successfully? >>>> >>>> select d_date_sk, d_date_id, d_date, d_month_seq, d_week_seq, >>>> d_quarter_seq, d_dow, d_moy, d_dom, d_qoy, d_fy_year, >>>> d_fy_quarter_seq, d_fy_week_seq, d_day_name, d_quarter_name, >>>> d_holiday, d_weekend, d_following_holiday, d_first_dom, d_last_dom, >>>> d_same_day_ly, d_same_day_lq, d_current_day, d_current_week, >>>> d_current_month, d_current_quarter, d_current_year, d_year >>>> from date_dim >>>> >>>> On Mon, Mar 4, 2013 at 5:37 PM, Dileep Kumar <dileepkumar...@gmail.com> >>>> wrote: >>>> > Hi All, >>>> > >>>> > I am using the schema in the Impala VM and trying to create a dynamic >>>> > partitioned table on date_dim. >>>> > New table is called date_dim_i and schema for that is defined as: >>>> > create table date_dim_i >>>> > ( >>>> > d_date_sk int, >>>> > d_date_id string, >>>> > d_date timestamp, >>>> > d_month_seq int, >>>> > d_week_seq int, >>>> > d_quarter_seq int, >>>> > d_dow int, >>>> > d_moy int, >>>> > d_dom int, >>>> > d_qoy int, >>>> > d_fy_year int, >>>> > d_fy_quarter_seq int, >>>> > d_fy_week_seq int, >>>> > d_day_name string, >>>> > d_quarter_name string, >>>> > d_holiday string, >>>> > d_weekend string, >>>> > d_following_holiday string, >>>> > d_first_dom int, >>>> > d_last_dom int, >>>> > d_same_day_ly int, >>>> > d_same_day_lq int, >>>> > d_current_day string, >>>> > d_current_week string, >>>> > d_current_month string, >>>> > d_current_quarter string, >>>> > d_current_year string >>>> > ) >>>> > PARTITIONED BY (d_year int) >>>> > stored as RCFILE; >>>> > >>>> > Then I do insert overwrite as: >>>> > insert overwrite table date_dim_i >>>> > PARTITION (d_year) >>>> > select d_date_sk, d_date_id, d_date, d_month_seq, d_week_seq, >>>> d_quarter_seq, >>>> > d_dow, d_moy, d_dom, d_qoy, d_fy_year, d_fy_quarter_seq, >>>> d_fy_week_seq, >>>> > d_day_name, d_quarter_name, d_holiday, d_weekend, d_following_holiday, >>>> > d_first_dom, d_last_dom, d_same_day_ly, d_same_day_lq, d_current_day, >>>> > d_current_week, d_current_month, d_current_quarter, d_current_year, >>>> d_year >>>> > from date_dim; >>>> > >>>> > The date_dim table schema is as : >>>> > create external table date_dim >>>> > ( >>>> > d_date_sk int, >>>> > d_date_id string, >>>> > d_date timestamp, >>>> > d_month_seq int, >>>> > d_week_seq int, >>>> > d_quarter_seq int, >>>> > d_year int, >>>> > d_dow int, >>>> > d_moy int, >>>> > d_dom int, >>>> > d_qoy int, >>>> > d_fy_year int, >>>> > d_fy_quarter_seq int, >>>> > d_fy_week_seq int, >>>> > d_day_name string, >>>> > d_quarter_name string, >>>> > d_holiday string, >>>> > d_weekend string, >>>> > d_following_holiday string, >>>> > d_first_dom int, >>>> > d_last_dom int, >>>> > d_same_day_ly int, >>>> > d_same_day_lq int, >>>> > d_current_day string, >>>> > d_current_week string, >>>> > d_current_month string, >>>> > d_current_quarter string, >>>> > d_current_year string >>>> > ) >>>> > row format delimited fields terminated by '|' >>>> > location '/hive/tpcds/date_dim'; >>>> > >>>> > >>>> > >>>> > >>>> > >>>> > It fails with following exception: >>>> > >>>> > Error: java.lang.RuntimeException: >>>> > org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error >>>> while >>>> > processing row >>>> > >>>> {"d_date_sk":2415022,"d_date_id":"AAAAAAAAOKJNECAA","d_date":"1969-12-31 >>>> > >>>> 19:00:00","d_month_seq":0,"d_week_seq":1,"d_quarter_seq":1,"d_year":1900,"d_dow":1,"d_moy":1,"d_dom":2,"d_qoy":1,"d_fy_year":1900,"d_fy_quarter_seq":1,"d_fy_week_seq":1,"d_day_name":"Monday","d_quarter_name":"1900Q1","d_holiday":"N","d_weekend":"N","d_following_holiday":"Y","d_first_dom":2415021,"d_last_dom":2415020,"d_same_day_ly":2414657,"d_same_day_lq":2414930,"d_current_day":"N","d_current_week":"N","d_current_month":"N","d_current_quarter":"N","d_current_year":"N"} >>>> > >>>> > at >>>> > org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:161) >>>> > >>>> > at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54) >>>> > >>>> > at >>>> org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:399) >>>> > >>>> > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:334) >>>> > >>>> > at >>>> org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:152) >>>> > >>>> > at java.security.AccessController.doPrivileged(Native Method) >>>> > >>>> > at javax.security.auth.Subject.doAs(Subject.java:396) >>>> > >>>> > at >>>> > >>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1332) >>>> > >>>> > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:147) >>>> > >>>> > Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive >>>> Runtime >>>> > Error while processing row >>>> > >>>> {"d_date_sk":2415022,"d_date_id":"AAAAAAAAOKJNECAA","d_date":"1969-12-31 >>>> > >>>> 19:00:00","d_month_seq":0,"d_week_seq":1,"d_quarter_seq":1,"d_year":1900,"d_dow":1,"d_moy":1,"d_dom":2,"d_qoy":1,"d_fy_year":1900,"d_fy_quarter_seq":1,"d_fy_week_seq":1,"d_day_name":"Monday","d_quarter_name":"1900Q1","d_holiday":"N","d_weekend":"N","d_following_holiday":"Y","d_first_dom":2415021,"d_last_dom":2415020,"d_same_day_ly":2414657,"d_same_day_lq":2414930,"d_current_day":"N","d_current_week":"N","d_current_month":"N","d_current_quarter":"N","d_current_year":"N"} >>>> > >>>> > at >>>> > >>>> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:548) >>>> > >>>> > at >>>> > org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:143) >>>> > >>>> > ... 8 more >>>> > >>>> > Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Error >>>> > evaluating d_date >>>> > >>>> > at >>>> > >>>> org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:80) >>>> > >>>> > at >>>> > org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:471) >>>> > >>>> > at >>>> > org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:762) >>>> > >>>> > at >>>> > >>>> org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:83) >>>> > >>>> > at >>>> > org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:471) >>>> > >>>> > at >>>> > org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:762) >>>> > >>>> > at >>>> > >>>> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:529) >>>> > >>>> > ... 9 more >>>> > >>>> > Caused by: java.lang.IllegalArgumentException: Timestamp format must >>>> be >>>> > yyyy-mm-dd hh:mm:ss[.fffffffff] >>>> > >>>> > at java.sql.Timestamp.valueOf(Timestamp.java:185) >>>> > >>>> > >>>> > Please suggest what could be wrong here as datatypes are exact same >>>> in both >>>> > cases. >>>> > >>>> > >>>> > Thanks ! >>>> >>> >>> >> >