Hari Sankar Sivarama Subramaniyan created HIVE-10919:
--------------------------------------------------------
Summary: Windows: create table with JsonSerDe failed via beeline
unless you add hcatalog core jar to classpath
Key: HIVE-10919
URL: https://issues.apache.org/jira/browse/HIVE-10919
Project: Hive
Issue Type: Bug
Reporter: Hari Sankar Sivarama Subramaniyan
Assignee: Hari Sankar Sivarama Subramaniyan
Before we run HiveServer2 tests, we create table via beeline.
And 'create table' with JsonSerDe failed on Winodws. It works on Linux:
{noformat}
0: jdbc:hive2://localhost:10001> create external table all100kjson(
0: jdbc:hive2://localhost:10001> s string,
0: jdbc:hive2://localhost:10001> i int,
0: jdbc:hive2://localhost:10001> d double,
0: jdbc:hive2://localhost:10001> m map<string, string>,
0: jdbc:hive2://localhost:10001> bb array<struct<a: int, b: string>>,
0: jdbc:hive2://localhost:10001> t timestamp)
0: jdbc:hive2://localhost:10001> row format serde
'org.apache.hive.hcatalog.data.JsonSerDe'
0: jdbc:hive2://localhost:10001> WITH SERDEPROPERTIES
('timestamp.formats'='yyyy-MM-dd\'T\'HH:mm:ss')
0: jdbc:hive2://localhost:10001> STORED AS TEXTFILE
0: jdbc:hive2://localhost:10001> location '/user/hcat/tests/data/all100kjson';
Error: Error while processing statement: FAILED: Execution Error, return code 1
from org.apache.hadoop.hive.ql.exec.DDLT
ask. Cannot validate serde: org.apache.hive.hcatalog.data.JsonSerDe
(state=08S01,code=1)
{noformat}
hive.log shows:
{noformat}
2015-05-21 21:59:17,004 ERROR operation.Operation (SQLOperation.java:run(209))
- Error running hive query:
org.apache.hive.service.cli.HiveSQLException: Error while processing statement:
FAILED: Execution Error, return code 1 from
org.apache.hadoop.hive.ql.exec.DDLTask. Cannot validate serde:
org.apache.hive.hcatalog.data.JsonSerDe
at
org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:315)
at
org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:156)
at
org.apache.hive.service.cli.operation.SQLOperation.access$100(SQLOperation.java:71)
at
org.apache.hive.service.cli.operation.SQLOperation$1$1.run(SQLOperation.java:206)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
at
org.apache.hive.service.cli.operation.SQLOperation$1.run(SQLOperation.java:218)
at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Cannot validate
serde: org.apache.hive.hcatalog.data.JsonSerDe
at
org.apache.hadoop.hive.ql.exec.DDLTask.validateSerDe(DDLTask.java:3871)
at org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4011)
at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:306)
at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:160)
at
org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:88)
at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1650)
at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1409)
at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1192)
at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1059)
at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1054)
at
org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:154)
... 11 more
Caused by: java.lang.ClassNotFoundException: Class
org.apache.hive.hcatalog.data.JsonSerDe not found
at
org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:2101)
at
org.apache.hadoop.hive.ql.exec.DDLTask.validateSerDe(DDLTask.java:3865)
... 21 more
{noformat}
If you do add the hcatalog jar to classpath, it works:
{noformat}0: jdbc:hive2://localhost:10001> add jar
hdfs:///tmp/testjars/hive-hcatalog-core-1.2.0.2.3.0.0-2079.jar;
INFO : converting to local
hdfs:///tmp/testjars/hive-hcatalog-core-1.2.0.2.3.0.0-2079.jar
INFO : Added
[/C:/Users/hadoop/AppData/Local/Temp/bc941dac-3bca-4287-a490-8a65c2dac220_resources/hive-hcatalog-core-1.2
.0.2.3.0.0-2079.jar] to class path
INFO : Added resources:
[hdfs:///tmp/testjars/hive-hcatalog-core-1.2.0.2.3.0.0-2079.jar]
No rows affected (0.304 seconds)
0: jdbc:hive2://localhost:10001> create external table all100kjson(
0: jdbc:hive2://localhost:10001> s string,
0: jdbc:hive2://localhost:10001> i int,
0: jdbc:hive2://localhost:10001> d double,
0: jdbc:hive2://localhost:10001> m map<string, string>,
0: jdbc:hive2://localhost:10001> bb array<struct<a: int, b: string>>,
0: jdbc:hive2://localhost:10001> t timestamp)
0: jdbc:hive2://localhost:10001> row format serde
'org.apache.hive.hcatalog.data.JsonSerDe'
0: jdbc:hive2://localhost:10001> WITH SERDEPROPERTIES
('timestamp.formats'='yyyy-MM-dd\'T\'HH:mm:ss')
0: jdbc:hive2://localhost:10001> STORED AS TEXTFILE
0: jdbc:hive2://localhost:10001> location '/user/hcat/tests/data/all100kjson';
No rows affected (2.141 seconds)
{noformat}
Thanks to [~taksaito] for uncovering this issue!
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)