[ 
https://issues.apache.org/jira/browse/HIVE-13976?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15335744#comment-15335744
 ] 

Kai Sasaki commented on HIVE-13976:
-----------------------------------

[~shenavandeh] Thanks for checking but I'm not sure HIVE-2 is related to this 
issue. 
Do you mean simply Hive version 2 instead of JIRA ticket HIVE-2?

> UNION ALL which takes actual source table in one side failed
> ------------------------------------------------------------
>
>                 Key: HIVE-13976
>                 URL: https://issues.apache.org/jira/browse/HIVE-13976
>             Project: Hive
>          Issue Type: Bug
>    Affects Versions: 0.13.0
>         Environment: Ubuntu 12.04, JDK 7
>            Reporter: Kai Sasaki
>
> UNION ALL must take actual source table in both side or none exclusively.
> * UNION ALL with actual table in both side -> Succeed as expected
> {code}
> SELECT 
>   1 AS id,
>   'Alice' AS name
> FROM
>   table1
> UNION ALL 
> SELECT 
>   2 AS id,
>   'Bob' AS name
> FROM
>   table2
> {code}
> * UNION ALL without actual table in both side -> Succeed as expected
> {code}
> SELECT 
>   1 AS id,
>   'Alice' AS name
> UNION ALL 
> SELECT 
>   2 AS id,
>   'Bob' AS name
> {code}
> * UNION ALL with actual table on one side -> Failed
> {code}
> SELECT 
>   1 AS id,
>   'Alice' AS name
> UNION ALL 
> SELECT 
>   2 AS id,
>   'Bob' AS name
> FROM
>    some_table
> {code}
> The error message from map task of third case is this.
> {code}
> Diagnostic Messages for this Task:
> Error: java.lang.IllegalArgumentException: Can not create a Path from an 
> empty string
>       at org.apache.hadoop.fs.Path.checkPathArg(Path.java:127)
>       at org.apache.hadoop.fs.Path.<init>(Path.java:135)
>       at 
> org.apache.hadoop.hive.ql.io.HiveInputFormat$HiveInputSplit.getPath(HiveInputFormat.java:116)
>       at org.apache.hadoop.mapred.MapTask.updateJobWithSplit(MapTask.java:458)
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to