External Table

2017-08-15 Thread Gayathri Devi
I have a table in MySQL . I sqoop to Hbase.

Then created an external table in Hive.

One column of data type decimal(32,1) from MySQL is not reflecting in Hive.
It's filled with null values

But hbase table has all values.

Only Hive external table one column alone is filled with null values


Review Request 61679: HIVE-17330 refactor TezSessionPoolManager to separate its multiple functions

2017-08-15 Thread Sergey Shelukhin

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/61679/
---

Review request for hive and Siddharth Seth.


Repository: hive-git


Description
---

see jira


Diffs
-

  ql/src/java/org/apache/hadoop/hive/ql/exec/tez/RestrictedConfigChecker.java 
PRE-CREATION 
  ql/src/java/org/apache/hadoop/hive/ql/exec/tez/SessionExpirationTracker.java 
PRE-CREATION 
  ql/src/java/org/apache/hadoop/hive/ql/exec/tez/TezSessionPool.java 
PRE-CREATION 
  ql/src/java/org/apache/hadoop/hive/ql/exec/tez/TezSessionPoolManager.java 
dabca3f4d0 
  ql/src/java/org/apache/hadoop/hive/ql/exec/tez/TezSessionPoolSession.java 
PRE-CREATION 
  ql/src/test/org/apache/hadoop/hive/ql/exec/tez/SampleTezSessionState.java 
2d1c687a4e 
  ql/src/test/org/apache/hadoop/hive/ql/exec/tez/TestTezSessionPool.java 
5a3eba3ad4 


Diff: https://reviews.apache.org/r/61679/diff/1/


Testing
---


Thanks,

Sergey Shelukhin



[jira] [Created] (HIVE-17330) refactor TezSessionPoolManager to separate its multiple functions

2017-08-15 Thread Sergey Shelukhin (JIRA)
Sergey Shelukhin created HIVE-17330:
---

 Summary: refactor TezSessionPoolManager to separate its multiple 
functions
 Key: HIVE-17330
 URL: https://issues.apache.org/jira/browse/HIVE-17330
 Project: Hive
  Issue Type: Bug
Reporter: Sergey Shelukhin


TezSessionPoolManager would retain things specific to current Hive session 
management. 
The session pool itself, as well as expiration tracking, the pool session 
implementation, and some config validation can be separated out and made 
independent from the pool.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


Re: Review Request 61625: HIVE-17308 Improvement in join cardinality estimation

2017-08-15 Thread Ashutosh Chauhan

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/61625/#review183018
---




ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/stats/HiveRelMdSelectivity.java
Line 124 (original), 116 (patched)


better name: ndvEstimate?



ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/stats/HiveRelMdSelectivity.java
Lines 187 (patched)


comment: max ndv across all column references from both sides of table.


- Ashutosh Chauhan


On Aug. 15, 2017, 1:20 a.m., Vineet Garg wrote:
> 
> ---
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/61625/
> ---
> 
> (Updated Aug. 15, 2017, 1:20 a.m.)
> 
> 
> Review request for hive, Ashutosh Chauhan and Jesús Camacho Rodríguez.
> 
> 
> Bugs: HIVE-17308
> https://issues.apache.org/jira/browse/HIVE-17308
> 
> 
> Repository: hive-git
> 
> 
> Description
> ---
> 
> See JIRA description
> 
> 
> Diffs
> -
> 
>   common/src/java/org/apache/hadoop/hive/conf/HiveConf.java 056f2d7834 
>   
> ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/HiveConfPlannerContext.java
>  PRE-CREATION 
>   
> ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/HivePlannerContext.java
>  bdf995548f 
>   
> ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/stats/HiveRelMdSelectivity.java
>  046f51b5a0 
>   ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java d6695ccbf2 
>   
> ql/src/test/org/apache/hadoop/hive/ql/optimizer/calcite/TestCBORuleFiredOnlyOnce.java
>  884e034731 
>   ql/src/test/results/clientpositive/annotate_stats_join.q.out d24324a175 
>   ql/src/test/results/clientpositive/join_alt_syntax.q.out 1c08e6a630 
>   ql/src/test/results/clientpositive/join_cond_pushdown_2.q.out 6b03800d4a 
>   ql/src/test/results/clientpositive/join_cond_pushdown_4.q.out 1408dad546 
>   ql/src/test/results/clientpositive/llap/correlationoptimizer1.q.out 
> 1e4e08453b 
>   ql/src/test/results/clientpositive/llap/dynamic_semijoin_reduction.q.out 
> 478b0828a3 
>   ql/src/test/results/clientpositive/llap/dynamic_semijoin_user_level.q.out 
> a73faebcb7 
>   ql/src/test/results/clientpositive/llap/explainuser_1.q.out b36fe2a83f 
>   ql/src/test/results/clientpositive/llap/explainuser_2.q.out 4f1a4a1f14 
>   ql/src/test/results/clientpositive/llap/join_max_hashtable.q.out 6520fd3f2b 
>   ql/src/test/results/clientpositive/llap/skewjoin.q.out f4fc43edfa 
>   ql/src/test/results/clientpositive/llap/subquery_exists.q.out 4d223923a7 
>   ql/src/test/results/clientpositive/llap/subquery_in.q.out df92bb2b93 
>   ql/src/test/results/clientpositive/llap/subquery_multi.q.out 5673f0ee8c 
>   ql/src/test/results/clientpositive/llap/subquery_notin.q.out 3c6e95ae70 
>   ql/src/test/results/clientpositive/llap/subquery_scalar.q.out 05d4fc8093 
>   ql/src/test/results/clientpositive/llap/subquery_select.q.out 013b0a3bdd 
>   ql/src/test/results/clientpositive/llap/subquery_views.q.out 094d6b29b5 
>   ql/src/test/results/clientpositive/perf/query17.q.out 6c9e00bb6a 
>   ql/src/test/results/clientpositive/perf/query24.q.out 3eb65b67c6 
>   ql/src/test/results/clientpositive/perf/query25.q.out ee94eecc2c 
>   ql/src/test/results/clientpositive/perf/query29.q.out b3b9c3ed53 
>   ql/src/test/results/clientpositive/perf/query50.q.out 68d6611b43 
>   ql/src/test/results/clientpositive/perf/query54.q.out b9d0b8b376 
>   ql/src/test/results/clientpositive/perf/query64.q.out 7464430c03 
>   ql/src/test/results/clientpositive/perf/query72.q.out 4d18fb3a4f 
>   ql/src/test/results/clientpositive/perf/query85.q.out f188fa2b2c 
>   ql/src/test/results/clientpositive/spark/annotate_stats_join.q.out 
> cd24cbb6df 
>   ql/src/test/results/clientpositive/spark/join_alt_syntax.q.out f1728ce1f7 
>   ql/src/test/results/clientpositive/spark/join_cond_pushdown_2.q.out 
> c3b805aeb0 
>   ql/src/test/results/clientpositive/spark/join_cond_pushdown_4.q.out 
> 9c3aed8932 
> 
> 
> Diff: https://reviews.apache.org/r/61625/diff/3/
> 
> 
> Testing
> ---
> 
> 
> Thanks,
> 
> Vineet Garg
> 
>



[jira] [Created] (HIVE-17329) ensure acid side file is not overwritten

2017-08-15 Thread Eugene Koifman (JIRA)
Eugene Koifman created HIVE-17329:
-

 Summary: ensure acid side file is not overwritten
 Key: HIVE-17329
 URL: https://issues.apache.org/jira/browse/HIVE-17329
 Project: Hive
  Issue Type: Improvement
  Components: Transactions
Reporter: Eugene Koifman
Assignee: Eugene Koifman
Priority: Minor
 Fix For: 3.0.0


OrcRecordUpdater() has 
{noformat}
  flushLengths = fs.create(OrcAcidUtils.getSideFile(this.path), true, 8,
  options.getReporter());
{noformat}

this should be the only place where the side file is created but to be safe we 
should set "overwrite" parameter to false.  If this file already exists that 
means there are 2 OrcRecordUpdates trying to write the same (primary) file - 
never ok.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (HIVE-17328) Remove special handling for Acid tables wherever possible

2017-08-15 Thread Eugene Koifman (JIRA)
Eugene Koifman created HIVE-17328:
-

 Summary: Remove special handling for Acid tables wherever possible
 Key: HIVE-17328
 URL: https://issues.apache.org/jira/browse/HIVE-17328
 Project: Hive
  Issue Type: Improvement
  Components: Transactions
Reporter: Eugene Koifman
Assignee: Eugene Koifman


There are various places in the code that do something like 
if(acid update or delete) {
 do something
}
else {
do something else
}

this complicates the code and makes it so that acid code path is not properly 
tested in many new non-acid features or bug fixes.

Some work to simplify this was done in HIVE-15844.

SortedDynPartitionOptimizer has some special logic
ReduceSinkOperator relies on partitioning columns for update/delete be 
UDFToInteger(RecordIdentifier) which is set up in SemanticAnalyzer.  
Consequently SemanticAnalyzer has special logic to set it up.
FileSinkOperator has some specialization.

AbstractCorrelationProcCtx makes changes specific to acid writes setting 
hive.optimize.reducededuplication.min.reducer=1


With acid 2.0 (HIVE-17089) a lot more of it can simplified/removed.
Generally, Acid Insert follows the same code path as regular insert except that 
the writer in FileSinkOperator is Acid specific.
So all the specialization is to route Update/Delete events to the right place.

We can do the U=D+I early in the operator pipeline so that an Update is a Hive 
multi-insert with 1 leg being the Insert leg and the other being the Delete leg 
(like Merge stmt).
The Delete events themselves don't need to be routed in any particular way if 
we always ship all delete_delta files for each split.  This is ok since delete 
events are very small and highly compressible.  What is shipped is independent 
of what needs to be loaded into memory.

This would allow removing almost all special code paths.
If need be we can also have the compactor rewrite the delete files so that the 
name of the file matches the contents and make it as if they were bucketed 
properly and use it reduce what needs to be shipped for each split.  This may 
help with some extreme cases where someone updates 1B rows.




--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (HIVE-17327) LLAP IO: restrict native file ID usage to default FS to avoid hypothetiocal collisions with HDFS federation

2017-08-15 Thread Sergey Shelukhin (JIRA)
Sergey Shelukhin created HIVE-17327:
---

 Summary: LLAP IO: restrict native file ID usage to default FS to 
avoid hypothetiocal collisions with HDFS federation
 Key: HIVE-17327
 URL: https://issues.apache.org/jira/browse/HIVE-17327
 Project: Hive
  Issue Type: Bug
Reporter: Gopal V
Assignee: Sergey Shelukhin






--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (HIVE-17326) Insert into HBase tables fails if hive.llap.execution.mode is set to only

2017-08-15 Thread Sailaja Navvluru (JIRA)
Sailaja Navvluru created HIVE-17326:
---

 Summary: Insert into HBase tables fails if 
hive.llap.execution.mode is set to only
 Key: HIVE-17326
 URL: https://issues.apache.org/jira/browse/HIVE-17326
 Project: Hive
  Issue Type: Bug
  Components: llap
Affects Versions: 2.1.0
 Environment: HDP 2.6.x
Reporter: Sailaja Navvluru


Inserting into a table created using HBase storage handler errors out if 
hive.llap.execution.mode=only. Works if the hive.llap.execution.mode value is 
none or auto or with MR execution engine.
Simple repro script
CREATE TABLE hbase_table_sai(id int, name string)
STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,cf1:name")
TBLPROPERTIES ("hbase.table.name" = "sai");

create table hive_tab1(c1 int, c2 string);
 insert into hive_tab1 values(1,'abc');
0: jdbc:hive2://localhost:10500/default> insert overwrite table hbase_table_sai 
select * from hive_tab1;
INFO  : Compiling 
command(queryId=hive_20170808125431_652dbcde-96d5-4afd-9359-bd71bfd6b01a): 
insert overwrite table hbase_table_sai select * from hive_tab1
INFO  : We are setting the hadoop caller context from 
HIVE_SSN_ID:7114abad-2ba2-410d-ad73-40d473a647af to 
hive_20170808125431_652dbcde-96d5-4afd-9359-bd71bfd6b01a
INFO  : Semantic Analysis Completed
INFO  : Returning Hive schema: 
Schema(fieldSchemas:[FieldSchema(name:hive_tab1.c1, type:int, comment:null), 
FieldSchema(name:hive_tab1.c2, type:string, comment:null)], properties:null)
INFO  : Completed compiling 
command(queryId=hive_20170808125431_652dbcde-96d5-4afd-9359-bd71bfd6b01a); Time 
taken: 0.36 seconds
INFO  : We are resetting the hadoop caller context to 
HIVE_SSN_ID:7114abad-2ba2-410d-ad73-40d473a647af
INFO  : Concurrency mode is disabled, not creating a lock manager
INFO  : Setting caller context to query id 
hive_20170808125431_652dbcde-96d5-4afd-9359-bd71bfd6b01a
INFO  : Executing 
command(queryId=hive_20170808125431_652dbcde-96d5-4afd-9359-bd71bfd6b01a): 
insert overwrite table hbase_table_sai select * from hive_tab1
INFO  : Query ID = hive_20170808125431_652dbcde-96d5-4afd-9359-bd71bfd6b01a
INFO  : Total jobs = 1
INFO  : Starting task [Stage-0:DDL] in serial mode
INFO  : Starting task [Stage-1:DDL] in serial mode
INFO  : Launching Job 1 out of 1
INFO  : Starting task [Stage-3:MAPRED] in serial mode
INFO  : Session is already open
INFO  : Tez session missing resources, adding additional necessary resources
INFO  : Dag name: insert overwrite table hbase_tab...hive_tab1(Stage-3)
INFO  : Dag submit failed due to There is conflicting local resource 
(guava-14.0.1.jar) between dag local resource and vertex Map 1 local resource.
Resource of dag : resource { scheme: "hdfs" host: "ulcer1.fyre.ibm.com" port: 
8020 file: 
"/tmp/hive/hive/7114abad-2ba2-410d-ad73-40d473a647af/hive_2017-08-08_12-54-31_225_8109820757632121978-7/hive/_tez_scratch_dir/guava-14.0.1.jar"
 } size: 2189117 timestamp: 150072247 type: FILE visibility: PRIVATE
Resource of vertex: resource { scheme: "hdfs" host: "ulcer1.fyre.ibm.com" port: 
8020 file: 
"/tmp/hive/hive/_tez_session_dir/8a93f7fd-b925-4684-a6b1-6561b5c8e344/guava-14.0.1.jar"
 } size: 2189117 timestamp: 1502211657919 type: FILE visibility: PRIVATE stack 
trace: [org.apache.tez.dag.api.DAG.verify(DAG.java:695), 
org.apache.tez.dag.api.DAG.createDag(DAG.java:796), 
org.apache.tez.client.TezClientUtils.prepareAndCreateDAGPlan(TezClientUtils.java:718),
 org.apache.tez.client.TezClient.submitDAGSession(TezClient.java:555), 
org.apache.tez.client.TezClient.submitDAG(TezClient.java:522), 
org.apache.hadoop.hive.ql.exec.tez.TezTask.submit(TezTask.java:506), 
org.apache.hadoop.hive.ql.exec.tez.TezTask.execute(TezTask.java:188), 
org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:197), 
org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:100), 
org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1905), 
org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1607), 
org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1354), 
org.apache.hadoop.hive.ql.Driver.run(Driver.java:1123), 
org.apache.hadoop.hive.ql.Driver.run(Driver.java:1116), 
org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:242),
 
org.apache.hive.service.cli.operation.SQLOperation.access$800(SQLOperation.java:91),
 
org.apache.hive.service.cli.operation.SQLOperation$BackgroundWork$1.run(SQLOperation.java:334),
 java.security.AccessController.doPrivileged(Native Method), 
javax.security.auth.Subject.doAs(Subject.java:422), 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1866),
 
org.apache.hive.service.cli.operation.SQLOperation$BackgroundWork.run(SQLOperation.java:348),
 java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511), 
java.util.concurrent.FutureTask.run(FutureTask.java:266), 

[jira] [Created] (HIVE-17325) Clean up intermittently failing uni tests

2017-08-15 Thread Alan Gates (JIRA)
Alan Gates created HIVE-17325:
-

 Summary: Clean up intermittently failing uni tests
 Key: HIVE-17325
 URL: https://issues.apache.org/jira/browse/HIVE-17325
 Project: Hive
  Issue Type: Test
  Components: Tests
Reporter: Alan Gates
Assignee: Alan Gates


We have a number of intermittently failing tests.  I propose to disable these 
so that we can get clean (or at least cleaner) CI runs.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


Re: Running tests against branches other than master.

2017-08-15 Thread Mithun RK
Ah, so. Thank you, Sergey. Let me try this for branch-2.

Mithun

On Tue, Aug 15, 2017 at 12:58 PM, Sergey Shelukhin 
wrote:

> And by jar I mean patch.
>
> On 17/8/15, 12:56, "Sergey Shelukhin"  wrote:
>
> >Specifying the branch name as part of the jar is the trigger. I’ve never
> >tried it for more than one branch at a time so I’m not sure if HiveQA
> >would run for every one; it seems more likely that it would pick the
> >latest patch, whatever that happens to be.
> >Also, for that to actually work the branches have to be added to some
> >ptest config that ptest admins can do. I think only the main branches
> >(e.g. branch-1/2) are set up by default.
> >
> >On 17/8/15, 12:54, "Mithun RK"  wrote:
> >
> >>Hello, hive-dev.
> >>
> >>I was wondering if anyone could clarify how to get pre-commit tests to
> >>run
> >>for branches other than master.
> >>
> >>The dev documentation
> >> HowToContribute#HowToCo
> >>n
> >>tribute-CreatingaPatch>
> >>indicates
> >>that if the branch-name is specified as per the naming convention,
> >>pre-commit tests should run on the appropriate branch. It doesn't mention
> >>anything about setting target-versions, etc. I only seem to have had
> >>tests
> >>run on master.
> >>
> >>Does one have to cancel-patch + submit-patch for every branch? That
> >>sounds
> >>unwieldy.
> >>
> >>Regards, etc.
> >>Mithun
> >
>
>


Re: Running tests against branches other than master.

2017-08-15 Thread Sergey Shelukhin
And by jar I mean patch.

On 17/8/15, 12:56, "Sergey Shelukhin"  wrote:

>Specifying the branch name as part of the jar is the trigger. I’ve never
>tried it for more than one branch at a time so I’m not sure if HiveQA
>would run for every one; it seems more likely that it would pick the
>latest patch, whatever that happens to be.
>Also, for that to actually work the branches have to be added to some
>ptest config that ptest admins can do. I think only the main branches
>(e.g. branch-1/2) are set up by default.
>
>On 17/8/15, 12:54, "Mithun RK"  wrote:
>
>>Hello, hive-dev.
>>
>>I was wondering if anyone could clarify how to get pre-commit tests to
>>run
>>for branches other than master.
>>
>>The dev documentation
>>>n
>>tribute-CreatingaPatch>
>>indicates
>>that if the branch-name is specified as per the naming convention,
>>pre-commit tests should run on the appropriate branch. It doesn't mention
>>anything about setting target-versions, etc. I only seem to have had
>>tests
>>run on master.
>>
>>Does one have to cancel-patch + submit-patch for every branch? That
>>sounds
>>unwieldy.
>>
>>Regards, etc.
>>Mithun
>



[jira] [Created] (HIVE-17324) Typo in Hive language manual for varchar max size

2017-08-15 Thread Chip Sands (JIRA)
Chip Sands created HIVE-17324:
-

 Summary: Typo in Hive language manual for varchar max size
 Key: HIVE-17324
 URL: https://issues.apache.org/jira/browse/HIVE-17324
 Project: Hive
  Issue Type: Bug
Reporter: Chip Sands
Priority: Trivial


Varchar Max length

 Hive code defines may max length of char as MAX_VARCHAR_LENGTH = 65535


Manual swapped two digits and document max size as 65355


Varchar
Varchar types are created with a length specifier (between 1 and 65355),



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


Re: Running tests against branches other than master.

2017-08-15 Thread Sergey Shelukhin
Specifying the branch name as part of the jar is the trigger. I’ve never
tried it for more than one branch at a time so I’m not sure if HiveQA
would run for every one; it seems more likely that it would pick the
latest patch, whatever that happens to be.
Also, for that to actually work the branches have to be added to some
ptest config that ptest admins can do. I think only the main branches
(e.g. branch-1/2) are set up by default.

On 17/8/15, 12:54, "Mithun RK"  wrote:

>Hello, hive-dev.
>
>I was wondering if anyone could clarify how to get pre-commit tests to run
>for branches other than master.
>
>The dev documentation
>tribute-CreatingaPatch>
>indicates
>that if the branch-name is specified as per the naming convention,
>pre-commit tests should run on the appropriate branch. It doesn't mention
>anything about setting target-versions, etc. I only seem to have had tests
>run on master.
>
>Does one have to cancel-patch + submit-patch for every branch? That sounds
>unwieldy.
>
>Regards, etc.
>Mithun



Running tests against branches other than master.

2017-08-15 Thread Mithun RK
Hello, hive-dev.

I was wondering if anyone could clarify how to get pre-commit tests to run
for branches other than master.

The dev documentation

indicates
that if the branch-name is specified as per the naming convention,
pre-commit tests should run on the appropriate branch. It doesn't mention
anything about setting target-versions, etc. I only seem to have had tests
run on master.

Does one have to cancel-patch + submit-patch for every branch? That sounds
unwieldy.

Regards, etc.
Mithun


JSON SerDe column name mapping

2017-08-15 Thread Jagrut Sharma
For org.apache.hive.hcatalog.data.JsonSerDe, is there a way to provide
column name mapping via serdeproperties? Trying to handle keys with dots in
them by renaming to underscores.

For a JSON key like key.name.dev, the Hive table will have column
key_name_dev. The mapping should be specified via the serde config when
creating the table.

This functionality is provided by this SerDe (https://github.com/rcongiu/
Hive-JSON-Serde), and trying to get something similar.

Thanks.

-- 
Jagrut


[jira] [Created] (HIVE-17323) Improve upon HIVE-16260

2017-08-15 Thread Deepak Jaiswal (JIRA)
Deepak Jaiswal created HIVE-17323:
-

 Summary: Improve upon HIVE-16260
 Key: HIVE-17323
 URL: https://issues.apache.org/jira/browse/HIVE-17323
 Project: Hive
  Issue Type: Bug
Reporter: Deepak Jaiswal
Assignee: Deepak Jaiswal


HIVE-16260 allows removal of parallel edges of semijoin with mapjoins.

https://issues.apache.org/jira/browse/HIVE-16260

However, it should also consider dynamic partition pruning edge like semijoin 
without removing it while traversing the query tree.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[GitHub] hive pull request #232: Update 039-HIVE-12274.oracle.sql

2017-08-15 Thread rmartin-rp
GitHub user rmartin-rp opened a pull request:

https://github.com/apache/hive/pull/232

Update 039-HIVE-12274.oracle.sql

You cannot modify a column from VARCHAR2(4000) to CLOB directly. You need 
to add a new column and drop the old one or recreate the table. I chose to 
recreate the table because it's more clean.


https://asktom.oracle.com/pls/asktom/f?p=100:11:0P11_QUESTION_ID:1770086700346491686

Also, if you do something like:
ALTER TABLE COLUMNS_V2 MODIFY (COLUMN_NAME VARCHAR(767) NOT NULL);
And the column is already NOT NULL, the ALTER TABLE fails.  I removed 2 
like this.

I hope it helps.

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/rmartin-rp/hive patch-1

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/hive/pull/232.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #232


commit 4ff45b78f16171a808a2635705d22536db34dfb4
Author: rmartin-rp 
Date:   2017-08-15T15:39:24Z

Update 039-HIVE-12274.oracle.sql

You cannot modify a column from VARCHAR2(4000) to CLOB directly. You need 
to add a new column and drop the old one or recreate the table. I chose to 
recreate the table because it's more clean.


https://asktom.oracle.com/pls/asktom/f?p=100:11:0P11_QUESTION_ID:1770086700346491686

Also, if you do something like:
ALTER TABLE COLUMNS_V2 MODIFY (COLUMN_NAME VARCHAR(767) NOT NULL);
And the column is already NOT NULL, the ALTER TABLE fails.  I removed 2 
like this.

I hope it helps.




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


Review Request 61663: WebUI query plan graphs

2017-08-15 Thread Karen Coppage

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/61663/
---

Review request for hive, Peter Vary and Xuefu Zhang.


Repository: hive-git


Description
---

I’m working on a feature of the Hive WebUI Query Plan tab that would provide 
the option to display the query plan as a nice graph (scroll down for 
screenshots). If you click on one of the graph’s stages, the plan for that 
stage appears as text below.
Stages are color-coded if they have a status (Success, Error, Running), and the 
rest are grayed out. Coloring is based on status already available in the 
WebUI, under the Stages tab.
There is an additional option to display stats for MapReduce tasks. This 
includes the job’s ID, tracking URL (where the logs are found), and mapper and 
reducer numbers/progress, among other info.
The library I’m using for the graph is called vis.js (http://visjs.org/). It 
has an Apache license, and the only necessary file to be included from this 
library is about 700 KB.
I tried to keep server-side changes minimal, and graph generation is taken care 
of by the client. Plans with more than a given number of stages (default: 25) 
won't be displayed in order to preserve resources.


Diffs
-

  common/src/java/org/apache/hadoop/hive/common/LogUtils.java 0a3e0c7201 
  common/src/java/org/apache/hadoop/hive/conf/HiveConf.java 056f2d7834 
  
itests/hive-unit/src/test/java/org/apache/hive/service/cli/session/TestQueryDisplay.java
 155c65dd26 
  ql/src/java/org/apache/hadoop/hive/ql/Driver.java 4e7c80f184 
  ql/src/java/org/apache/hadoop/hive/ql/MapRedStats.java 4b6051485e 
  ql/src/java/org/apache/hadoop/hive/ql/QueryDisplay.java 5bf22107dc 
  ql/src/java/org/apache/hadoop/hive/ql/exec/mr/HadoopJobExecHelper.java 
3c0719717c 
  ql/src/java/org/apache/hadoop/hive/ql/exec/mr/MapRedTask.java 1bd4db7805 
  service/src/jamon/org/apache/hive/tmpl/QueryProfileTmpl.jamon 5e2d68c4a4 
  service/src/java/org/apache/hive/service/servlet/QueryProfileServlet.java 
27a3b1423a 
  service/src/resources/hive-webapps/static/css/query-plan-graph.css 
PRE-CREATION 
  service/src/resources/hive-webapps/static/js/query-plan-graph.js PRE-CREATION 
  service/src/resources/hive-webapps/static/js/vis.min.js PRE-CREATION 


Diff: https://reviews.apache.org/r/61663/diff/1/


Testing
---


Thanks,

Karen Coppage



Re: Review Request 61656: HIVE-17316: Use regular expressions for the hidden configuration variables

2017-08-15 Thread Barna Zsombor Klara

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/61656/
---

(Updated Aug. 15, 2017, 3:30 p.m.)


Review request for hive and Peter Vary.


Changes
---

Fixed failing tests and updated based on review comments. Thank you Peter for 
the review!


Repository: hive-git


Description
---

HIVE-17316: Use regular expressions for the hidden configuration variables


Diffs (updated)
-

  common/src/java/org/apache/hadoop/hive/conf/HiveConf.java 
056f2d78346b6b306d34dfb610e3a7fed4ca68aa 
  common/src/java/org/apache/hadoop/hive/conf/HiveConfUtil.java 
9084fed9fa5121285709874923b3e03e887cb20d 
  common/src/test/org/apache/hadoop/hive/conf/TestHiveConfRestrictList.java 
980fa761541d72a0b37bc5d687e1fbc4551b5282 
  common/src/test/org/apache/hadoop/hive/conf/TestHiveConfUtil.java 
PRE-CREATION 
  
service/src/test/org/apache/hive/service/cli/session/TestSessionGlobalInitFile.java
 5760e66786b1521e439f7abc3597776449b6336d 


Diff: https://reviews.apache.org/r/61656/diff/2/

Changes: https://reviews.apache.org/r/61656/diff/1-2/


Testing
---

Added unit tests and checked that the current set of restricted variables will 
not restrict any new ones due to the change in the check from equals to 
contains.


Thanks,

Barna Zsombor Klara



Re: Review Request 61656: HIVE-17316: Use regular expressions for the hidden configuration variables

2017-08-15 Thread Peter Vary

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/61656/#review182946
---




common/src/java/org/apache/hadoop/hive/conf/HiveConfUtil.java
Lines 21 (patched)


nit: Is this needed?



common/src/java/org/apache/hadoop/hive/conf/HiveConfUtil.java
Lines 42 (patched)


nit: Is this needed?



common/src/test/org/apache/hadoop/hive/conf/TestHiveConfRestrictList.java
Line 44 (original), 44 (patched)


Why does this needed?



common/src/test/org/apache/hadoop/hive/conf/TestHiveConfUntil.java
Lines 29 (patched)


nit: Until -> Util? :)



common/src/test/org/apache/hadoop/hive/conf/TestHiveConfUntil.java
Lines 44 (patched)


What happens if the user uses conf.get("dummyBoolean")


- Peter Vary


On Aug. 15, 2017, 1:33 p.m., Barna Zsombor Klara wrote:
> 
> ---
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/61656/
> ---
> 
> (Updated Aug. 15, 2017, 1:33 p.m.)
> 
> 
> Review request for hive and Peter Vary.
> 
> 
> Repository: hive-git
> 
> 
> Description
> ---
> 
> HIVE-17316: Use regular expressions for the hidden configuration variables
> 
> 
> Diffs
> -
> 
>   common/src/java/org/apache/hadoop/hive/conf/HiveConf.java 
> 056f2d78346b6b306d34dfb610e3a7fed4ca68aa 
>   common/src/java/org/apache/hadoop/hive/conf/HiveConfUtil.java 
> 9084fed9fa5121285709874923b3e03e887cb20d 
>   common/src/test/org/apache/hadoop/hive/conf/TestHiveConfRestrictList.java 
> 980fa761541d72a0b37bc5d687e1fbc4551b5282 
>   common/src/test/org/apache/hadoop/hive/conf/TestHiveConfUntil.java 
> PRE-CREATION 
> 
> 
> Diff: https://reviews.apache.org/r/61656/diff/1/
> 
> 
> Testing
> ---
> 
> Added unit tests and checked that the current set of restricted variables 
> will not restrict any new ones due to the change in the check from equals to 
> contains.
> 
> 
> Thanks,
> 
> Barna Zsombor Klara
> 
>



Review Request 61656: HIVE-17316: Use regular expressions for the hidden configuration variables

2017-08-15 Thread Barna Zsombor Klara

---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/61656/
---

Review request for hive and Peter Vary.


Repository: hive-git


Description
---

HIVE-17316: Use regular expressions for the hidden configuration variables


Diffs
-

  common/src/java/org/apache/hadoop/hive/conf/HiveConf.java 
056f2d78346b6b306d34dfb610e3a7fed4ca68aa 
  common/src/java/org/apache/hadoop/hive/conf/HiveConfUtil.java 
9084fed9fa5121285709874923b3e03e887cb20d 
  common/src/test/org/apache/hadoop/hive/conf/TestHiveConfRestrictList.java 
980fa761541d72a0b37bc5d687e1fbc4551b5282 
  common/src/test/org/apache/hadoop/hive/conf/TestHiveConfUntil.java 
PRE-CREATION 


Diff: https://reviews.apache.org/r/61656/diff/1/


Testing
---

Added unit tests and checked that the current set of restricted variables will 
not restrict any new ones due to the change in the check from equals to 
contains.


Thanks,

Barna Zsombor Klara



[jira] [Created] (HIVE-17322) Execute BeeLine qtests in a serial manner to prevent flakyness

2017-08-15 Thread Barna Zsombor Klara (JIRA)
Barna Zsombor Klara created HIVE-17322:
--

 Summary: Execute BeeLine qtests in a serial manner to prevent 
flakyness
 Key: HIVE-17322
 URL: https://issues.apache.org/jira/browse/HIVE-17322
 Project: Hive
  Issue Type: Bug
Affects Versions: 3.0.0
Reporter: Barna Zsombor Klara
Assignee: Barna Zsombor Klara
Priority: Minor






--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (HIVE-17321) HoS: analyze ORC table doesn't compute raw data size when noscan/partialscan is not specified

2017-08-15 Thread Rui Li (JIRA)
Rui Li created HIVE-17321:
-

 Summary: HoS: analyze ORC table doesn't compute raw data size when 
noscan/partialscan is not specified
 Key: HIVE-17321
 URL: https://issues.apache.org/jira/browse/HIVE-17321
 Project: Hive
  Issue Type: Bug
  Components: Spark
Reporter: Rui Li
Assignee: Rui Li






--
This message was sent by Atlassian JIRA
(v6.4.14#64029)