[
https://issues.apache.org/jira/browse/HIVE-26317?focusedWorklogId=781185&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-781185
]
ASF GitHub Bot logged work on HIVE-26317:
-----------------------------------------
Author: ASF GitHub Bot
Created on: 14/Jun/22 13:35
Start Date: 14/Jun/22 13:35
Worklog Time Spent: 10m
Work Description: github-actions[bot] commented on PR #3366:
URL: https://github.com/apache/hive/pull/3366#issuecomment-1155196943
# @check-spelling-bot Report
### :red_circle: Please review
See the [files](3366/files/) view or the [action
log](https://github.com/apache/hive/actions/runs/2495572724) for details.
#### Unrecognized words (3)
api
esri
wkid
<details><summary>Previously acknowledged words that are now absent
</summary>aarry timestamplocal yyyy </details>
<details><summary>Available dictionaries could cover words not in the
dictionary</summary>
[cspell:cpp/cpp.txt](https://raw.githubusercontent.com/check-spelling/cspell-dicts/HEAD/dictionaries/cpp/cpp.txt)
(104293) covers 81 of them
[cspell:django/django.txt](https://raw.githubusercontent.com/check-spelling/cspell-dicts/HEAD/dictionaries/django/django.txt)
(2342) covers 14 of them
[cspell:golang/go.txt](https://raw.githubusercontent.com/check-spelling/cspell-dicts/HEAD/dictionaries/golang/go.txt)
(7745) covers 12 of them
[cspell:java/java.txt](https://raw.githubusercontent.com/check-spelling/cspell-dicts/HEAD/dictionaries/java/java.txt)
(33524) covers 11 of them
[cspell:filetypes/filetypes.txt](https://raw.githubusercontent.com/check-spelling/cspell-dicts/HEAD/dictionaries/filetypes/filetypes.txt)
(337) covers 10 of them
[cspell:aws/aws.txt](https://raw.githubusercontent.com/check-spelling/cspell-dicts/HEAD/dictionaries/aws/aws.txt)
(1485) covers 10 of them
[cspell:css/css.txt](https://raw.githubusercontent.com/check-spelling/cspell-dicts/HEAD/dictionaries/css/css.txt)
(993) covers 9 of them
[cspell:rust/rust.txt](https://raw.githubusercontent.com/check-spelling/cspell-dicts/HEAD/dictionaries/rust/rust.txt)
(112) covers 8 of them
[cspell:npm/npm.txt](https://raw.githubusercontent.com/check-spelling/cspell-dicts/HEAD/dictionaries/npm/npm.txt)
(671) covers 8 of them
[cspell:html/html.txt](https://raw.githubusercontent.com/check-spelling/cspell-dicts/HEAD/dictionaries/html/html.txt)
(542) covers 8 of them
[cspell:scala/scala.txt](https://raw.githubusercontent.com/check-spelling/cspell-dicts/HEAD/dictionaries/scala/scala.txt)
(2752) covers 7 of them
[cspell:php/php.txt](https://raw.githubusercontent.com/check-spelling/cspell-dicts/HEAD/dictionaries/php/php.txt)
(9785) covers 6 of them
[cspell:fullstack/fullstack.txt](https://raw.githubusercontent.com/check-spelling/cspell-dicts/HEAD/dictionaries/fullstack/fullstack.txt)
(181) covers 5 of them
[cspell:csharp/csharp.txt](https://raw.githubusercontent.com/check-spelling/cspell-dicts/HEAD/dictionaries/csharp/csharp.txt)
(123) covers 5 of them
[cspell:python/python.txt](https://raw.githubusercontent.com/check-spelling/cspell-dicts/HEAD/dictionaries/python/python.txt)
(364) covers 3 of them
[cspell:lua/lua.txt](https://raw.githubusercontent.com/check-spelling/cspell-dicts/HEAD/dictionaries/lua/lua.txt)
(391) covers 3 of them
[cspell:dotnet/dotnet.txt](https://raw.githubusercontent.com/check-spelling/cspell-dicts/HEAD/dictionaries/dotnet/dotnet.txt)
(9824) covers 2 of them
[cspell:ruby/ruby.txt](https://raw.githubusercontent.com/check-spelling/cspell-dicts/HEAD/dictionaries/ruby/ruby.txt)
(354) covers 1 of them
[cspell:bash/bash-words.txt](https://raw.githubusercontent.com/check-spelling/cspell-dicts/HEAD/dictionaries/bash/bash-words.txt)
(22) covers 1 of them
Consider adding them using:
```
with:
extra_dictionaries:
cspell:cpp/cpp.txt
cspell:django/django.txt
cspell:golang/go.txt
cspell:java/java.txt
cspell:filetypes/filetypes.txt
cspell:aws/aws.txt
cspell:css/css.txt
cspell:rust/rust.txt
cspell:npm/npm.txt
cspell:html/html.txt
cspell:scala/scala.txt
cspell:php/php.txt
cspell:fullstack/fullstack.txt
cspell:csharp/csharp.txt
cspell:python/python.txt
cspell:lua/lua.txt
cspell:dotnet/dotnet.txt
cspell:ruby/ruby.txt
cspell:bash/bash-words.txt
```
To stop checking additional dictionaries, add:
```
with:
check_extra_dictionaries: ''
```
</details>
<details><summary>To accept these unrecognized words as correct (and remove
the previously acknowledged and now absent words),
run the following commands</summary>
... in a clone of the
[[email protected]:lcspinter/hive-1.git](https://github.com/lcspinter/hive-1.git)
repository
on the `HIVE-26317` branch:
```
update_files() {
perl -e '
my @expect_files=qw('".github/actions/spelling/expect.txt"');
@ARGV=@expect_files;
my @stale=qw('"$patch_remove"');
my $re=join "|", @stale;
my $suffix=".".time();
my $previous="";
sub maybe_unlink { unlink($_[0]) if $_[0]; }
while (<>) {
if ($ARGV ne $old_argv) { maybe_unlink($previous); $previous="$ARGV$suffix";
rename($ARGV, $previous); open(ARGV_OUT, ">$ARGV"); select(ARGV_OUT); $old_argv
= $ARGV; }
next if /^(?:$re)(?:(?:\r|\n)*$| .*)/; print;
}; maybe_unlink($previous);'
perl -e '
my $new_expect_file=".github/actions/spelling/expect.txt";
use File::Path qw(make_path);
use File::Basename qw(dirname);
make_path (dirname($new_expect_file));
open FILE, q{<}, $new_expect_file; chomp(my @words = <FILE>); close FILE;
my @add=qw('"$patch_add"');
my %items; @items{@words} = @words x (1); @items{@add} = @add x (1);
@words = sort {lc($a)."-".$a cmp lc($b)."-".$b} keys %items;
open FILE, q{>}, $new_expect_file; for my $word (@words) { print FILE
"$word\n" if $word =~ /\w/; };
close FILE;
system("git", "add", $new_expect_file);
'
}
comment_json=$(mktemp)
curl -L -s -S \
-H "Content-Type: application/json" \
"COMMENT_URL" > "$comment_json"
comment_body=$(mktemp)
jq -r ".body // empty" "$comment_json" > $comment_body
rm $comment_json
patch_remove=$(perl -ne 'next unless s{^</summary>(.*)</details>$}{$1};
print' < "$comment_body")
patch_add=$(perl -e '$/=undef; $_=<>; if (m{Unrecognized
words[^<]*</summary>\n*```\n*([^<]*)```\n*</details>$}m) { print "$1" } elsif
(m{Unrecognized words[^<]*\n\n((?:\w.*\n)+)\n}m) { print "$1" };' <
"$comment_body")
update_files
rm $comment_body
git add -u
```
</details>
<!
Issue Time Tracking
-------------------
Worklog Id: (was: 781185)
Time Spent: 20m (was: 10m)
> Select on iceberg table stored as parquet and vectorization enabled fails
> with Runtime Exception
> ------------------------------------------------------------------------------------------------
>
> Key: HIVE-26317
> URL: https://issues.apache.org/jira/browse/HIVE-26317
> Project: Hive
> Issue Type: Bug
> Reporter: László Pintér
> Assignee: László Pintér
> Priority: Major
> Labels: pull-request-available
> Time Spent: 20m
> Remaining Estimate: 0h
>
> Create an iceberg table having the following schema:
> {code:sql}
> CREATE EXTERNAL tbl_complex TABLE (a int, arrayofarrays
> array<array<string>>) STORED BY ICEBERG STORED AS PARQUET
> {code}
> Insert some data into it:
> {code:sql}
> INSERT INTO tbl_complex VALUES (1, array(array('a'), array('b', 'c')))
> {code}
> Turn on vectorization and run a simple query:
> {code:sql}
> set hive.vectorized.execution.enabled=true;
> SELECT * FROM tbl_complex;
> {code}
> The query will fail with
> {code:java}
> Vertex did not succeed due to OWN_TASK_FAILURE, failedTasks:1 killedTasks:0,
> Vertex vertex_1655109552551_0001_2_00 [Map 1] killed/failed due
> to:OWN_TASK_FAILURE]DAG did not succeed due to VERTEX_FAILURE.
> failedVertices:1 killedVertices:0
> at
> org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:367)
> at
> org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:246)
> at
> org.apache.hive.service.cli.operation.SQLOperation.runInternal(SQLOperation.java:270)
> at
> org.apache.hive.service.cli.operation.Operation.run(Operation.java:281)
> at
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:545)
> at
> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatement(HiveSessionImpl.java:513)
> at
> org.apache.hive.service.cli.CLIService.executeStatement(CLIService.java:271)
> at
> org.apache.iceberg.mr.hive.TestHiveShell.executeStatement(TestHiveShell.java:142)
> ... 16 more
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Vertex failed,
> vertexName=Map 1, vertexId=vertex_1655109552551_0001_2_00, diagnostics=[Task
> failed, taskId=task_1655109552551_0001_2_00_000000, diagnostics=[TaskAttempt
> 0 failed, info=[Error: Error while running task ( failure ) :
> attempt_1655109552551_0001_2_00_000000_0:java.lang.RuntimeException:
> org.apache.hadoop.hive.ql.metadata.HiveException: java.io.IOException:
> java.lang.RuntimeException: Unsupported type used in list:array<string>
> at
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:348)
> at
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:276)
> at
> org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:381)
> at
> org.apache.tez.runtime.task.TaskRunner2Callable$1.run(TaskRunner2Callable.java:82)
> at
> org.apache.tez.runtime.task.TaskRunner2Callable$1.run(TaskRunner2Callable.java:69)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:422)
> at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1682)
> at
> org.apache.tez.runtime.task.TaskRunner2Callable.callInternal(TaskRunner2Callable.java:69)
> at
> org.apache.tez.runtime.task.TaskRunner2Callable.callInternal(TaskRunner2Callable.java:39)
> at org.apache.tez.common.CallableWithNdc.call(CallableWithNdc.java:36)
> at
> com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:108)
> at
> com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:41)
> at
> com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:77)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> at java.lang.Thread.run(Thread.java:748)
> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException:
> java.io.IOException: java.lang.RuntimeException: Unsupported type used in
> list:array<string>
> at
> org.apache.hadoop.hive.ql.exec.tez.MapRecordSource.pushRecord(MapRecordSource.java:89)
> at
> org.apache.hadoop.hive.ql.exec.tez.MapRecordProcessor.run(MapRecordProcessor.java:414)
> at
> org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:293)
> ... 16 more
> Caused by: java.io.IOException: java.lang.RuntimeException: Unsupported type
> used in list:array<string>
> at
> org.apache.hadoop.hive.io.HiveIOExceptionHandlerChain.handleRecordReaderNextException(HiveIOExceptionHandlerChain.java:121)
> at
> org.apache.hadoop.hive.io.HiveIOExceptionHandlerUtil.handleRecordReaderNextException(HiveIOExceptionHandlerUtil.java:77)
> at
> org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.doNext(HiveContextAwareRecordReader.java:381)
> at
> org.apache.hadoop.hive.ql.io.HiveRecordReader.doNext(HiveRecordReader.java:82)
> at
> org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.next(HiveContextAwareRecordReader.java:119)
> at
> org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.next(HiveContextAwareRecordReader.java:59)
> at
> org.apache.hadoop.mapred.split.TezGroupedSplitsInputFormat$TezGroupedSplitsRecordReader.next(TezGroupedSplitsInputFormat.java:145)
> at
> org.apache.tez.mapreduce.lib.MRReaderMapred.next(MRReaderMapred.java:116)
> at
> org.apache.hadoop.hive.ql.exec.tez.MapRecordSource.pushRecord(MapRecordSource.java:75)
> ... 18 more
> Caused by: java.lang.RuntimeException: Unsupported type used in
> list:array<string>
> at
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.checkListColumnSupport(VectorizedParquetRecordReader.java:632)
> at
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.buildVectorizedParquetReader(VectorizedParquetRecordReader.java:568)
> at
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.checkEndOfRowGroup(VectorizedParquetRecordReader.java:470)
> at
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.nextBatch(VectorizedParquetRecordReader.java:431)
> at
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:379)
> at
> org.apache.hadoop.hive.ql.io.parquet.vector.VectorizedParquetRecordReader.next(VectorizedParquetRecordReader.java:100)
> at
> org.apache.iceberg.mr.hive.vector.VectorizedRowBatchIterator.advance(VectorizedRowBatchIterator.java:62)
> at
> org.apache.iceberg.mr.hive.vector.VectorizedRowBatchIterator.hasNext(VectorizedRowBatchIterator.java:84)
> at
> org.apache.iceberg.mr.mapreduce.IcebergInputFormat$IcebergRecordReader.nextKeyValue(IcebergInputFormat.java:273)
> at
> org.apache.iceberg.mr.hive.vector.HiveIcebergVectorizedRecordReader.next(HiveIcebergVectorizedRecordReader.java:48)
> at
> org.apache.iceberg.mr.hive.vector.HiveIcebergVectorizedRecordReader.next(HiveIcebergVectorizedRecordReader.java:34)
> at
> org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.doNext(HiveContextAwareRecordReader.java:376)
> ... 24 more
> {code}
--
This message was sent by Atlassian Jira
(v8.20.7#820007)