Github user HyukjinKwon commented on a diff in the pull request:
https://github.com/apache/spark/pull/23014#discussion_r232893546
--- Diff:
sql/core/src/main/java/org/apache/spark/sql/execution/vectorized/WritableColumnVector.java
---
@@ -101,10 +101,11 @@ private void throwUnsupportedException(int
requiredCapacity, Throwable cause) {
String message = "Cannot reserve additional contiguous bytes in the
vectorized reader (" +
(requiredCapacity >= 0 ? "requested " + requiredCapacity + "
bytes" : "integer overflow") +
"). As a workaround, you can reduce the vectorized reader batch
size, or disable the " +
- "vectorized reader. For parquet file format, refer to " +
+ "vectorized reader, or disable " +
SQLConf.BUCKETING_ENABLED().key() + " if you read " +
+ "from bucket table. For Parquet file format, refer to " +
SQLConf.PARQUET_VECTORIZED_READER_BATCH_SIZE().key() +
" (default " +
SQLConf.PARQUET_VECTORIZED_READER_BATCH_SIZE().defaultValueString() +
- ") and " + SQLConf.PARQUET_VECTORIZED_READER_ENABLED().key() + ";
for orc file format, " +
+ ") and " + SQLConf.PARQUET_VECTORIZED_READER_ENABLED().key() + ";
for Orc file format, " +
--- End diff --
`Orc` is `ORC` BTW :-).
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]