This is an automated email from the ASF dual-hosted git repository.

git-site-role pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/drill-site.git


The following commit(s) were added to refs/heads/asf-site by this push:
     new 1c66e4e  Automatic Site Publish by Buildbot
1c66e4e is described below

commit 1c66e4e4ee29d27f90cf5c8362030cb4e626d712
Author: buildbot <[email protected]>
AuthorDate: Thu Dec 2 12:02:12 2021 +0000

    Automatic Site Publish by Buildbot
---
 output/docs/mongo-metastore/index.html    |  4 +-
 output/docs/parquet-format/index.html     | 75 ++++++++++++++++++++++++-------
 output/feed.xml                           |  4 +-
 output/zh/docs/mongo-metastore/index.html |  4 +-
 output/zh/docs/parquet-format/index.html  | 75 ++++++++++++++++++++++++-------
 output/zh/feed.xml                        |  4 +-
 6 files changed, 130 insertions(+), 36 deletions(-)

diff --git a/output/docs/mongo-metastore/index.html 
b/output/docs/mongo-metastore/index.html
index 5c437be..604f620 100644
--- a/output/docs/mongo-metastore/index.html
+++ b/output/docs/mongo-metastore/index.html
@@ -1436,7 +1436,9 @@
 
     <div class="int_text" align="left">
       
-        <p>The Mongo Metastore implementation allows you store Drill Metastore 
metadata in a configured
+        <p><strong>Introduced in release:</strong> 1.20.</p>
+
+<p>The Mongo Metastore implementation allows you store Drill Metastore 
metadata in a configured
  MongoDB.</p>
 
 <h2 id="configuration">Configuration</h2>
diff --git a/output/docs/parquet-format/index.html 
b/output/docs/parquet-format/index.html
index aef5194..dd3d168 100644
--- a/output/docs/parquet-format/index.html
+++ b/output/docs/parquet-format/index.html
@@ -1453,6 +1453,51 @@
   <li>Handling Parquet data types</li>
 </ul>
 
+<h2 id="configuration-options">Configuration Options</h2>
+
+<table>
+  <thead>
+    <tr>
+      <th>Option</th>
+      <th>Description</th>
+    </tr>
+  </thead>
+  <tbody>
+    <tr>
+      <td>enableStringsSignedMinMax</td>
+      <td>See config opt store.parquet.reader.strings_signed_min_max</td>
+    </tr>
+    <tr>
+      <td>blockSize</td>
+      <td>See config opt store.parquet.block-size</td>
+    </tr>
+    <tr>
+      <td>pageSize</td>
+      <td>See config opt store.parquet.page-size</td>
+    </tr>
+    <tr>
+      <td>useSingleFsBlock</td>
+      <td>See config opt store.parquet.writer.use_single_fs_block</td>
+    </tr>
+    <tr>
+      <td>writerCompressionType</td>
+      <td>See config opt store.parquet.compression</td>
+    </tr>
+    <tr>
+      <td>writerLogicalTypeForDecimals</td>
+      <td>See config opt store.parquet.writer.logical_type_for_decimals</td>
+    </tr>
+    <tr>
+      <td>writerUsePrimitivesForDecimals</td>
+      <td>See config opt 
store.parquet.writer.use_primitive_types_for_decimals</td>
+    </tr>
+    <tr>
+      <td>writerFormatVersion</td>
+      <td>See config opt store.parquet.writer.format_version</td>
+    </tr>
+  </tbody>
+</table>
+
 <h2 id="reading-parquet-files">Reading Parquet Files</h2>
 <p>When a read of Parquet data occurs, Drill loads only the necessary columns 
of data, which reduces I/O. Reading only a small piece of the Parquet data from 
a data file or table, Drill can examine and analyze all values for a column 
across multiple files. You can create a Drill table from one format and store 
the data in another format, including Parquet.</p>
 
@@ -1472,13 +1517,13 @@
        "parquet": {
          "type": "parquet",
          "autoCorrectCorruptDates": false
-       }  
+       }
 </code></pre></div></div>
 
 <p>Alternatively, you can set the option to false when you issue a query, as 
shown in the following example:</p>
 
-<div class="language-plaintext highlighter-rouge"><div class="highlight"><pre 
class="highlight"><code>   SELECT l_shipdate, l_commitdate FROM 
table(dfs.`/drill/testdata/parquet_date/dates_nodrillversion/drillgen2_lineitem`
 
-   (type =&gt; 'parquet', autoCorrectCorruptDates =&gt; false)) LIMIT 1; 
+<div class="language-plaintext highlighter-rouge"><div class="highlight"><pre 
class="highlight"><code>   SELECT l_shipdate, l_commitdate FROM 
table(dfs.`/drill/testdata/parquet_date/dates_nodrillversion/drillgen2_lineitem`
+   (type =&gt; 'parquet', autoCorrectCorruptDates =&gt; false)) LIMIT 1;
 </code></pre></div></div>
 
 <h3 id="configuring-the-parquet-storage-format">Configuring the Parquet 
Storage Format</h3>
@@ -1504,7 +1549,7 @@
 <h3 id="configuring-the-hdfs-block-size-for-parquet-files">Configuring the 
HDFS Block Size for Parquet Files</h3>
 <p>Drill 1.11 introduces the <code class="language-plaintext 
highlighter-rouge">store.parquet.writer.use_single_fs_block</code> option, 
which enables Drill to write a Parquet file as a single file system block 
without changing the default file system block size. Query performance improves 
when Drill reads Parquet files as a single block on the file system. When the 
<code class="language-plaintext 
highlighter-rouge">store.parquet.writer.use_single_fs_block</code> option is 
enabled, the <c [...]
 
-<div class="language-plaintext highlighter-rouge"><div class="highlight"><pre 
class="highlight"><code>ALTER SYSTEM|SESSION SET 
store.parquet.writer.use_single_fs_block = 'true|false';  
+<div class="language-plaintext highlighter-rouge"><div class="highlight"><pre 
class="highlight"><code>ALTER SYSTEM|SESSION SET 
store.parquet.writer.use_single_fs_block = 'true|false';
 </code></pre></div></div>
 
 <h3 id="type-mapping">Type Mapping</h3>
@@ -1522,10 +1567,10 @@
 <h3 id="example-read-json-write-parquet">Example: Read JSON, Write Parquet</h3>
 <p>This example demonstrates a storage plugin definition, a sample row of data 
from a JSON file, and a Drill query that writes the JSON input to Parquet 
output.</p>
 
-<h3 id="storage-plugin-definition">Storage Plugin Definition</h3>
+<h4 id="storage-plugin-definition">Storage Plugin Definition</h4>
 <p>You can use the default dfs storage plugin installed with Drill for reading 
and writing Parquet files. The storage plugin needs to configure the writable 
option of the workspace to true, so Drill can write the Parquet output. The dfs 
storage plugin defines the tmp writable workspace, which you can use in the 
CTAS command to create a Parquet table.</p>
 
-<h3 id="sample-row-of-json-data">Sample Row of JSON Data</h3>
+<h4 id="sample-row-of-json-data">Sample Row of JSON Data</h4>
 <p>A JSON file called sample.json contains data consisting of strings, typical 
of JSON data. The following example shows one row of the JSON file:</p>
 
 <div class="language-plaintext highlighter-rouge"><div class="highlight"><pre 
class="highlight"><code>    
{"trans_id":0,"date":"2013-07-26","time":"04:56:59","amount":80.5,"user_info":
@@ -1540,16 +1585,16 @@
     }
 </code></pre></div></div>
 
-<h3 id="ctas-query">CTAS Query</h3>
+<h4 id="ctas-query">CTAS Query</h4>
 <p>The following example shows a CTAS query that creates a table from JSON 
data shown in the last example. The command casts the date, time, and amount 
strings to SQL types DATE, TIME, and DOUBLE. String-to-VARCHAR casting of the 
other strings occurs automatically.</p>
 
-<div class="language-plaintext highlighter-rouge"><div class="highlight"><pre 
class="highlight"><code>CREATE TABLE dfs.tmp.sampleparquet AS 
-(SELECT trans_id, 
-cast(`date` AS date) transdate, 
-cast(`time` AS time) transtime, 
-cast(amount AS double) amountm,
-user_info, marketing_info, trans_info 
-FROM dfs.`/Users/drilluser/sample.json`);
+<div class="language-sql highlighter-rouge"><div class="highlight"><pre 
class="highlight"><code><span class="k">CREATE</span> <span 
class="k">TABLE</span> <span class="n">dfs</span><span class="p">.</span><span 
class="n">tmp</span><span class="p">.</span><span 
class="n">sampleparquet</span> <span class="k">AS</span>
+<span class="p">(</span><span class="k">SELECT</span> <span 
class="n">trans_id</span><span class="p">,</span>
+<span class="k">cast</span><span class="p">(</span><span 
class="nv">`date`</span> <span class="k">AS</span> <span 
class="nb">date</span><span class="p">)</span> <span 
class="n">transdate</span><span class="p">,</span>
+<span class="k">cast</span><span class="p">(</span><span 
class="nv">`time`</span> <span class="k">AS</span> <span 
class="nb">time</span><span class="p">)</span> <span 
class="n">transtime</span><span class="p">,</span>
+<span class="k">cast</span><span class="p">(</span><span 
class="n">amount</span> <span class="k">AS</span> <span 
class="nb">double</span><span class="p">)</span> <span 
class="n">amountm</span><span class="p">,</span>
+<span class="n">user_info</span><span class="p">,</span> <span 
class="n">marketing_info</span><span class="p">,</span> <span 
class="n">trans_info</span>
+<span class="k">FROM</span> <span class="n">dfs</span><span 
class="p">.</span><span class="nv">`/Users/drilluser/sample.json`</span><span 
class="p">);</span>
 </code></pre></div></div>
 
 <p>The CTAS query does not specify a file name extension for the output. Drill 
creates a parquet file by default, as indicated by the file name in the 
output:</p>
@@ -1564,7 +1609,7 @@ FROM dfs.`/Users/drilluser/sample.json`);
 
 <p>You can query the Parquet file to verify that Drill now interprets the 
converted string as a date.</p>
 
-<div class="language-plaintext highlighter-rouge"><div class="highlight"><pre 
class="highlight"><code>SELECT extract(year from transdate) AS `Year`, 
t.user_info.cust_id AS Customer 
+<div class="language-plaintext highlighter-rouge"><div class="highlight"><pre 
class="highlight"><code>SELECT extract(year from transdate) AS `Year`, 
t.user_info.cust_id AS Customer
 FROM dfs.tmp.`sampleparquet` t;
 
 |------------|------------|
diff --git a/output/feed.xml b/output/feed.xml
index 3cec40f..23c189e 100644
--- a/output/feed.xml
+++ b/output/feed.xml
@@ -6,8 +6,8 @@
 </description>
     <link>/</link>
     <atom:link href="/feed.xml" rel="self" type="application/rss+xml"/>
-    <pubDate>Wed, 01 Dec 2021 07:58:25 +0000</pubDate>
-    <lastBuildDate>Wed, 01 Dec 2021 07:58:25 +0000</lastBuildDate>
+    <pubDate>Thu, 02 Dec 2021 11:59:26 +0000</pubDate>
+    <lastBuildDate>Thu, 02 Dec 2021 11:59:26 +0000</lastBuildDate>
     <generator>Jekyll v3.9.1</generator>
     
       <item>
diff --git a/output/zh/docs/mongo-metastore/index.html 
b/output/zh/docs/mongo-metastore/index.html
index 5a67a1c..197bd28 100644
--- a/output/zh/docs/mongo-metastore/index.html
+++ b/output/zh/docs/mongo-metastore/index.html
@@ -1436,7 +1436,9 @@
 
     <div class="int_text" align="left">
       
-        <p>The Mongo Metastore implementation allows you store Drill Metastore 
metadata in a configured
+        <p><strong>Introduced in release:</strong> 1.20.</p>
+
+<p>The Mongo Metastore implementation allows you store Drill Metastore 
metadata in a configured
  MongoDB.</p>
 
 <h2 id="configuration">Configuration</h2>
diff --git a/output/zh/docs/parquet-format/index.html 
b/output/zh/docs/parquet-format/index.html
index 4b9227c..b53e425 100644
--- a/output/zh/docs/parquet-format/index.html
+++ b/output/zh/docs/parquet-format/index.html
@@ -1453,6 +1453,51 @@
   <li>Handling Parquet data types</li>
 </ul>
 
+<h2 id="configuration-options">Configuration Options</h2>
+
+<table>
+  <thead>
+    <tr>
+      <th>Option</th>
+      <th>Description</th>
+    </tr>
+  </thead>
+  <tbody>
+    <tr>
+      <td>enableStringsSignedMinMax</td>
+      <td>See config opt store.parquet.reader.strings_signed_min_max</td>
+    </tr>
+    <tr>
+      <td>blockSize</td>
+      <td>See config opt store.parquet.block-size</td>
+    </tr>
+    <tr>
+      <td>pageSize</td>
+      <td>See config opt store.parquet.page-size</td>
+    </tr>
+    <tr>
+      <td>useSingleFsBlock</td>
+      <td>See config opt store.parquet.writer.use_single_fs_block</td>
+    </tr>
+    <tr>
+      <td>writerCompressionType</td>
+      <td>See config opt store.parquet.compression</td>
+    </tr>
+    <tr>
+      <td>writerLogicalTypeForDecimals</td>
+      <td>See config opt store.parquet.writer.logical_type_for_decimals</td>
+    </tr>
+    <tr>
+      <td>writerUsePrimitivesForDecimals</td>
+      <td>See config opt 
store.parquet.writer.use_primitive_types_for_decimals</td>
+    </tr>
+    <tr>
+      <td>writerFormatVersion</td>
+      <td>See config opt store.parquet.writer.format_version</td>
+    </tr>
+  </tbody>
+</table>
+
 <h2 id="reading-parquet-files">Reading Parquet Files</h2>
 <p>When a read of Parquet data occurs, Drill loads only the necessary columns 
of data, which reduces I/O. Reading only a small piece of the Parquet data from 
a data file or table, Drill can examine and analyze all values for a column 
across multiple files. You can create a Drill table from one format and store 
the data in another format, including Parquet.</p>
 
@@ -1472,13 +1517,13 @@
        "parquet": {
          "type": "parquet",
          "autoCorrectCorruptDates": false
-       }  
+       }
 </code></pre></div></div>
 
 <p>Alternatively, you can set the option to false when you issue a query, as 
shown in the following example:</p>
 
-<div class="language-plaintext highlighter-rouge"><div class="highlight"><pre 
class="highlight"><code>   SELECT l_shipdate, l_commitdate FROM 
table(dfs.`/drill/testdata/parquet_date/dates_nodrillversion/drillgen2_lineitem`
 
-   (type =&gt; 'parquet', autoCorrectCorruptDates =&gt; false)) LIMIT 1; 
+<div class="language-plaintext highlighter-rouge"><div class="highlight"><pre 
class="highlight"><code>   SELECT l_shipdate, l_commitdate FROM 
table(dfs.`/drill/testdata/parquet_date/dates_nodrillversion/drillgen2_lineitem`
+   (type =&gt; 'parquet', autoCorrectCorruptDates =&gt; false)) LIMIT 1;
 </code></pre></div></div>
 
 <h3 id="configuring-the-parquet-storage-format">Configuring the Parquet 
Storage Format</h3>
@@ -1504,7 +1549,7 @@
 <h3 id="configuring-the-hdfs-block-size-for-parquet-files">Configuring the 
HDFS Block Size for Parquet Files</h3>
 <p>Drill 1.11 introduces the <code class="language-plaintext 
highlighter-rouge">store.parquet.writer.use_single_fs_block</code> option, 
which enables Drill to write a Parquet file as a single file system block 
without changing the default file system block size. Query performance improves 
when Drill reads Parquet files as a single block on the file system. When the 
<code class="language-plaintext 
highlighter-rouge">store.parquet.writer.use_single_fs_block</code> option is 
enabled, the <c [...]
 
-<div class="language-plaintext highlighter-rouge"><div class="highlight"><pre 
class="highlight"><code>ALTER SYSTEM|SESSION SET 
store.parquet.writer.use_single_fs_block = 'true|false';  
+<div class="language-plaintext highlighter-rouge"><div class="highlight"><pre 
class="highlight"><code>ALTER SYSTEM|SESSION SET 
store.parquet.writer.use_single_fs_block = 'true|false';
 </code></pre></div></div>
 
 <h3 id="type-mapping">Type Mapping</h3>
@@ -1522,10 +1567,10 @@
 <h3 id="example-read-json-write-parquet">Example: Read JSON, Write Parquet</h3>
 <p>This example demonstrates a storage plugin definition, a sample row of data 
from a JSON file, and a Drill query that writes the JSON input to Parquet 
output.</p>
 
-<h3 id="storage-plugin-definition">Storage Plugin Definition</h3>
+<h4 id="storage-plugin-definition">Storage Plugin Definition</h4>
 <p>You can use the default dfs storage plugin installed with Drill for reading 
and writing Parquet files. The storage plugin needs to configure the writable 
option of the workspace to true, so Drill can write the Parquet output. The dfs 
storage plugin defines the tmp writable workspace, which you can use in the 
CTAS command to create a Parquet table.</p>
 
-<h3 id="sample-row-of-json-data">Sample Row of JSON Data</h3>
+<h4 id="sample-row-of-json-data">Sample Row of JSON Data</h4>
 <p>A JSON file called sample.json contains data consisting of strings, typical 
of JSON data. The following example shows one row of the JSON file:</p>
 
 <div class="language-plaintext highlighter-rouge"><div class="highlight"><pre 
class="highlight"><code>    
{"trans_id":0,"date":"2013-07-26","time":"04:56:59","amount":80.5,"user_info":
@@ -1540,16 +1585,16 @@
     }
 </code></pre></div></div>
 
-<h3 id="ctas-query">CTAS Query</h3>
+<h4 id="ctas-query">CTAS Query</h4>
 <p>The following example shows a CTAS query that creates a table from JSON 
data shown in the last example. The command casts the date, time, and amount 
strings to SQL types DATE, TIME, and DOUBLE. String-to-VARCHAR casting of the 
other strings occurs automatically.</p>
 
-<div class="language-plaintext highlighter-rouge"><div class="highlight"><pre 
class="highlight"><code>CREATE TABLE dfs.tmp.sampleparquet AS 
-(SELECT trans_id, 
-cast(`date` AS date) transdate, 
-cast(`time` AS time) transtime, 
-cast(amount AS double) amountm,
-user_info, marketing_info, trans_info 
-FROM dfs.`/Users/drilluser/sample.json`);
+<div class="language-sql highlighter-rouge"><div class="highlight"><pre 
class="highlight"><code><span class="k">CREATE</span> <span 
class="k">TABLE</span> <span class="n">dfs</span><span class="p">.</span><span 
class="n">tmp</span><span class="p">.</span><span 
class="n">sampleparquet</span> <span class="k">AS</span>
+<span class="p">(</span><span class="k">SELECT</span> <span 
class="n">trans_id</span><span class="p">,</span>
+<span class="k">cast</span><span class="p">(</span><span 
class="nv">`date`</span> <span class="k">AS</span> <span 
class="nb">date</span><span class="p">)</span> <span 
class="n">transdate</span><span class="p">,</span>
+<span class="k">cast</span><span class="p">(</span><span 
class="nv">`time`</span> <span class="k">AS</span> <span 
class="nb">time</span><span class="p">)</span> <span 
class="n">transtime</span><span class="p">,</span>
+<span class="k">cast</span><span class="p">(</span><span 
class="n">amount</span> <span class="k">AS</span> <span 
class="nb">double</span><span class="p">)</span> <span 
class="n">amountm</span><span class="p">,</span>
+<span class="n">user_info</span><span class="p">,</span> <span 
class="n">marketing_info</span><span class="p">,</span> <span 
class="n">trans_info</span>
+<span class="k">FROM</span> <span class="n">dfs</span><span 
class="p">.</span><span class="nv">`/Users/drilluser/sample.json`</span><span 
class="p">);</span>
 </code></pre></div></div>
 
 <p>The CTAS query does not specify a file name extension for the output. Drill 
creates a parquet file by default, as indicated by the file name in the 
output:</p>
@@ -1564,7 +1609,7 @@ FROM dfs.`/Users/drilluser/sample.json`);
 
 <p>You can query the Parquet file to verify that Drill now interprets the 
converted string as a date.</p>
 
-<div class="language-plaintext highlighter-rouge"><div class="highlight"><pre 
class="highlight"><code>SELECT extract(year from transdate) AS `Year`, 
t.user_info.cust_id AS Customer 
+<div class="language-plaintext highlighter-rouge"><div class="highlight"><pre 
class="highlight"><code>SELECT extract(year from transdate) AS `Year`, 
t.user_info.cust_id AS Customer
 FROM dfs.tmp.`sampleparquet` t;
 
 |------------|------------|
diff --git a/output/zh/feed.xml b/output/zh/feed.xml
index 661f624..1d59166 100644
--- a/output/zh/feed.xml
+++ b/output/zh/feed.xml
@@ -6,8 +6,8 @@
 </description>
     <link>/</link>
     <atom:link href="/zh/feed.xml" rel="self" type="application/rss+xml"/>
-    <pubDate>Wed, 01 Dec 2021 07:58:25 +0000</pubDate>
-    <lastBuildDate>Wed, 01 Dec 2021 07:58:25 +0000</lastBuildDate>
+    <pubDate>Thu, 02 Dec 2021 11:59:26 +0000</pubDate>
+    <lastBuildDate>Thu, 02 Dec 2021 11:59:26 +0000</lastBuildDate>
     <generator>Jekyll v3.9.1</generator>
     
       <item>

Reply via email to