This is an automated email from the ASF dual-hosted git repository.

vinoth pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/incubator-hudi.git


The following commit(s) were added to refs/heads/asf-site by this push:
     new d0c3b9f  Travis CI build asf-site
d0c3b9f is described below

commit d0c3b9fb1095eaae1cda48bbef0de32defdb04b8
Author: CI <[email protected]>
AuthorDate: Thu May 21 04:46:35 2020 +0000

    Travis CI build asf-site
---
 content/cn/docs/0.5.2-querying_data.html | 4 ++--
 content/cn/docs/querying_data.html       | 4 ++--
 content/cn/docs/quick-start-guide.html   | 2 ++
 content/docs/0.5.2-querying_data.html    | 4 ++--
 content/docs/querying_data.html          | 4 ++--
 content/docs/quick-start-guide.html      | 6 ++++--
 6 files changed, 14 insertions(+), 10 deletions(-)

diff --git a/content/cn/docs/0.5.2-querying_data.html 
b/content/cn/docs/0.5.2-querying_data.html
index eeaf7a1..4ed98db 100644
--- a/content/cn/docs/0.5.2-querying_data.html
+++ b/content/cn/docs/0.5.2-querying_data.html
@@ -360,7 +360,7 @@
     </ul>
   </li>
   <li><a href="#presto">Presto</a></li>
-  <li><a href="#impala此功能还未正式发布">Impala(此功能还未正式发布)</a>
+  <li><a href="#impala-34-or-later">Impala (3.4 or later)</a>
     <ul>
       <li><a href="#读优化表-1">读优化表</a></li>
     </ul>
@@ -677,7 +677,7 @@ Upsert实用程序(<code 
class="highlighter-rouge">HoodieDeltaStreamer</code>
 <p>Presto是一种常用的查询引擎,可提供交互式查询性能。 Hudi RO表可以在Presto中无缝查询。
 这需要在整个安装过程中将<code class="highlighter-rouge">hudi-presto-bundle</code> 
jar放入<code 
class="highlighter-rouge">&lt;presto_install&gt;/plugin/hive-hadoop2/</code>中。</p>
 
-<h2 id="impala此功能还未正式发布">Impala(此功能还未正式发布)</h2>
+<h2 id="impala-34-or-later">Impala (3.4 or later)</h2>
 
 <h3 id="读优化表-1">读优化表</h3>
 
diff --git a/content/cn/docs/querying_data.html 
b/content/cn/docs/querying_data.html
index e33d18b..002c34c 100644
--- a/content/cn/docs/querying_data.html
+++ b/content/cn/docs/querying_data.html
@@ -360,7 +360,7 @@
     </ul>
   </li>
   <li><a href="#presto">Presto</a></li>
-  <li><a href="#impala此功能还未正式发布">Impala(此功能还未正式发布)</a>
+  <li><a href="#impala-34-or-later">Impala (3.4 or later)</a>
     <ul>
       <li><a href="#读优化表-1">读优化表</a></li>
     </ul>
@@ -677,7 +677,7 @@ Upsert实用程序(<code 
class="highlighter-rouge">HoodieDeltaStreamer</code>
 <p>Presto是一种常用的查询引擎,可提供交互式查询性能。 Hudi RO表可以在Presto中无缝查询。
 这需要在整个安装过程中将<code class="highlighter-rouge">hudi-presto-bundle</code> 
jar放入<code 
class="highlighter-rouge">&lt;presto_install&gt;/plugin/hive-hadoop2/</code>中。</p>
 
-<h2 id="impala此功能还未正式发布">Impala(此功能还未正式发布)</h2>
+<h2 id="impala-34-or-later">Impala (3.4 or later)</h2>
 
 <h3 id="读优化表-1">读优化表</h3>
 
diff --git a/content/cn/docs/quick-start-guide.html 
b/content/cn/docs/quick-start-guide.html
index 984639a..adc2bfc 100644
--- a/content/cn/docs/quick-start-guide.html
+++ b/content/cn/docs/quick-start-guide.html
@@ -410,6 +410,8 @@
     <span class="n">read</span><span class="o">.</span>
     <span class="nf">format</span><span class="o">(</span><span 
class="s">"org.apache.hudi"</span><span class="o">).</span>
     <span class="nf">load</span><span class="o">(</span><span 
class="n">basePath</span> <span class="o">+</span> <span 
class="s">"/*/*/*/*"</span><span class="o">)</span>
+    <span class="c1">//load(basePath) 如果使用 "/partitionKey=partitionValue" 
文件夹命名格式,Spark将自动识别分区信息
+</span>
 <span class="nv">roViewDF</span><span class="o">.</span><span 
class="py">registerTempTable</span><span class="o">(</span><span 
class="s">"hudi_ro_table"</span><span class="o">)</span>
 <span class="nv">spark</span><span class="o">.</span><span 
class="py">sql</span><span class="o">(</span><span class="s">"select fare, 
begin_lon, begin_lat, ts from  hudi_ro_table where fare &gt; 20.0"</span><span 
class="o">).</span><span class="py">show</span><span class="o">()</span>
 <span class="nv">spark</span><span class="o">.</span><span 
class="py">sql</span><span class="o">(</span><span class="s">"select 
_hoodie_commit_time, _hoodie_record_key, _hoodie_partition_path, rider, driver, 
fare from  hudi_ro_table"</span><span class="o">).</span><span 
class="py">show</span><span class="o">()</span>
diff --git a/content/docs/0.5.2-querying_data.html 
b/content/docs/0.5.2-querying_data.html
index 859bd0c..4315d03 100644
--- a/content/docs/0.5.2-querying_data.html
+++ b/content/docs/0.5.2-querying_data.html
@@ -357,7 +357,7 @@
     </ul>
   </li>
   <li><a href="#presto">Presto</a></li>
-  <li><a href="#impala-not-officially-released">Impala (Not Officially 
Released)</a>
+  <li><a href="#impala-34-or-later">Impala (3.4 or later)</a>
     <ul>
       <li><a href="#snapshot-query">Snapshot Query</a></li>
     </ul>
@@ -672,7 +672,7 @@ Please refer to <a 
href="/docs/0.5.2-configurations.html#spark-datasource">confi
 <p>Presto is a popular query engine, providing interactive query performance. 
Presto currently supports snapshot queries on COPY_ON_WRITE and read optimized 
queries 
 on MERGE_ON_READ Hudi tables. This requires the <code 
class="highlighter-rouge">hudi-presto-bundle</code> jar to be placed into <code 
class="highlighter-rouge">&lt;presto_install&gt;/plugin/hive-hadoop2/</code>, 
across the installation.</p>
 
-<h2 id="impala-not-officially-released">Impala (Not Officially Released)</h2>
+<h2 id="impala-34-or-later">Impala (3.4 or later)</h2>
 
 <h3 id="snapshot-query">Snapshot Query</h3>
 
diff --git a/content/docs/querying_data.html b/content/docs/querying_data.html
index e8dbe1a..2cfa722 100644
--- a/content/docs/querying_data.html
+++ b/content/docs/querying_data.html
@@ -357,7 +357,7 @@
     </ul>
   </li>
   <li><a href="#presto">Presto</a></li>
-  <li><a href="#impala-not-officially-released">Impala (Not Officially 
Released)</a>
+  <li><a href="#impala-34-or-later">Impala (3.4 or later)</a>
     <ul>
       <li><a href="#snapshot-query">Snapshot Query</a></li>
     </ul>
@@ -672,7 +672,7 @@ Please refer to <a 
href="/docs/configurations.html#spark-datasource">configurati
 <p>Presto is a popular query engine, providing interactive query performance. 
Presto currently supports snapshot queries on COPY_ON_WRITE and read optimized 
queries 
 on MERGE_ON_READ Hudi tables. This requires the <code 
class="highlighter-rouge">hudi-presto-bundle</code> jar to be placed into <code 
class="highlighter-rouge">&lt;presto_install&gt;/plugin/hive-hadoop2/</code>, 
across the installation.</p>
 
-<h2 id="impala-not-officially-released">Impala (Not Officially Released)</h2>
+<h2 id="impala-34-or-later">Impala (3.4 or later)</h2>
 
 <h3 id="snapshot-query">Snapshot Query</h3>
 
diff --git a/content/docs/quick-start-guide.html 
b/content/docs/quick-start-guide.html
index fa00061..76f0967 100644
--- a/content/docs/quick-start-guide.html
+++ b/content/docs/quick-start-guide.html
@@ -446,7 +446,8 @@ Here we are using the default write operation : <code 
class="highlighter-rouge">
   <span class="n">read</span><span class="o">.</span>
   <span class="nf">format</span><span class="o">(</span><span 
class="s">"hudi"</span><span class="o">).</span>
   <span class="nf">load</span><span class="o">(</span><span 
class="n">basePath</span> <span class="o">+</span> <span 
class="s">"/*/*/*/*"</span><span class="o">)</span>
-<span class="nv">tripsSnapshotDF</span><span class="o">.</span><span 
class="py">createOrReplaceTempView</span><span class="o">(</span><span 
class="s">"hudi_trips_snapshot"</span><span class="o">)</span>
+<span class="c1">//load(basePath) use "/partitionKey=partitionValue" folder 
structure for Spark auto partition discovery
+</span><span class="nv">tripsSnapshotDF</span><span class="o">.</span><span 
class="py">createOrReplaceTempView</span><span class="o">(</span><span 
class="s">"hudi_trips_snapshot"</span><span class="o">)</span>
 
 <span class="nv">spark</span><span class="o">.</span><span 
class="py">sql</span><span class="o">(</span><span class="s">"select fare, 
begin_lon, begin_lat, ts from  hudi_trips_snapshot where fare &gt; 
20.0"</span><span class="o">).</span><span class="py">show</span><span 
class="o">()</span>
 <span class="nv">spark</span><span class="o">.</span><span 
class="py">sql</span><span class="o">(</span><span class="s">"select 
_hoodie_commit_time, _hoodie_record_key, _hoodie_partition_path, rider, driver, 
fare from  hudi_trips_snapshot"</span><span class="o">).</span><span 
class="py">show</span><span class="o">()</span>
@@ -637,7 +638,8 @@ Here we are using the default write operation : <code 
class="highlighter-rouge">
   <span class="n">read</span><span class="o">.</span> \
   <span class="nb">format</span><span class="p">(</span><span 
class="s">"hudi"</span><span class="p">)</span><span class="o">.</span> \
   <span class="n">load</span><span class="p">(</span><span 
class="n">basePath</span> <span class="o">+</span> <span 
class="s">"/*/*/*/*"</span><span class="p">)</span>
-
+<span class="c1"># load(basePath) use "/partitionKey=partitionValue" folder 
structure for Spark auto partition discovery
+</span>
 <span class="n">tripsSnapshotDF</span><span class="o">.</span><span 
class="n">createOrReplaceTempView</span><span class="p">(</span><span 
class="s">"hudi_trips_snapshot"</span><span class="p">)</span>
 
 <span class="n">spark</span><span class="o">.</span><span 
class="n">sql</span><span class="p">(</span><span class="s">"select fare, 
begin_lon, begin_lat, ts from  hudi_trips_snapshot where fare &gt; 
20.0"</span><span class="p">)</span><span class="o">.</span><span 
class="n">show</span><span class="p">()</span>

Reply via email to