This is an automated email from the ASF dual-hosted git repository.

leesf pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/incubator-hudi.git


The following commit(s) were added to refs/heads/asf-site by this push:
     new 6fc935a  [DOCS] Update site (#1295)
6fc935a is described below

commit 6fc935a949d2371edd6fd613649e63297ec3aa6d
Author: Bhavani Sudha Saktheeswaran <bhasu...@uber.com>
AuthorDate: Fri Jan 31 00:45:24 2020 -0800

    [DOCS] Update site (#1295)
---
 content/docs/quick-start-guide.html | 52 ++++++++++++++++++-------------------
 content/index.html                  |  6 ++---
 content/releases.html               |  4 +--
 docs/_docs/1_1_quick_start_guide.md |  2 +-
 docs/_pages/index.md                |  2 +-
 docs/_pages/releases.md             |  2 +-
 6 files changed, 34 insertions(+), 34 deletions(-)

diff --git a/content/docs/quick-start-guide.html 
b/content/docs/quick-start-guide.html
index 6d8816a..9212e27 100644
--- a/content/docs/quick-start-guide.html
+++ b/content/docs/quick-start-guide.html
@@ -502,32 +502,32 @@ specific commit time and beginTime to “000” (denoting 
earliest possible comm
 <h2 id="deletes">Delete data</h2>
 <p>Delete records for the HoodieKeys passed in.</p>
 
-<div class="highlighter-rouge"><div class="highlight"><pre 
class="highlight"><code>// fetch total records count
-spark.sql("select uuid, partitionPath from hudi_ro_table").count()
-// fetch two records to be deleted
-val ds = spark.sql("select uuid, partitionPath from hudi_ro_table").limit(2)
-
-// issue deletes
-val deletes = dataGen.generateDeletes(ds.collectAsList())
-val df = spark.read.json(spark.sparkContext.parallelize(deletes, 2));
-df.write.format("org.apache.hudi").
-options(getQuickstartWriteConfigs).
-option(OPERATION_OPT_KEY,"delete").
-option(PRECOMBINE_FIELD_OPT_KEY, "ts").
-option(RECORDKEY_FIELD_OPT_KEY, "uuid").
-option(PARTITIONPATH_FIELD_OPT_KEY, "partitionpath").
-option(TABLE_NAME, tableName).
-mode(Append).
-save(basePath);
-
-// run the same read query as above.
-val roAfterDeleteViewDF = spark.
-    read.
-    format("org.apache.hudi").
-    load(basePath + "/*/*/*/*")
-roAfterDeleteViewDF.registerTempTable("hudi_ro_table")
-// fetch should return (total - 2) records
-spark.sql("select uuid, partitionPath from hudi_ro_table").count()
+<div class="language-scala highlighter-rouge"><div class="highlight"><pre 
class="highlight"><code><span class="c1">// fetch total records count
+</span><span class="nv">spark</span><span class="o">.</span><span 
class="py">sql</span><span class="o">(</span><span class="s">"select uuid, 
partitionPath from hudi_ro_table"</span><span class="o">).</span><span 
class="py">count</span><span class="o">()</span>
+<span class="c1">// fetch two records to be deleted
+</span><span class="k">val</span> <span class="nv">ds</span> <span 
class="k">=</span> <span class="nv">spark</span><span class="o">.</span><span 
class="py">sql</span><span class="o">(</span><span class="s">"select uuid, 
partitionPath from hudi_ro_table"</span><span class="o">).</span><span 
class="py">limit</span><span class="o">(</span><span class="mi">2</span><span 
class="o">)</span>
+
+<span class="c1">// issue deletes
+</span><span class="k">val</span> <span class="nv">deletes</span> <span 
class="k">=</span> <span class="nv">dataGen</span><span class="o">.</span><span 
class="py">generateDeletes</span><span class="o">(</span><span 
class="nv">ds</span><span class="o">.</span><span 
class="py">collectAsList</span><span class="o">())</span>
+<span class="k">val</span> <span class="nv">df</span> <span class="k">=</span> 
<span class="nv">spark</span><span class="o">.</span><span 
class="py">read</span><span class="o">.</span><span class="py">json</span><span 
class="o">(</span><span class="nv">spark</span><span class="o">.</span><span 
class="py">sparkContext</span><span class="o">.</span><span 
class="py">parallelize</span><span class="o">(</span><span 
class="n">deletes</span><span class="o">,</span> <span class="mi">2</span><spa 
[...]
+<span class="nv">df</span><span class="o">.</span><span 
class="py">write</span><span class="o">.</span><span 
class="py">format</span><span class="o">(</span><span 
class="s">"org.apache.hudi"</span><span class="o">).</span>
+<span class="nf">options</span><span class="o">(</span><span 
class="n">getQuickstartWriteConfigs</span><span class="o">).</span>
+<span class="nf">option</span><span class="o">(</span><span 
class="nc">OPERATION_OPT_KEY</span><span class="o">,</span><span 
class="s">"delete"</span><span class="o">).</span>
+<span class="nf">option</span><span class="o">(</span><span 
class="nc">PRECOMBINE_FIELD_OPT_KEY</span><span class="o">,</span> <span 
class="s">"ts"</span><span class="o">).</span>
+<span class="nf">option</span><span class="o">(</span><span 
class="nc">RECORDKEY_FIELD_OPT_KEY</span><span class="o">,</span> <span 
class="s">"uuid"</span><span class="o">).</span>
+<span class="nf">option</span><span class="o">(</span><span 
class="nc">PARTITIONPATH_FIELD_OPT_KEY</span><span class="o">,</span> <span 
class="s">"partitionpath"</span><span class="o">).</span>
+<span class="nf">option</span><span class="o">(</span><span 
class="nc">TABLE_NAME</span><span class="o">,</span> <span 
class="n">tableName</span><span class="o">).</span>
+<span class="nf">mode</span><span class="o">(</span><span 
class="nc">Append</span><span class="o">).</span>
+<span class="nf">save</span><span class="o">(</span><span 
class="n">basePath</span><span class="o">);</span>
+
+<span class="c1">// run the same read query as above.
+</span><span class="k">val</span> <span class="nv">roAfterDeleteViewDF</span> 
<span class="k">=</span> <span class="n">spark</span><span class="o">.</span>
+    <span class="n">read</span><span class="o">.</span>
+    <span class="nf">format</span><span class="o">(</span><span 
class="s">"org.apache.hudi"</span><span class="o">).</span>
+    <span class="nf">load</span><span class="o">(</span><span 
class="n">basePath</span> <span class="o">+</span> <span 
class="s">"/*/*/*/*"</span><span class="o">)</span>
+<span class="nv">roAfterDeleteViewDF</span><span class="o">.</span><span 
class="py">registerTempTable</span><span class="o">(</span><span 
class="s">"hudi_ro_table"</span><span class="o">)</span>
+<span class="c1">// fetch should return (total - 2) records
+</span><span class="nv">spark</span><span class="o">.</span><span 
class="py">sql</span><span class="o">(</span><span class="s">"select uuid, 
partitionPath from hudi_ro_table"</span><span class="o">).</span><span 
class="py">count</span><span class="o">()</span>
 </code></pre></div></div>
 <p>Note: Only <code class="highlighter-rouge">Append</code> mode is supported 
for delete operation.</p>
 
diff --git a/content/index.html b/content/index.html
index 657c8bc..554f792 100644
--- a/content/index.html
+++ b/content/index.html
@@ -4,7 +4,7 @@
     <meta charset="utf-8">
 
 <!-- begin _includes/seo.html --><title>Welcome to Apache Hudi ! - Apache 
Hudi</title>
-<meta name="description" content="Apache Hudi ingests &amp; manages storage of 
large analytical datasets over DFS (hdfs or cloud stores). Latest release 
0.5.0-incubating">
+<meta name="description" content="Apache Hudi ingests &amp; manages storage of 
large analytical datasets over DFS (hdfs or cloud stores). Latest release 
0.5.1-incubating">
 
 <meta property="og:type" content="website">
 <meta property="og:locale" content="en_US">
@@ -13,7 +13,7 @@
 <meta property="og:url" content="https://hudi.apache.org/";>
 
 
-  <meta property="og:description" content="Apache Hudi ingests &amp; manages 
storage of large analytical datasets over DFS (hdfs or cloud stores). Latest 
release 0.5.0-incubating">
+  <meta property="og:description" content="Apache Hudi ingests &amp; manages 
storage of large analytical datasets over DFS (hdfs or cloud stores). Latest 
release 0.5.1-incubating">
 
 
 
@@ -111,7 +111,7 @@
           Welcome to Apache Hudi !
         </h1>
 
-        <p class="page__lead">Apache Hudi ingests & manages storage of large 
analytical datasets over DFS (hdfs or cloud stores).<br /> <small><a 
href="https://github.com/apache/incubator-hudi/releases/tag/release-0.5.0-incubating";
 target="_blank">Latest release 0.5.0-incubating</a></small>
+        <p class="page__lead">Apache Hudi ingests & manages storage of large 
analytical datasets over DFS (hdfs or cloud stores).<br /> <small><a 
href="https://github.com/apache/incubator-hudi/releases/tag/release-0.5.1-incubating";
 target="_blank">Latest release 0.5.1-incubating</a></small>
 </p>
         <p>
           <a href="/docs/quick-start-guide.html" class="btn btn--light-outline 
btn--large"><i class="fa fa-paper-plane"></i> Get Started</a>
diff --git a/content/releases.html b/content/releases.html
index 8ef2aa7..2293ba9 100644
--- a/content/releases.html
+++ b/content/releases.html
@@ -181,7 +181,7 @@
           <nav class="toc">
             <header><h4 class="nav__title"><i class="fas fa-file-alt"></i> IN 
THIS PAGE</h4></header>
             <ul class="toc__menu">
-  <li><a href="#release-051-incubating">[Release 0.5.1-incubating]</a>
+  <li><a href="#release-051-incubating-docs">Release 0.5.1-incubating 
(docs)</a>
     <ul>
       <li><a href="#download-information">Download Information</a></li>
       <li><a href="#release-highlights">Release Highlights</a></li>
@@ -206,7 +206,7 @@
           </nav>
         </aside>
         
-        <h2 id="release-051-incubating">[Release 0.5.1-incubating]</h2>
+        <h2 id="release-051-incubating-docs"><a 
href="https://github.com/apache/incubator-hudi/releases/tag/release-0.5.1-incubating";>Release
 0.5.1-incubating</a> (<a href="/docs/quick-start-guide.html">docs</a>)</h2>
 
 <h3 id="download-information">Download Information</h3>
 <ul>
diff --git a/docs/_docs/1_1_quick_start_guide.md 
b/docs/_docs/1_1_quick_start_guide.md
index d7d645e..4a9c1b3 100644
--- a/docs/_docs/1_1_quick_start_guide.md
+++ b/docs/_docs/1_1_quick_start_guide.md
@@ -173,7 +173,7 @@ spark.sql("select `_hoodie_commit_time`, fare, begin_lon, 
begin_lat, ts from  hu
 ## Delete data {#deletes}
 Delete records for the HoodieKeys passed in.
 
-```
+```scala
 // fetch total records count
 spark.sql("select uuid, partitionPath from hudi_ro_table").count()
 // fetch two records to be deleted
diff --git a/docs/_pages/index.md b/docs/_pages/index.md
index 1f57a65..7190d6b 100644
--- a/docs/_pages/index.md
+++ b/docs/_pages/index.md
@@ -4,7 +4,7 @@ permalink: /
 title: Welcome to Apache Hudi !
 excerpt: >
   Apache Hudi ingests & manages storage of large analytical datasets over DFS 
(hdfs or cloud stores).<br />
-  <small><a 
href="https://github.com/apache/incubator-hudi/releases/tag/release-0.5.0-incubating";
 target="_blank">Latest release 0.5.0-incubating</a></small>
+  <small><a 
href="https://github.com/apache/incubator-hudi/releases/tag/release-0.5.1-incubating";
 target="_blank">Latest release 0.5.1-incubating</a></small>
 power_items:
   - img_path: /assets/images/powers/aws.jpg
   - img_path: /assets/images/powers/emis.jpg
diff --git a/docs/_pages/releases.md b/docs/_pages/releases.md
index 88a5d3e..8797e84 100644
--- a/docs/_pages/releases.md
+++ b/docs/_pages/releases.md
@@ -6,7 +6,7 @@ toc: true
 last_modified_at: 2019-12-30T15:59:57-04:00
 ---
 
-## [Release 0.5.1-incubating]
+## [Release 
0.5.1-incubating](https://github.com/apache/incubator-hudi/releases/tag/release-0.5.1-incubating)
 ([docs](/docs/quick-start-guide.html))
 
 ### Download Information
  * Source Release : [Apache Hudi(incubating) 0.5.1-incubating Source 
Release](https://www.apache.org/dist/incubator/hudi/0.5.1-incubating/hudi-0.5.1-incubating.src.tgz)
 
([asc](https://www.apache.org/dist/incubator/hudi/0.5.1-incubating/hudi-0.5.1-incubating.src.tgz.asc),
 
[sha512](https://www.apache.org/dist/incubator/hudi/0.5.1-incubating/hudi-0.5.1-incubating.src.tgz.sha512))

Reply via email to