This is an automated email from the ASF dual-hosted git repository. leesf pushed a commit to branch asf-site in repository https://gitbox.apache.org/repos/asf/incubator-hudi.git
The following commit(s) were added to refs/heads/asf-site by this push: new 6fc935a [DOCS] Update site (#1295) 6fc935a is described below commit 6fc935a949d2371edd6fd613649e63297ec3aa6d Author: Bhavani Sudha Saktheeswaran <bhasu...@uber.com> AuthorDate: Fri Jan 31 00:45:24 2020 -0800 [DOCS] Update site (#1295) --- content/docs/quick-start-guide.html | 52 ++++++++++++++++++------------------- content/index.html | 6 ++--- content/releases.html | 4 +-- docs/_docs/1_1_quick_start_guide.md | 2 +- docs/_pages/index.md | 2 +- docs/_pages/releases.md | 2 +- 6 files changed, 34 insertions(+), 34 deletions(-) diff --git a/content/docs/quick-start-guide.html b/content/docs/quick-start-guide.html index 6d8816a..9212e27 100644 --- a/content/docs/quick-start-guide.html +++ b/content/docs/quick-start-guide.html @@ -502,32 +502,32 @@ specific commit time and beginTime to “000” (denoting earliest possible comm <h2 id="deletes">Delete data</h2> <p>Delete records for the HoodieKeys passed in.</p> -<div class="highlighter-rouge"><div class="highlight"><pre class="highlight"><code>// fetch total records count -spark.sql("select uuid, partitionPath from hudi_ro_table").count() -// fetch two records to be deleted -val ds = spark.sql("select uuid, partitionPath from hudi_ro_table").limit(2) - -// issue deletes -val deletes = dataGen.generateDeletes(ds.collectAsList()) -val df = spark.read.json(spark.sparkContext.parallelize(deletes, 2)); -df.write.format("org.apache.hudi"). -options(getQuickstartWriteConfigs). -option(OPERATION_OPT_KEY,"delete"). -option(PRECOMBINE_FIELD_OPT_KEY, "ts"). -option(RECORDKEY_FIELD_OPT_KEY, "uuid"). -option(PARTITIONPATH_FIELD_OPT_KEY, "partitionpath"). -option(TABLE_NAME, tableName). -mode(Append). -save(basePath); - -// run the same read query as above. -val roAfterDeleteViewDF = spark. - read. - format("org.apache.hudi"). - load(basePath + "/*/*/*/*") -roAfterDeleteViewDF.registerTempTable("hudi_ro_table") -// fetch should return (total - 2) records -spark.sql("select uuid, partitionPath from hudi_ro_table").count() +<div class="language-scala highlighter-rouge"><div class="highlight"><pre class="highlight"><code><span class="c1">// fetch total records count +</span><span class="nv">spark</span><span class="o">.</span><span class="py">sql</span><span class="o">(</span><span class="s">"select uuid, partitionPath from hudi_ro_table"</span><span class="o">).</span><span class="py">count</span><span class="o">()</span> +<span class="c1">// fetch two records to be deleted +</span><span class="k">val</span> <span class="nv">ds</span> <span class="k">=</span> <span class="nv">spark</span><span class="o">.</span><span class="py">sql</span><span class="o">(</span><span class="s">"select uuid, partitionPath from hudi_ro_table"</span><span class="o">).</span><span class="py">limit</span><span class="o">(</span><span class="mi">2</span><span class="o">)</span> + +<span class="c1">// issue deletes +</span><span class="k">val</span> <span class="nv">deletes</span> <span class="k">=</span> <span class="nv">dataGen</span><span class="o">.</span><span class="py">generateDeletes</span><span class="o">(</span><span class="nv">ds</span><span class="o">.</span><span class="py">collectAsList</span><span class="o">())</span> +<span class="k">val</span> <span class="nv">df</span> <span class="k">=</span> <span class="nv">spark</span><span class="o">.</span><span class="py">read</span><span class="o">.</span><span class="py">json</span><span class="o">(</span><span class="nv">spark</span><span class="o">.</span><span class="py">sparkContext</span><span class="o">.</span><span class="py">parallelize</span><span class="o">(</span><span class="n">deletes</span><span class="o">,</span> <span class="mi">2</span><spa [...] +<span class="nv">df</span><span class="o">.</span><span class="py">write</span><span class="o">.</span><span class="py">format</span><span class="o">(</span><span class="s">"org.apache.hudi"</span><span class="o">).</span> +<span class="nf">options</span><span class="o">(</span><span class="n">getQuickstartWriteConfigs</span><span class="o">).</span> +<span class="nf">option</span><span class="o">(</span><span class="nc">OPERATION_OPT_KEY</span><span class="o">,</span><span class="s">"delete"</span><span class="o">).</span> +<span class="nf">option</span><span class="o">(</span><span class="nc">PRECOMBINE_FIELD_OPT_KEY</span><span class="o">,</span> <span class="s">"ts"</span><span class="o">).</span> +<span class="nf">option</span><span class="o">(</span><span class="nc">RECORDKEY_FIELD_OPT_KEY</span><span class="o">,</span> <span class="s">"uuid"</span><span class="o">).</span> +<span class="nf">option</span><span class="o">(</span><span class="nc">PARTITIONPATH_FIELD_OPT_KEY</span><span class="o">,</span> <span class="s">"partitionpath"</span><span class="o">).</span> +<span class="nf">option</span><span class="o">(</span><span class="nc">TABLE_NAME</span><span class="o">,</span> <span class="n">tableName</span><span class="o">).</span> +<span class="nf">mode</span><span class="o">(</span><span class="nc">Append</span><span class="o">).</span> +<span class="nf">save</span><span class="o">(</span><span class="n">basePath</span><span class="o">);</span> + +<span class="c1">// run the same read query as above. +</span><span class="k">val</span> <span class="nv">roAfterDeleteViewDF</span> <span class="k">=</span> <span class="n">spark</span><span class="o">.</span> + <span class="n">read</span><span class="o">.</span> + <span class="nf">format</span><span class="o">(</span><span class="s">"org.apache.hudi"</span><span class="o">).</span> + <span class="nf">load</span><span class="o">(</span><span class="n">basePath</span> <span class="o">+</span> <span class="s">"/*/*/*/*"</span><span class="o">)</span> +<span class="nv">roAfterDeleteViewDF</span><span class="o">.</span><span class="py">registerTempTable</span><span class="o">(</span><span class="s">"hudi_ro_table"</span><span class="o">)</span> +<span class="c1">// fetch should return (total - 2) records +</span><span class="nv">spark</span><span class="o">.</span><span class="py">sql</span><span class="o">(</span><span class="s">"select uuid, partitionPath from hudi_ro_table"</span><span class="o">).</span><span class="py">count</span><span class="o">()</span> </code></pre></div></div> <p>Note: Only <code class="highlighter-rouge">Append</code> mode is supported for delete operation.</p> diff --git a/content/index.html b/content/index.html index 657c8bc..554f792 100644 --- a/content/index.html +++ b/content/index.html @@ -4,7 +4,7 @@ <meta charset="utf-8"> <!-- begin _includes/seo.html --><title>Welcome to Apache Hudi ! - Apache Hudi</title> -<meta name="description" content="Apache Hudi ingests & manages storage of large analytical datasets over DFS (hdfs or cloud stores). Latest release 0.5.0-incubating"> +<meta name="description" content="Apache Hudi ingests & manages storage of large analytical datasets over DFS (hdfs or cloud stores). Latest release 0.5.1-incubating"> <meta property="og:type" content="website"> <meta property="og:locale" content="en_US"> @@ -13,7 +13,7 @@ <meta property="og:url" content="https://hudi.apache.org/"> - <meta property="og:description" content="Apache Hudi ingests & manages storage of large analytical datasets over DFS (hdfs or cloud stores). Latest release 0.5.0-incubating"> + <meta property="og:description" content="Apache Hudi ingests & manages storage of large analytical datasets over DFS (hdfs or cloud stores). Latest release 0.5.1-incubating"> @@ -111,7 +111,7 @@ Welcome to Apache Hudi ! </h1> - <p class="page__lead">Apache Hudi ingests & manages storage of large analytical datasets over DFS (hdfs or cloud stores).<br /> <small><a href="https://github.com/apache/incubator-hudi/releases/tag/release-0.5.0-incubating" target="_blank">Latest release 0.5.0-incubating</a></small> + <p class="page__lead">Apache Hudi ingests & manages storage of large analytical datasets over DFS (hdfs or cloud stores).<br /> <small><a href="https://github.com/apache/incubator-hudi/releases/tag/release-0.5.1-incubating" target="_blank">Latest release 0.5.1-incubating</a></small> </p> <p> <a href="/docs/quick-start-guide.html" class="btn btn--light-outline btn--large"><i class="fa fa-paper-plane"></i> Get Started</a> diff --git a/content/releases.html b/content/releases.html index 8ef2aa7..2293ba9 100644 --- a/content/releases.html +++ b/content/releases.html @@ -181,7 +181,7 @@ <nav class="toc"> <header><h4 class="nav__title"><i class="fas fa-file-alt"></i> IN THIS PAGE</h4></header> <ul class="toc__menu"> - <li><a href="#release-051-incubating">[Release 0.5.1-incubating]</a> + <li><a href="#release-051-incubating-docs">Release 0.5.1-incubating (docs)</a> <ul> <li><a href="#download-information">Download Information</a></li> <li><a href="#release-highlights">Release Highlights</a></li> @@ -206,7 +206,7 @@ </nav> </aside> - <h2 id="release-051-incubating">[Release 0.5.1-incubating]</h2> + <h2 id="release-051-incubating-docs"><a href="https://github.com/apache/incubator-hudi/releases/tag/release-0.5.1-incubating">Release 0.5.1-incubating</a> (<a href="/docs/quick-start-guide.html">docs</a>)</h2> <h3 id="download-information">Download Information</h3> <ul> diff --git a/docs/_docs/1_1_quick_start_guide.md b/docs/_docs/1_1_quick_start_guide.md index d7d645e..4a9c1b3 100644 --- a/docs/_docs/1_1_quick_start_guide.md +++ b/docs/_docs/1_1_quick_start_guide.md @@ -173,7 +173,7 @@ spark.sql("select `_hoodie_commit_time`, fare, begin_lon, begin_lat, ts from hu ## Delete data {#deletes} Delete records for the HoodieKeys passed in. -``` +```scala // fetch total records count spark.sql("select uuid, partitionPath from hudi_ro_table").count() // fetch two records to be deleted diff --git a/docs/_pages/index.md b/docs/_pages/index.md index 1f57a65..7190d6b 100644 --- a/docs/_pages/index.md +++ b/docs/_pages/index.md @@ -4,7 +4,7 @@ permalink: / title: Welcome to Apache Hudi ! excerpt: > Apache Hudi ingests & manages storage of large analytical datasets over DFS (hdfs or cloud stores).<br /> - <small><a href="https://github.com/apache/incubator-hudi/releases/tag/release-0.5.0-incubating" target="_blank">Latest release 0.5.0-incubating</a></small> + <small><a href="https://github.com/apache/incubator-hudi/releases/tag/release-0.5.1-incubating" target="_blank">Latest release 0.5.1-incubating</a></small> power_items: - img_path: /assets/images/powers/aws.jpg - img_path: /assets/images/powers/emis.jpg diff --git a/docs/_pages/releases.md b/docs/_pages/releases.md index 88a5d3e..8797e84 100644 --- a/docs/_pages/releases.md +++ b/docs/_pages/releases.md @@ -6,7 +6,7 @@ toc: true last_modified_at: 2019-12-30T15:59:57-04:00 --- -## [Release 0.5.1-incubating] +## [Release 0.5.1-incubating](https://github.com/apache/incubator-hudi/releases/tag/release-0.5.1-incubating) ([docs](/docs/quick-start-guide.html)) ### Download Information * Source Release : [Apache Hudi(incubating) 0.5.1-incubating Source Release](https://www.apache.org/dist/incubator/hudi/0.5.1-incubating/hudi-0.5.1-incubating.src.tgz) ([asc](https://www.apache.org/dist/incubator/hudi/0.5.1-incubating/hudi-0.5.1-incubating.src.tgz.asc), [sha512](https://www.apache.org/dist/incubator/hudi/0.5.1-incubating/hudi-0.5.1-incubating.src.tgz.sha512))