This is an automated email from the ASF dual-hosted git repository.
vinoth pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/hudi.git
The following commit(s) were added to refs/heads/asf-site by this push:
new b5b966c Travis CI build asf-site
b5b966c is described below
commit b5b966c5ec6f73a0888da11f9df4a413a7494027
Author: CI <[email protected]>
AuthorDate: Mon May 25 15:58:54 2020 +0000
Travis CI build asf-site
---
content/assets/js/lunr/lunr-store.js | 2 +-
content/docs/powered_by.html | 34 +++++++++++++++++++++++-----------
2 files changed, 24 insertions(+), 12 deletions(-)
diff --git a/content/assets/js/lunr/lunr-store.js
b/content/assets/js/lunr/lunr-store.js
index e5f3028..d133410 100644
--- a/content/assets/js/lunr/lunr-store.js
+++ b/content/assets/js/lunr/lunr-store.js
@@ -575,7 +575,7 @@ var store = [{
"url": "https://hudi.apache.org/cn/docs/powered_by.html",
"teaser":"https://hudi.apache.org/assets/images/500x300.png"},{
"title": "Talks & Powered By",
- "excerpt":"Adoption Uber Apache Hudi was originally developed at Uber,
to achieve low latency database ingestion, with high efficiency. It has been in
production since Aug 2016, powering the massive 100PB data lake, including
highly business critical tables like core trips,riders,partners. It also powers
several incremental Hive ETL pipelines and being...","categories": [],
+ "excerpt":"Adoption Alibaba Cloud Alibaba Cloud provides cloud
computing services to online businesses and Alibaba’s own e-commerce ecosystem,
Apache Hudi is integrated into Alibaba Cloud Data Lake Analytics offering
real-time analysis on hudi dataset. Amazon Web Services Amazon Web Services is
the World’s leading cloud services provider. Apache Hudi is
pre-installed...","categories": [],
"tags": [],
"url": "https://hudi.apache.org/docs/powered_by.html",
"teaser":"https://hudi.apache.org/assets/images/500x300.png"},{
diff --git a/content/docs/powered_by.html b/content/docs/powered_by.html
index ec4f001..3421d7c 100644
--- a/content/docs/powered_by.html
+++ b/content/docs/powered_by.html
@@ -343,12 +343,6 @@
<h2 id="adoption">Adoption</h2>
-<h3 id="uber">Uber</h3>
-
-<p>Apache Hudi was originally developed at <a
href="https://uber.com">Uber</a>, to achieve <a
href="http://www.slideshare.net/vinothchandar/hadoop-strata-talk-uber-your-hadoop-has-arrived/32">low
latency database ingestion, with high efficiency</a>.
-It has been in production since Aug 2016, powering the massive <a
href="https://eng.uber.com/uber-big-data-platform/">100PB data lake</a>,
including highly business critical tables like core trips,riders,partners. It
also
-powers several incremental Hive ETL pipelines and being currently integrated
into Uber’s data dispersal system.</p>
-
<h3 id="alibaba-cloud">Alibaba Cloud</h3>
<p>Alibaba Cloud provides cloud computing services to online businesses and
Alibaba’s own e-commerce ecosystem, Apache Hudi is integrated into Alibaba
Cloud <a href="https://www.alibabacloud.com/help/product/70174.htm">Data Lake
Analytics</a>
offering real-time analysis on hudi dataset.</p>
@@ -361,17 +355,35 @@ offering, providing means for AWS users to perform
record-level updates/deletes
<p><a href="https://www.emishealth.com/">EMIS Health</a> is the largest
provider of Primary Care IT software in the UK with datasets including more
than 500Bn healthcare records. HUDI is used to manage their analytics dataset
in production and keeping them up-to-date with their upstream source. Presto is
being used to query the data written in HUDI format.</p>
+<h3 id="logical-clocks">Logical Clocks</h3>
+
+<p><a
href="https://www.logicalclocks.com/blog/introducing-the-hopsworks-1-x-series">Hopsworks
1.x series</a> supports Apache Hudi feature groups, to enable upserts and time
travel.</p>
+
+<h3 id="tathastuai">Tathastu.ai</h3>
+
+<p><a href="https://www.tathastu.ai">Tathastu.ai</a> offers the largest AI/ML
playground of consumer data for data scientists, AI experts and technologists
to build upon. They have built a CDC pipeline using Apache Hudi and Debezium.
Data from Hudi datasets is being queried using Hive, Presto and Spark.</p>
+
+<h3 id="tencent">Tencent</h3>
+
+<p><a href="https://intl.cloud.tencent.com/product/emr">EMR from Tencent</a>
Cloud has integrated Hudi as one of its BigData components <a
href="https://intl.cloud.tencent.com/document/product/1026/35587">since
V2.2.0</a>. Using Hudi, the end-users can handle either read-heavy or
write-heavy use cases, and Hudi will manage the underlying data stored on
HDFS/COS/CHDFS using Apache Parquet and Apache Avro.</p>
+
+<h3 id="uber">Uber</h3>
+
+<p>Apache Hudi was originally developed at <a
href="https://uber.com">Uber</a>, to achieve <a
href="http://www.slideshare.net/vinothchandar/hadoop-strata-talk-uber-your-hadoop-has-arrived/32">low
latency database ingestion, with high efficiency</a>.
+It has been in production since Aug 2016, powering the massive <a
href="https://eng.uber.com/uber-big-data-platform/">100PB data lake</a>,
including highly business critical tables like core trips,riders,partners. It
also
+powers several incremental Hive ETL pipelines and being currently integrated
into Uber’s data dispersal system.</p>
+
+<h3 id="udemy">Udemy</h3>
+
+<p>At <a href="https://www.udemy.com/">Udemy</a>, Apache Hudi on AWS EMR is
used to perform ingest MySQL change data capture.</p>
+
<h3 id="yieldsio">Yields.io</h3>
<p>Yields.io is the first FinTech platform that uses AI for automated model
validation and real-time monitoring on an enterprise-wide scale. Their <a
href="https://www.yields.io/Blog/Apache-Hudi-at-Yields">data lake</a> is
managed by Hudi. They are also actively building their infrastructure for
incremental, cross language/platform machine learning using Hudi.</p>
<h3 id="yotpo">Yotpo</h3>
-<p>Using Hudi at Yotpo for several usages. Firstly, integrated Hudi as a
writer in their open source ETL framework https://github.com/YotpoLtd/metorikku
and using as an output writer for a CDC pipeline, with events that are being
generated from a database binlog streams to Kafka and then are written to
S3.</p>
-
-<h3 id="tathastuai">Tathastu.ai</h3>
-
-<p><a href="https://www.tathastu.ai">Tathastu.ai</a> offers the largest AI/ML
playground of consumer data for data scientists, AI experts and technologists
to build upon. They have built a CDC pipeline using Apache Hudi and Debezium.
Data from Hudi datasets is being queried using Hive, Presto and Spark.</p>
+<p>Using Hudi at Yotpo for several usages. Firstly, integrated Hudi as a
writer in their open source ETL framework, <a
href="https://github.com/YotpoLtd/metorikku">Metorikku</a> and using as an
output writer for a CDC pipeline, with events that are being generated from a
database binlog streams to Kafka and then are written to S3.</p>
<h2 id="talks--presentations">Talks & Presentations</h2>