This is an automated email from the ASF dual-hosted git repository.

vinoth pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/hudi.git


The following commit(s) were added to refs/heads/asf-site by this push:
     new b5b966c  Travis CI build asf-site
b5b966c is described below

commit b5b966c5ec6f73a0888da11f9df4a413a7494027
Author: CI <[email protected]>
AuthorDate: Mon May 25 15:58:54 2020 +0000

    Travis CI build asf-site
---
 content/assets/js/lunr/lunr-store.js |  2 +-
 content/docs/powered_by.html         | 34 +++++++++++++++++++++++-----------
 2 files changed, 24 insertions(+), 12 deletions(-)

diff --git a/content/assets/js/lunr/lunr-store.js 
b/content/assets/js/lunr/lunr-store.js
index e5f3028..d133410 100644
--- a/content/assets/js/lunr/lunr-store.js
+++ b/content/assets/js/lunr/lunr-store.js
@@ -575,7 +575,7 @@ var store = [{
         "url": "https://hudi.apache.org/cn/docs/powered_by.html";,
         "teaser":"https://hudi.apache.org/assets/images/500x300.png"},{
         "title": "Talks & Powered By",
-        "excerpt":"Adoption Uber Apache Hudi was originally developed at Uber, 
to achieve low latency database ingestion, with high efficiency. It has been in 
production since Aug 2016, powering the massive 100PB data lake, including 
highly business critical tables like core trips,riders,partners. It also powers 
several incremental Hive ETL pipelines and being...","categories": [],
+        "excerpt":"Adoption Alibaba Cloud Alibaba Cloud provides cloud 
computing services to online businesses and Alibaba’s own e-commerce ecosystem, 
Apache Hudi is integrated into Alibaba Cloud Data Lake Analytics offering 
real-time analysis on hudi dataset. Amazon Web Services Amazon Web Services is 
the World’s leading cloud services provider. Apache Hudi is 
pre-installed...","categories": [],
         "tags": [],
         "url": "https://hudi.apache.org/docs/powered_by.html";,
         "teaser":"https://hudi.apache.org/assets/images/500x300.png"},{
diff --git a/content/docs/powered_by.html b/content/docs/powered_by.html
index ec4f001..3421d7c 100644
--- a/content/docs/powered_by.html
+++ b/content/docs/powered_by.html
@@ -343,12 +343,6 @@
         
         <h2 id="adoption">Adoption</h2>
 
-<h3 id="uber">Uber</h3>
-
-<p>Apache Hudi was originally developed at <a 
href="https://uber.com";>Uber</a>, to achieve <a 
href="http://www.slideshare.net/vinothchandar/hadoop-strata-talk-uber-your-hadoop-has-arrived/32";>low
 latency database ingestion, with high efficiency</a>.
-It has been in production since Aug 2016, powering the massive <a 
href="https://eng.uber.com/uber-big-data-platform/";>100PB data lake</a>, 
including highly business critical tables like core trips,riders,partners. It 
also 
-powers several incremental Hive ETL pipelines and being currently integrated 
into Uber’s data dispersal system.</p>
-
 <h3 id="alibaba-cloud">Alibaba Cloud</h3>
 <p>Alibaba Cloud provides cloud computing services to online businesses and 
Alibaba’s own e-commerce ecosystem, Apache Hudi is integrated into Alibaba 
Cloud <a href="https://www.alibabacloud.com/help/product/70174.htm";>Data Lake 
Analytics</a>
 offering real-time analysis on hudi dataset.</p>
@@ -361,17 +355,35 @@ offering, providing means for AWS users to perform 
record-level updates/deletes
 
 <p><a href="https://www.emishealth.com/";>EMIS Health</a> is the largest 
provider of Primary Care IT software in the UK with datasets including more 
than 500Bn healthcare records. HUDI is used to manage their analytics dataset 
in production and keeping them up-to-date with their upstream source. Presto is 
being used to query the data written in HUDI format.</p>
 
+<h3 id="logical-clocks">Logical Clocks</h3>
+
+<p><a 
href="https://www.logicalclocks.com/blog/introducing-the-hopsworks-1-x-series";>Hopsworks
 1.x series</a> supports Apache Hudi feature groups, to enable upserts and time 
travel.</p>
+
+<h3 id="tathastuai">Tathastu.ai</h3>
+
+<p><a href="https://www.tathastu.ai";>Tathastu.ai</a> offers the largest AI/ML 
playground of consumer data for data scientists, AI experts and technologists 
to build upon. They have built a CDC pipeline using Apache Hudi and Debezium. 
Data from Hudi datasets is being queried using Hive, Presto and Spark.</p>
+
+<h3 id="tencent">Tencent</h3>
+
+<p><a href="https://intl.cloud.tencent.com/product/emr";>EMR from Tencent</a> 
Cloud has integrated Hudi as one of its BigData components <a 
href="https://intl.cloud.tencent.com/document/product/1026/35587";>since 
V2.2.0</a>. Using Hudi, the end-users can handle either read-heavy or 
write-heavy use cases, and Hudi will manage the underlying data stored on 
HDFS/COS/CHDFS using Apache Parquet and Apache Avro.</p>
+
+<h3 id="uber">Uber</h3>
+
+<p>Apache Hudi was originally developed at <a 
href="https://uber.com";>Uber</a>, to achieve <a 
href="http://www.slideshare.net/vinothchandar/hadoop-strata-talk-uber-your-hadoop-has-arrived/32";>low
 latency database ingestion, with high efficiency</a>.
+It has been in production since Aug 2016, powering the massive <a 
href="https://eng.uber.com/uber-big-data-platform/";>100PB data lake</a>, 
including highly business critical tables like core trips,riders,partners. It 
also 
+powers several incremental Hive ETL pipelines and being currently integrated 
into Uber’s data dispersal system.</p>
+
+<h3 id="udemy">Udemy</h3>
+
+<p>At <a href="https://www.udemy.com/";>Udemy</a>, Apache Hudi on AWS EMR is 
used to perform ingest MySQL change data capture.</p>
+
 <h3 id="yieldsio">Yields.io</h3>
 
 <p>Yields.io is the first FinTech platform that uses AI for automated model 
validation and real-time monitoring on an enterprise-wide scale. Their <a 
href="https://www.yields.io/Blog/Apache-Hudi-at-Yields";>data lake</a> is 
managed by Hudi. They are also actively building their infrastructure for 
incremental, cross language/platform machine learning using Hudi.</p>
 
 <h3 id="yotpo">Yotpo</h3>
 
-<p>Using Hudi at Yotpo for several usages. Firstly, integrated Hudi as a 
writer in their open source ETL framework https://github.com/YotpoLtd/metorikku 
and using as an output writer for a CDC pipeline, with events that are being 
generated from a database binlog streams to Kafka and then are written to 
S3.</p>
-
-<h3 id="tathastuai">Tathastu.ai</h3>
-
-<p><a href="https://www.tathastu.ai";>Tathastu.ai</a> offers the largest AI/ML 
playground of consumer data for data scientists, AI experts and technologists 
to build upon. They have built a CDC pipeline using Apache Hudi and Debezium. 
Data from Hudi datasets is being queried using Hive, Presto and Spark.</p>
+<p>Using Hudi at Yotpo for several usages. Firstly, integrated Hudi as a 
writer in their open source ETL framework, <a 
href="https://github.com/YotpoLtd/metorikku";>Metorikku</a> and using as an 
output writer for a CDC pipeline, with events that are being generated from a 
database binlog streams to Kafka and then are written to S3.</p>
 
 <h2 id="talks--presentations">Talks &amp; Presentations</h2>
 

Reply via email to