This is an automated email from the ASF dual-hosted git repository.
vinoth pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/hudi.git
The following commit(s) were added to refs/heads/asf-site by this push:
new cc08429 Travis CI build asf-site
cc08429 is described below
commit cc084298091d453a1383807e2fe9e516058eb946
Author: CI <[email protected]>
AuthorDate: Sat Feb 13 08:01:25 2021 +0000
Travis CI build asf-site
---
content/docs/0.7.0-writing_data.html | 2 +-
content/docs/writing_data.html | 2 +-
2 files changed, 2 insertions(+), 2 deletions(-)
diff --git a/content/docs/0.7.0-writing_data.html
b/content/docs/0.7.0-writing_data.html
index a5a4a9d..abdca7e 100644
--- a/content/docs/0.7.0-writing_data.html
+++ b/content/docs/0.7.0-writing_data.html
@@ -357,7 +357,7 @@ can be chosen/changed across each commit/deltacommit issued
against the table.</
<ul>
<li><strong>UPSERT</strong> : This is the default operation where the input
records are first tagged as inserts or updates by looking up the index.
The records are ultimately written after heuristics are run to determine how
best to pack them on storage to optimize for things like file sizing.
- This operation is recommended for use-cases like database change capture
where the input almost certainly contains updates.</li>
+ This operation is recommended for use-cases like database change capture
where the input almost certainly contains updates. The target table will never
show duplicates.</li>
<li><strong>INSERT</strong> : This operation is very similar to upsert in
terms of heuristics/file sizing but completely skips the index lookup step.
Thus, it can be a lot faster than upserts
for use-cases like log de-duplication (in conjunction with options to filter
duplicates mentioned below). This is also suitable for use-cases where the
table can tolerate duplicates, but just
need the transactional writes/incremental pull/storage management
capabilities of Hudi.</li>
diff --git a/content/docs/writing_data.html b/content/docs/writing_data.html
index 0ab4f2b..11ce29f 100644
--- a/content/docs/writing_data.html
+++ b/content/docs/writing_data.html
@@ -357,7 +357,7 @@ can be chosen/changed across each commit/deltacommit issued
against the table.</
<ul>
<li><strong>UPSERT</strong> : This is the default operation where the input
records are first tagged as inserts or updates by looking up the index.
The records are ultimately written after heuristics are run to determine how
best to pack them on storage to optimize for things like file sizing.
- This operation is recommended for use-cases like database change capture
where the input almost certainly contains updates.</li>
+ This operation is recommended for use-cases like database change capture
where the input almost certainly contains updates. The target table will never
show duplicates.</li>
<li><strong>INSERT</strong> : This operation is very similar to upsert in
terms of heuristics/file sizing but completely skips the index lookup step.
Thus, it can be a lot faster than upserts
for use-cases like log de-duplication (in conjunction with options to filter
duplicates mentioned below). This is also suitable for use-cases where the
table can tolerate duplicates, but just
need the transactional writes/incremental pull/storage management
capabilities of Hudi.</li>