This is an automated email from the ASF dual-hosted git repository.
madhan pushed a commit to branch branch-2.0
in repository https://gitbox.apache.org/repos/asf/atlas.git
The following commit(s) were added to refs/heads/branch-2.0 by this push:
new 9734248 ATLAS-4045: fixed typos and other grammatical mistakes in
export/import API documentation
9734248 is described below
commit 97342489eb504e0766d3ef44945624877d5ad1d6
Author: Umesh Padashetty <[email protected]>
AuthorDate: Fri Nov 20 21:08:05 2020 +0530
ATLAS-4045: fixed typos and other grammatical mistakes in export/import API
documentation
Signed-off-by: Madhan Neethiraj <[email protected]>
(cherry picked from commit 099324e801accaf3b4be24028a9d6db2c4b36ed9)
---
docs/src/documents/Import-Export/ExportAPI.md | 14 +++++------
docs/src/documents/Import-Export/ExportHDFSAPI.md | 6 ++---
.../documents/Import-Export/ExportImportAudits.md | 2 +-
docs/src/documents/Import-Export/ImportAPI.md | 8 +++----
.../documents/Import-Export/ImportAPIOptions.md | 28 +++++++++++-----------
.../Import-Export/ImportEntityTransforms.md | 22 ++++++++---------
.../src/documents/Import-Export/ImportExportAPI.md | 6 ++---
.../documents/Import-Export/IncrementalExport.md | 4 ++--
8 files changed, 45 insertions(+), 45 deletions(-)
diff --git a/docs/src/documents/Import-Export/ExportAPI.md
b/docs/src/documents/Import-Export/ExportAPI.md
index f240e02..849d452 100644
--- a/docs/src/documents/Import-Export/ExportAPI.md
+++ b/docs/src/documents/Import-Export/ExportAPI.md
@@ -23,7 +23,7 @@ See [here](#/ExportHDFSAPI) for details on exporting
*hdfs_path* entities.
| _URL_ |_api/atlas/admin/export_ |
| _Method_ |_POST_ |
| _URL Parameters_ |_None_ |
-| _Data Parameters_| The class _AtlasExportRequest_ is used to specify the
items to export. The list of _AtlasObjectId_(s) allow for specifying the
multiple items to export in a session. The _AtlasObjectId_ is a tuple of entity
type, name of unique attribute, value of unique attribute. Several items can be
specified. See examples below.|
+| _Data Parameters_| The class _AtlasExportRequest_ is used to specify the
items to export. The list of _AtlasObjectId_(s) allows for specifying the
multiple items to export in a session. The _AtlasObjectId_ is a tuple of entity
type, name of unique attribute, value of unique attribute. Several items can be
specified. See examples below.|
| _Success Response_|File stream as _application/zip_.|
|_Error Response_|Errors that are handled within the system will be returned
as _AtlasBaseException_. |
| _Notes_ | Consumer could choose to consume the output of the API by
programmatically using _java.io.ByteOutputStream_ or by manually, save the
contents of the stream to a file on the disk.|
@@ -39,10 +39,10 @@ __Method Signature__
### Additional Options
It is possible to specify additional parameters for the _Export_ operation.
-Current implementation has 2 options. Both are optional:
+The current implementation has 2 options. Both are optional:
-* _matchType_ This option configures the approach used for fetching the
starting entity. It has follow values:
+* _matchType_ This option configures the approach used for fetching the
starting entity. It has the following values:
* _startsWith_ Search for an entity that is prefixed with the specified
criteria.
* _endsWith_ Search for an entity that is suffixed with the specified
criteria.
* _contains_ Search for an entity that has the specified criteria as a
sub-string.
@@ -52,20 +52,20 @@ Current implementation has 2 options. Both are optional:
-* _fetchType_ This option configures the approach used for fetching entities.
It has following values:
+* _fetchType_ This option configures the approach used for fetching entities.
It has the following values:
* _FULL_: This fetches all the entities that are connected directly and
indirectly to the starting entity. E.g. If a starting entity specified is a
table, then this option will fetch the table, database and all the other tables
within the database.
* _CONNECTED_: This fetches all the etnties that are connected directly to
the starting entity. E.g. If a starting entity specified is a table, then this
option will fetch the table and the database entity only.
* _INCREMENTAL_: See [here](#/IncrementalExport) for details.
-If no _matchType_ is specified, exact match is used. Which means, that the
entire string is used in the search criteria.
+If no _matchType_ is specified, an exact match is used. Which means, that the
entire string is used in the search criteria.
-Searching using _matchType_ applies for all types of entities. It is
particularly useful for matching entities of type hdfs_path (see
[here](#/ExportHDFSAPI)).
+Searching using _matchType_ applies to all types of entities. It is
particularly useful for matching entities of type hdfs_path (see
[here](#/ExportHDFSAPI)).
The _fetchType_ option defaults to _FULL_.
-For complete example see section below.
+For a complete example see the section below.
### Contents of Exported ZIP File
diff --git a/docs/src/documents/Import-Export/ExportHDFSAPI.md
b/docs/src/documents/Import-Export/ExportHDFSAPI.md
index 9f94501..bce0065 100644
--- a/docs/src/documents/Import-Export/ExportHDFSAPI.md
+++ b/docs/src/documents/Import-Export/ExportHDFSAPI.md
@@ -30,10 +30,10 @@ __Sample HDFS Setup__
### Export API Using matchType
-To export entities that represent HDFS path, use the Export API using the
_matchType_ option. Details can be found [here](#/ExportAPI).
+To export entities that represent the HDFS path, use the Export API using the
_matchType_ option. Details can be found [here](#/ExportAPI).
### Example Using CURL Calls
-Below are sample CURL calls that performs export operation on the _Sample HDFS
Setup_ shown above.
+Below are sample CURL calls perform an export operation on the _Sample HDFS
Setup_ shown above.
<SyntaxHighlighter wrapLines={true} language="shell" style={theme.dark}>
{`curl -X POST -u adminuser:password -H "Content-Type: application/json" -H
"Cache-Control: no-cache" -d '{
@@ -48,4 +48,4 @@ Below are sample CURL calls that performs export operation on
the _Sample HDFS S
</SyntaxHighlighter>
### Automatic Creation of HDFS entities
-Given that HDFS entity creation is a manual process. The Export API offers a
mechanism for creation of requested HDFS entities.
+Given that HDFS entity creation is a manual process. The Export API offers a
mechanism for the creation of requested HDFS entities.
diff --git a/docs/src/documents/Import-Export/ExportImportAudits.md
b/docs/src/documents/Import-Export/ExportImportAudits.md
index c481582..615d0f0 100644
--- a/docs/src/documents/Import-Export/ExportImportAudits.md
+++ b/docs/src/documents/Import-Export/ExportImportAudits.md
@@ -13,7 +13,7 @@ import SyntaxHighlighter from 'react-syntax-highlighter';
#### Background
-The new audits for Export and Import operations also have corresponding REST
APIs to programatically fetch the audit entries.
+The new audits for Export and Import operations also have corresponding REST
APIs to programmatically fetch the audit entries.
#### REST APIs
diff --git a/docs/src/documents/Import-Export/ImportAPI.md
b/docs/src/documents/Import-Export/ImportAPI.md
index 58cf136..0cae7f8 100644
--- a/docs/src/documents/Import-Export/ImportAPI.md
+++ b/docs/src/documents/Import-Export/ImportAPI.md
@@ -14,7 +14,7 @@ import SyntaxHighlighter from 'react-syntax-highlighter';
The general approach is:
* Consumer makes a ZIP file available for import operation. See details
below for the 2 flavors of the API.
* The API if successful, will return the results of the operation.
- * Error will be returned on failure of the call.
+ * Error will be returned on the failure of the call.
### Import ZIP File Using POST
@@ -67,9 +67,9 @@ Please see [here](#/ImportAPIOptions) for the available
options during import pr
__AtlasImportResult Response__
The API will return the results of the import operation in the format defined
by the _AtlasImportResult_:
- * _AtlasImportParameters_: This contains a collection of name value pair of
the options that are applied during the import operation.
+ * _AtlasImportParameters_: This contains a collection of name-value pair of
the options that are applied during the import operation.
* _Metrics_: Operation metrics. These include details on the number of
types imported, number of entities imported, etc.
- * _Processed Entities_: Contains list of GUIDs for the entities that were
processed.
+ * _Processed Entities_: Contains a list of GUIDs for the entities that were
processed.
* _Operation Status_: Overall status of the operation. Values are
_SUCCESS_, PARTIAL_SUCCESS, _FAIL_.
### Examples Using CURL Calls
@@ -93,7 +93,7 @@ The _request_ parameter is optional. If import has to be run
without any options
</SyntaxHighlighter>
-The call below performs Import of _QuickStart_ database using a ZIP file
available on server.
+The call below performs Import of _QuickStart_ database using a ZIP file
available on the server.
<SyntaxHighlighter wrapLines={true} language="shell" style={theme.dark}>
{`curl -X POST -u adminuser:password -H "Cache-Control: no-cache" -d
./importOptions.json
diff --git a/docs/src/documents/Import-Export/ImportAPIOptions.md
b/docs/src/documents/Import-Export/ImportAPIOptions.md
index dab0031..1f6c8e3 100644
--- a/docs/src/documents/Import-Export/ImportAPIOptions.md
+++ b/docs/src/documents/Import-Export/ImportAPIOptions.md
@@ -11,7 +11,7 @@ import SyntaxHighlighter from 'react-syntax-highlighter';
# Import API Options
-Import API options are specified as _options_ JSON. Since the API accepts
multi-part form data, it is possible to sepecify multipls input streams within
the CURL call.
+Import API options are specified as _options_ JSON. Since the API accepts
multi-part form data, it is possible to specify multiple input streams within
the CURL call.
### Examples Using CURL Calls
<SyntaxHighlighter wrapLines={true} language="shell" style={theme.dark}>
@@ -35,7 +35,7 @@ To use the defaults, set the contents of _importOptions.json_
to:
### Options
Following options are supported for Import process:
- * Specify transforms during import operation.
+ * Specify transforms during the import operation.
* Resume import by specifying starting entity guid.
* Optionally import type definition.
* Handling large imports.
@@ -52,7 +52,7 @@ Right now these are the transforms that can be applied:
Example:
-The example below applies couple of transforms to the the _qualifiedName_
attribute of hive_table. It converts the value to lower case, then searches for
'cl1', if found, replaces it with 'cl2'.
+The example below applies a couple of transforms to the _qualifiedName_
attribute of hive_table. It converts the value to lower case, then searches for
'cl1', if found, replaces it with 'cl2'.
To use the option, set the contents of _importOptions.json_ to:
@@ -70,7 +70,7 @@ Please refer to
[ATLAS-1825](https://issues.apache.org/jira/browse/ATLAS-1825) f
When an import operation is in progress and the server goes down, it would be
possible to resume import from the last successfully imported entity. This
would allow the import to resume from where it left off.
-Server-side logging is improved to display the detail of the last successfully
imported entity, this includes the index within the import list and the
entity's guid. Either can be used specify the point to resume import.
+Server-side logging is improved to display the detail of the last successfully
imported entity, this includes the index within the import list and the
entity's guid. Either can be used to specify the point to resume import.
To use the option, set the contents of _importOptions.json_ to:
@@ -103,19 +103,19 @@ Steps to use the behavior:
The output of Export has _atlas-typedef.json_ that contains the type
definitions for the entities exported.
-By default (that is if no options is specified), the type definitions are
imported and applied to the system being imported to. The entity import is
performed after this.
+By default (that is if no options are specified), the type definitions are
imported and applied to the system being imported to. The entity import is
performed after this.
-In some cases, you would not want to modify the type definitions. Import may
be better off failing than the types be modified.
+In some cases, you would not want to modify the type definitions. The import
may be better off failing than the types be modified.
-This option allows for optionally importing of type definition. The option is
set to _true_ by default, which means that type definition is imported. With
this option set to _false_, type definitions preseneraent in the source will
not be imported. In case of mismatch between the entities being imported the
types present in the system where the import is being performed, the operation
will fail.
+This option allows for optionally importing of the type definition. The option
is set to _true_ by default, which means that type definition is imported. With
this option set to _false_, type definitions present in the source will not be
imported. In case of mismatch between the entities being imported the types
present in the system where the import is being performed, the operation will
fail.
Table below enumerates the conditions that get addressed as part of type
definition import:
|**Condition**|**Action**|
|-------------|----------|
-| Incoming type does not exist in target system | Type is created. |
-|Type to be imported and type in target system are same | No change |
-|Type to be imported and type in target system differ by some attributes|
Target system type is updated to the attributes present in the source.<br /> It
is possible that the target system will have attributes in addition to the one
present in the source.<br /> In that case, the target system's type attributes
will be an union of the attributes.<br /> Attributes in target system will not
be deleted to match the source. <br />If the type of the attribute differ,
import process will be abo [...]
+| Incoming type does not exist in the target system | Type is created. |
+|Type to be imported and type in the target system are same | No change |
+|Type to be imported and type in target system differ by some attributes|
Target system type is updated to the attributes present in the source.<br /> It
is possible that the target system will have attributes in addition to the one
present in the source.<br /> In that case, the target system's type attributes
will be a union of the attributes.<br /> Attributes in target system will not
be deleted to match the source. <br />If the type of the attribute differ, the
import process will be [...]
To use the option, set the contents of _importOptions.json_ to:
@@ -129,7 +129,7 @@ To use the option, set the contents of _importOptions.json_
to:
#### Specifying File to be Imported From Server Location
-In scenario where the file to be imported is present at a location on the
server, the _importfile_ API can be used. It behaves like the Import API.
+In a scenario where the file to be imported is present at a location on the
server, the _importfile_ API can be used. It behaves like the Import API.
To use the option, set the contents of _importOptions.json_ to:
@@ -152,10 +152,10 @@ _CURL_
#### Handling Large Imports
-By default, the Import Service stores all of the data in memory. This may be
limiting for ZIPs containing large amount of data.
+By default, the Import Service stores all of the data in memory. This may be
limiting for ZIPs containing a large amount of data.
-To configure temporary directory use the application property
_atlas.import.temp.directory_. If this property is left blank, default
in-memory implementation is used.
+To configure the temporary directory use the application property
_atlas.import.temp.directory_. If this property is left blank, the default
in-memory implementation is used.
Please ensure that there is sufficient disk space available for the operation.
-The contents of the directory created as backing store for the import
operation will be erased after the operation is over.
+The contents of the directory created as a backing store for the import
operation will be erased after the operation is over.
diff --git a/docs/src/documents/Import-Export/ImportEntityTransforms.md
b/docs/src/documents/Import-Export/ImportEntityTransforms.md
index 1a9b895..f1676b7 100644
--- a/docs/src/documents/Import-Export/ImportEntityTransforms.md
+++ b/docs/src/documents/Import-Export/ImportEntityTransforms.md
@@ -19,7 +19,7 @@ During Import Process, entity transforms are required to make
changes to the ent
A transformation framework allows a mechanism to selectively transform an
entity or specific attributes of that entity.
-To achieve this, the framework, provides:
+To achieve this, the framework provides:
* Way to set a condition that needs to be satisfied for a transformation to be
applied.
* Action to be taken on the entity once the condition is met.
@@ -28,14 +28,14 @@ The existing transformation frameworks allowed this to
happen.
#### Reason for New Transformation Framework
-While the existing framework provided the basic benefits of transformation
framework, it did not have support for some of the commonly used Atlas types.
Which meant that users of this framework would have to meticulously define
transformations for every type they are working with. This can be tedious and
potentially error prone.
+While the existing framework provided the basic benefits of the transformation
framework, it did not have support for some of the commonly used Atlas types.
Which meant that users of this framework would have to meticulously define
transformations for every type they are working with. This can be tedious and
potentially error-prone.
The new framework addresses this problem by providing built-in transformations
for some of the commonly used types. It can also be extended to accommodate new
types.
#### Approach
The approach used by the new transformation framework creates a transformation
by:
* Specifying a condition.
-* Specifying action(s) to be taken if condition is met.
+* Specifying action(s) to be taken if the condition is met.
##### Conditions
@@ -55,12 +55,12 @@ HAS_VALUE | Entity attribute has value. |
##### Actions
|**Action Type** | *Description** |
--------------------|----------------------------------------------|
-ADD_CLASSIFICATION | Add classifiction |
-REPLACE_PREFIX | Replace value starting with another value. |
-TO_LOWER | Convert value of an attribute to lower case. |
-SET | Set the value of an attribute |
-CLEAR | Clear value of an attribute |
+-------------------|------------------------------------------------|
+ADD_CLASSIFICATION | Add classification |
+REPLACE_PREFIX | Replace value starting with another value. |
+TO_LOWER | Convert a value of an attribute to lower case. |
+SET | Set the value of an attribute |
+CLEAR | Clear value of an attribute |
#### Built-in Transforms
@@ -79,7 +79,7 @@ During import, hive_db entity whose _qualifiedName_ is
_stocks@cl1_ will get the
}`}
</SyntaxHighlighter>
-Every imported entity will get the classification by simply changing the
condition. The __entity is special condition which matches entity.
+Every imported entity will get the classification by simply changing the
condition. The __entity is a special condition which matches entity.
<SyntaxHighlighter wrapLines={true} language="json" style={theme.dark}>
@@ -93,7 +93,7 @@ Every imported entity will get the classification by simply
changing the conditi
}`}
</SyntaxHighlighter>
-To add classification to only the top-level entity (entity that is used as
starting point for an export), use:
+To add the classification to only the top-level entity (an entity that is used
as a starting point for an export), use:
<SyntaxHighlighter wrapLines={true} language="json" style={theme.dark}>
{`{
diff --git a/docs/src/documents/Import-Export/ImportExportAPI.md
b/docs/src/documents/Import-Export/ImportExportAPI.md
index 8cea230..c903c17 100644
--- a/docs/src/documents/Import-Export/ImportExportAPI.md
+++ b/docs/src/documents/Import-Export/ImportExportAPI.md
@@ -22,16 +22,16 @@ The release of 0.8.3 includes the following improvements to
Export and Import AP
* New [ExportImportAudits](#/ExportImportAudits) for Export & Import
operations.
### Background
-The Import-Export APIs for Atlas facilitate transfer of data to and from a
cluster that has Atlas provisioned.
+The Import-Export APIs for Atlas facilitate the transfer of data to and from a
cluster that has Atlas provisioned.
The APIs when integrated with backup and/or disaster recovery process will
ensure participation of Atlas.
### Introduction
-There are 2 broad categories viz. Export & Import. The details of the APIs are
discussed below.
+There are 2 broad categories viz. Export & Import. The details of the APIs are
as discussed below.
The APIs are available only to _admin_ user.
-Only a single import or export operation can be performed at a given time. The
operations have a potential for generating large amount. They can also put
pressure on resources. This restriction tries to alleviate this problem.
+Only a single import or export operation can be performed at a given time. The
operations have the potential for generating large amount of data. They can
also put pressure on resources. This restriction tries to alleviate this
problem.
For Import-Export APIs relating to HDFS path, can be found
[here](#/ExportHDFSAPI).
diff --git a/docs/src/documents/Import-Export/IncrementalExport.md
b/docs/src/documents/Import-Export/IncrementalExport.md
index ed60f52..5b28905 100644
--- a/docs/src/documents/Import-Export/IncrementalExport.md
+++ b/docs/src/documents/Import-Export/IncrementalExport.md
@@ -31,7 +31,7 @@ New _fetchType_ added to indicate incremental export. This
option can be used wi
#### Getting Change Marker
-The very first call to export with _fetchType_ set to _incremental_ should be
made with _changeMarker_ set to 0. This will perform a full export. The
_AtlasExportResult_ will have the _changeMarker_ set to a value. This is the
value that should be used for subsequent call to Export.
+The very first call to export with _fetchType_ set to _incremental_ should be
made with _changeMarker_ set to 0. This will perform a full export. The
_AtlasExportResult_ will have the _changeMarker_ set to a value. This is the
value that should be used for a subsequent call to Export.
#### Skip Lineage Option
@@ -39,5 +39,5 @@ Export can be performed by skipping lineage information. This
avoids all lineage
#### Benefit of Incremental Export
-The real benefit of incremental export comes when export is done with
_skipLineage_ option set to _true_. This greatly improves performance when
fetching entities that have changed since the last export.
+The real benefit of incremental export comes when the export is done with
_skipLineage_ option set to _true_. This greatly improves performance when
fetching entities that have changed since the last export.