This is an automated email from the ASF dual-hosted git repository.
weibin pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/incubator-graphar.git
The following commit(s) were added to refs/heads/main by this push:
new f3a6f86 [Minor][Doc] Fix issues in the documentation (#450)
f3a6f86 is described below
commit f3a6f868cc4e26192699d698e62ea8070d41d6a9
Author: lixueclaire <[email protected]>
AuthorDate: Fri Apr 12 10:23:51 2024 +0800
[Minor][Doc] Fix issues in the documentation (#450)
---
docs/libraries/cpp/examples/snap-to-graphar.md | 2 +-
docs/libraries/cpp/getting-started.md | 4 ++--
docs/libraries/java/java.md | 6 +++---
docs/libraries/pyspark/how-to.md | 8 ++------
docs/libraries/spark/examples.md | 18 ++++++++++--------
docs/libraries/spark/spark.md | 20 ++++++++++----------
docs/specification/format.md | 2 +-
7 files changed, 29 insertions(+), 31 deletions(-)
diff --git a/docs/libraries/cpp/examples/snap-to-graphar.md
b/docs/libraries/cpp/examples/snap-to-graphar.md
index 91ed659..6e09273 100644
--- a/docs/libraries/cpp/examples/snap-to-graphar.md
+++ b/docs/libraries/cpp/examples/snap-to-graphar.md
@@ -87,5 +87,5 @@ e_builder->Clear();
For comprehensive insights into this example, please consult the
accompanying [source
-code](https://github.com/apache/incubator-graphar/tree/main/docs/cpp/examples/snap_dataset_to_graphar.cc)
+code](https://github.com/apache/incubator-graphar/tree/main/cpp/examples/snap_dataset_to_graphar.cc)
.
diff --git a/docs/libraries/cpp/getting-started.md
b/docs/libraries/cpp/getting-started.md
index f7dc186..5f741fb 100644
--- a/docs/libraries/cpp/getting-started.md
+++ b/docs/libraries/cpp/getting-started.md
@@ -9,7 +9,7 @@ sidebar_position: 1
This article is a quick guide that explains how to work with GraphAr
C++. To begin with, please refer to the [Building
-Steps](https://github.com/apache/incubator-graphar/blob/main/README.rst#building-libraries)
+Steps](https://github.com/apache/incubator-graphar/blob/main/README.md#building-libraries)
to install GraphAr.
## GAR Information Files
@@ -79,7 +79,7 @@ file type, as well as all the property groups attached.
It is allowed to store different types of adjLists for a group of
edges at the same time.
-::
+:::
## GAR Data Files
diff --git a/docs/libraries/java/java.md b/docs/libraries/java/java.md
index 66967a5..2276922 100644
--- a/docs/libraries/java/java.md
+++ b/docs/libraries/java/java.md
@@ -120,7 +120,7 @@ if (!graphInfoResult.hasError()) {
```
See [test for
-graphinfo](https://github.com/apache/incubator-graphar/tree/main/java/src/test/java/com/apache/incubator-graphar/graphinfo)
+graphinfo](https://github.com/apache/incubator-graphar/tree/main/java/src/test/java/org/apache/graphar/graphinfo)
for the complete example.
### Writers
@@ -177,7 +177,7 @@ writer.sortAndWriteAdjListTable(table, 0, 0); // Write adj
list of vertex chunk
```
See [test for
-writers](https://github.com/apache/incubator-graphar/tree/main/java/src/test/java/com/apache/incubator-graphar/writers)
+writers](https://github.com/apache/incubator-graphar/tree/main/java/src/test/java/org/apache/graphar/writers)
for the complete example.
### Readers
@@ -211,5 +211,5 @@ StdPair<Long, Long> range = reader.getRange().value();
```
See [test for
-readers](https://github.com/apache/incubator-graphar/tree/main/java/src/test/java/com/apache/incubator-graphar/readers)
+readers](https://github.com/apache/incubator-graphar/tree/main/java/src/test/java/org/apache/graphar/readers)
for the complete example.
\ No newline at end of file
diff --git a/docs/libraries/pyspark/how-to.md b/docs/libraries/pyspark/how-to.md
index 2df0764..aebeb32 100644
--- a/docs/libraries/pyspark/how-to.md
+++ b/docs/libraries/pyspark/how-to.md
@@ -27,10 +27,6 @@ spark = (
)
```
-
- .. rubric:: GraphAr PySpark initialize
- :name: graphar-pyspark-initialize
-
## GraphAr PySpark initialize
PySpark bindings are heavily relying on JVM-calls via ``py4j``. To
@@ -70,9 +66,9 @@ for the more detailed information.
GraphAr PySpark package provide two main ways how to initiate
objects, like ``GraphInfo``:
-#. ``from_python(**args)`` when you create an object based on
+- ``from_python(**args)`` when you create an object based on
python-arguments
-#. ``from_scala(jvm_ref)`` when you create an object from the
+- ``from_scala(jvm_ref)`` when you create an object from the
corresponded JVM-object (``py4j.java_gateway.JavaObject``)
diff --git a/docs/libraries/spark/examples.md b/docs/libraries/spark/examples.md
index ced535e..a6d6873 100644
--- a/docs/libraries/spark/examples.md
+++ b/docs/libraries/spark/examples.md
@@ -199,13 +199,15 @@ Finally, you will see the graph in Neo4j Browser after
running the above code.
See [GraphAr2Neo4j.scala][graphar2neo4j] for the complete example.
-> **tip:**
-> - The Neo4j Spark Connector offers different save modes and writing options,
such as Append(CREATE) or Overwrite(MERGE). Please refer to its
[documentation](https://neo4j.com/docs/spark/current/writing/) for more
information and take the most appropriate method while using.
-> - The Neo4j Spark Connector supports to use [Spark structured streaming
API](https://neo4j.com/docs/spark/current/streaming), which works differently
from Spark batching. One can utilize this API to read/write a stream from/to
Neo4j, avoiding to maintain all data in the memory.
+:::tip
+- The Neo4j Spark Connector offers different save modes and writing options,
such as Append(CREATE) or Overwrite(MERGE). Please refer to its
[documentation](https://neo4j.com/docs/spark/current/writing/) for more
information and take the most appropriate method while using.
+- The Neo4j Spark Connector supports to use [Spark structured streaming
API](https://neo4j.com/docs/spark/current/streaming), which works differently
from Spark batching. One can utilize this API to read/write a stream from/to
Neo4j, avoiding to maintain all data in the memory.
-[test-graph-transformer]:
https://github.com/apache/incubator-graphar/blob/main/spark/src/test/scala/com/apache/incubator-graphar/TestGraphTransformer.scala
-[transformer-example]:
https://github.com/apache/incubator-graphar/blob/main/spark/src/test/scala/com/apache/incubator-graphar/TransformExample.scala
-[compute-example]:
https://github.com/apache/incubator-graphar/blob/main/spark/src/test/scala/com/apache/incubator-graphar/ComputeExample.scala
-[neo4j2graphar]:
https://github.com/apache/incubator-graphar/blob/main/spark/src/main/scala/com/apache/incubator-graphar/example/Neo4j2GraphAr.scala
-[graphar2neo4j]:
https://github.com/apache/incubator-graphar/blob/main/spark/src/main/scala/com/apache/incubator-graphar/example/GraphAr2Neo4j.scala
\ No newline at end of file
+:::
+
+[test-graph-transformer]:
https://github.com/apache/incubator-graphar/blob/main/spark/graphar/src/test/scala/org/apache/graphar/TestGraphTransformer.scala
+[transformer-example]:
https://github.com/apache/incubator-graphar/blob/main/spark/graphar/src/test/scala/org/apache/graphar/TransformExample.scala
+[compute-example]:
https://github.com/apache/incubator-graphar/blob/main/spark/graphar/src/test/scala/org/apache/graphar/ComputeExample.scala
+[neo4j2graphar]:
https://github.com/apache/incubator-graphar/blob/main/spark/graphar/src/main/scala/org/apache/graphar/example/Neo4j2GraphAr.scala
+[graphar2neo4j]:
https://github.com/apache/incubator-graphar/blob/main/spark/graphar/src/main/scala/org/apache/graphar/example/GraphAr2Neo4j.scala
\ No newline at end of file
diff --git a/docs/libraries/spark/spark.md b/docs/libraries/spark/spark.md
index b313584..b48b96a 100644
--- a/docs/libraries/spark/spark.md
+++ b/docs/libraries/spark/spark.md
@@ -213,7 +213,7 @@ There are certain limitations while using the Graph
Transformer:
- The vertices (or edges) of the source and destination graphs are aligned
by labels, meaning each vertex/edge label included in the destination graph
must have an equivalent in the source graph, in order for the related chunks to
be loaded as the data source.
- For each group of vertices/edges (i.e., each single label), each property
included in the destination graph (defined in the relevant VertexInfo/EdgeInfo)
must also be present in the source graph.
- In addition, users can use the GraphAr Spark Reader/Writer to conduct data
transformation more flexibly at the vertex/edge table level, as opposed to the
graph level. This allows for a more granular approach to transforming data, as
`TransformExample.scala`_ shows.
+ In addition, users can use the GraphAr Spark Reader/Writer to conduct data
transformation more flexibly at the vertex/edge table level, as opposed to the
graph level. This allows for a more granular approach to transforming data, as
[TransformExample.scala][transform-example] shows.
:::
@@ -226,12 +226,12 @@ For more information on usage, please refer to the
examples:
- [Neo4j2GraphAr.scala][neo4j2graphar] and
[GraphAr2Neo4j.scala][graphar2neo4j] are examples to conduct data
importing/exporting for Neo4j.
-[test-graph-info]:
https://github.com/apache/incubator-graphar/blob/main/spark/src/test/scala/com/apache/incubator-graphar/TestGraphInfo.scala
-[test-index-generator]:
https://github.com/apache/incubator-graphar/blob/main/spark/src/test/scala/com/apache/incubator-graphar/TestIndexGenerator.scala
-[test-writer]:
https://github.com/apache/incubator-graphar/blob/main/spark/src/test/scala/com/apache/incubator-graphar/TestWriter.scala
-[test-reader]:
https://github.com/apache/incubator-graphar/blob/main/spark/src/test/scala/com/apache/incubator-graphar/TestReader.scala
-[test-graph-transformer]:
https://github.com/apache/incubator-graphar/blob/main/spark/src/test/scala/com/apache/incubator-graphar/TestGraphTransformer.scala
-[compute-example]:
https://github.com/apache/incubator-graphar/blob/main/spark/src/test/scala/com/apache/incubator-graphar/ComputeExample.scala
-[transform-example]:
https://github.com/apache/incubator-graphar/blob/main/spark/src/test/scala/com/apache/incubator-graphar/TransformExample.scala
-[neo4j2graphar]:
https://github.com/apache/incubator-graphar/blob/main/spark/src/main/scala/com/apache/incubator-graphar/example/Neo4j2GraphAr.scala
-[graphar2neo4j]:
https://github.com/apache/incubator-graphar/blob/main/spark/src/main/scala/com/apache/incubator-graphar/example/GraphAr2Neo4j.scala
\ No newline at end of file
+[test-graph-info]:
https://github.com/apache/incubator-graphar/blob/main/spark/graphar/src/test/scala/org/apache/graphar/TestGraphInfo.scala
+[test-index-generator]:
https://github.com/apache/incubator-graphar/blob/main/spark/graphar/src/test/scala/org/apache/graphar/TestIndexGenerator.scala
+[test-writer]:
https://github.com/apache/incubator-graphar/blob/main/spark/graphar/src/test/scala/org/apache/graphar/TestWriter.scala
+[test-reader]:
https://github.com/apache/incubator-graphar/blob/main/spark/graphar/src/test/scala/org/apache/graphar/TestReader.scala
+[test-graph-transformer]:
https://github.com/apache/incubator-graphar/blob/main/spark/graphar/src/test/scala/org/apache/graphar/TestGraphTransformer.scala
+[compute-example]:
https://github.com/apache/incubator-graphar/blob/main/spark/graphar/src/test/scala/org/apache/graphar/ComputeExample.scala
+[transform-example]:
https://github.com/apache/incubator-graphar/blob/main/spark/graphar/src/test/scala/org/apache/graphar/TransformExample.scala
+[neo4j2graphar]:
https://github.com/apache/incubator-graphar/blob/main/spark/graphar/src/main/scala/org/apache/graphar/example/Neo4j2GraphAr.scala
+[graphar2neo4j]:
https://github.com/apache/incubator-graphar/blob/main/spark/graphar/src/main/scala/org/apache/graphar/example/GraphAr2Neo4j.scala
\ No newline at end of file
diff --git a/docs/specification/format.md b/docs/specification/format.md
index e482cde..110cca8 100644
--- a/docs/specification/format.md
+++ b/docs/specification/format.md
@@ -110,7 +110,7 @@ Take the "person" vertex table as an example, if the chunk
size is set to be 500
<img src="../images/vertex_physical_table.png" alt="vertex physical table"
width="700" align="center"/>
-::: note
+:::note
For efficiently utilize the filter push-down of the payload file format like
Parquet, the internal vertex id is stored in the payload file as a column. And
since the internal vertex id is continuous, the payload file format can use the
delta encoding for the internal vertex id column, which would not bring too
much overhead for the storage.
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]