Github user cloud-fan commented on a diff in the pull request:
https://github.com/apache/spark/pull/13592#discussion_r66873712
--- Diff: docs/sql-programming-guide.md ---
@@ -12,130 +12,130 @@ title: Spark SQL and DataFrames
Spark SQL is a Spark module for structured data processing. Unlike the
basic Spark RDD API, the interfaces provided
by Spark SQL provide Spark with more information about the structure of
both the data and the computation being performed. Internally,
Spark SQL uses this extra information to perform extra optimizations.
There are several ways to
-interact with Spark SQL including SQL, the DataFrames API and the Datasets
API. When computing a result
+interact with Spark SQL including SQL and the Datasets API. When computing
a result
the same execution engine is used, independent of which API/language you
are using to express the
-computation. This unification means that developers can easily switch back
and forth between the
-various APIs based on which provides the most natural way to express a
given transformation.
+computation. This unification means that developers can easily switch back
and forth between
+different APIs based on which provides the most natural way to express a
given transformation.
All of the examples on this page use sample data included in the Spark
distribution and can be run in
the `spark-shell`, `pyspark` shell, or `sparkR` shell.
## SQL
-One use of Spark SQL is to execute SQL queries written using either a
basic SQL syntax or HiveQL.
--- End diff --
why change this line?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]