Repository: flink
Updated Branches:
  refs/heads/master 048cda72b -> 34a42681d


[hotfix] [docs] Update python.md

This closes #1814.


Project: http://git-wip-us.apache.org/repos/asf/flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/flink/commit/34a42681
Tree: http://git-wip-us.apache.org/repos/asf/flink/tree/34a42681
Diff: http://git-wip-us.apache.org/repos/asf/flink/diff/34a42681

Branch: refs/heads/master
Commit: 34a42681d331bcb4089b4b3467043b1e848f8931
Parents: 048cda7
Author: zentol <[email protected]>
Authored: Thu Mar 17 15:05:15 2016 +0100
Committer: Chiwan Park <[email protected]>
Committed: Tue Mar 22 14:07:33 2016 +0900

----------------------------------------------------------------------
 docs/apis/batch/python.md | 11 +++++++----
 1 file changed, 7 insertions(+), 4 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/flink/blob/34a42681/docs/apis/batch/python.md
----------------------------------------------------------------------
diff --git a/docs/apis/batch/python.md b/docs/apis/batch/python.md
index 359d2ed..646683d 100644
--- a/docs/apis/batch/python.md
+++ b/docs/apis/batch/python.md
@@ -150,7 +150,10 @@ Project setup
 
 Apart from setting up Flink, no additional work is required. The python 
package can be found in the /resource folder of your Flink distribution. The 
flink package, along with the plan and optional packages are automatically 
distributed among the cluster via HDFS when running a job.
 
-The Python API was tested on Linux systems that have Python 2.7 or 3.4 
installed.
+The Python API was tested on Linux/Windows systems that have Python 2.7 or 3.4 
installed.
+
+By default Flink will start python processes by calling "python" or "python3", 
depending on which start-script
+was used. By setting the "python.binary.python[2/3]" key in the 
flink-conf.yaml you can modify this behaviour to use a binary of your choice.
 
 {% top %}
 
@@ -216,7 +219,7 @@ data.flat_map(
       <td>
         <p>Transforms a parallel partition in a single function call. The 
function get the partition
         as an `Iterator` and can produce an arbitrary number of result values. 
The number of
-        elements in each partition depends on the degree-of-parallelism and 
previous operations.</p>
+        elements in each partition depends on the parallelism and previous 
operations.</p>
 {% highlight python %}
 data.map_partition(lambda x,c: [value * 2 for value in x])
 {% endhighlight %}
@@ -566,13 +569,13 @@ it executes. Execution environment parallelism can be 
overwritten by explicitly
 parallelism of an operator.
 
 The default parallelism of an execution environment can be specified by 
calling the
-`set_degree_of_parallelism()` method. To execute all operators, data sources, 
and data sinks of the
+`set_parallelism()` method. To execute all operators, data sources, and data 
sinks of the
 [WordCount](#example-program) example program with a parallelism of `3`, set 
the default parallelism of the
 execution environment as follows:
 
 {% highlight python %}
 env = get_environment()
-env.set_degree_of_parallelism(3)
+env.set_parallelism(3)
 
 text.flat_map(lambda x,c: x.lower().split()) \
     .group_by(1) \

Reply via email to