Repository: spark
Updated Branches:
  refs/heads/branch-1.2 d117f8fa4 -> 42b9d0d31


[SPARK-4535][Streaming] Fix the error in comments

change `NetworkInputDStream` to `ReceiverInputDStream`
change `ReceiverInputTracker` to `ReceiverTracker`

Author: q00251598 <qiyad...@huawei.com>

Closes #3400 from watermen/fix-comments and squashes the following commits:

75d795c [q00251598] change 'NetworkInputDStream' to 'ReceiverInputDStream' && 
change 'ReceiverInputTracker' to 'ReceiverTracker'

(cherry picked from commit a51118a34a4617c07373480c4b021e53124c3c00)
Signed-off-by: Tathagata Das <tathagata.das1...@gmail.com>

Conflicts:
        
examples/src/main/scala/org/apache/spark/examples/streaming/StatefulNetworkWordCount.scala


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/42b9d0d3
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/42b9d0d3
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/42b9d0d3

Branch: refs/heads/branch-1.2
Commit: 42b9d0d31eae8d992301bcd36665d01ef1a00a06
Parents: d117f8f
Author: q00251598 <qiyad...@huawei.com>
Authored: Tue Nov 25 04:01:56 2014 -0800
Committer: Tathagata Das <tathagata.das1...@gmail.com>
Committed: Tue Nov 25 04:07:13 2014 -0800

----------------------------------------------------------------------
 .../spark/examples/streaming/StatefulNetworkWordCount.scala  | 2 +-
 .../examples/streaming/clickstream/PageViewStream.scala      | 2 +-
 .../scala/org/apache/spark/streaming/StreamingContext.scala  | 2 +-
 .../spark/streaming/api/java/JavaStreamingContext.scala      | 2 +-
 .../spark/streaming/dstream/ReceiverInputDStream.scala       | 8 ++++----
 .../apache/spark/streaming/scheduler/ReceiverTracker.scala   | 2 +-
 6 files changed, 9 insertions(+), 9 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/spark/blob/42b9d0d3/examples/src/main/scala/org/apache/spark/examples/streaming/StatefulNetworkWordCount.scala
----------------------------------------------------------------------
diff --git 
a/examples/src/main/scala/org/apache/spark/examples/streaming/StatefulNetworkWordCount.scala
 
b/examples/src/main/scala/org/apache/spark/examples/streaming/StatefulNetworkWordCount.scala
index a4d159b..26b950f 100644
--- 
a/examples/src/main/scala/org/apache/spark/examples/streaming/StatefulNetworkWordCount.scala
+++ 
b/examples/src/main/scala/org/apache/spark/examples/streaming/StatefulNetworkWordCount.scala
@@ -56,7 +56,7 @@ object StatefulNetworkWordCount {
     val ssc = new StreamingContext(sparkConf, Seconds(1))
     ssc.checkpoint(".")
 
-    // Create a NetworkInputDStream on target ip:port and count the
+    // Create a ReceiverInputDStream on target ip:port and count the
     // words in input stream of \n delimited test (eg. generated by 'nc')
     val lines = ssc.socketTextStream(args(0), args(1).toInt)
     val words = lines.flatMap(_.split(" "))

http://git-wip-us.apache.org/repos/asf/spark/blob/42b9d0d3/examples/src/main/scala/org/apache/spark/examples/streaming/clickstream/PageViewStream.scala
----------------------------------------------------------------------
diff --git 
a/examples/src/main/scala/org/apache/spark/examples/streaming/clickstream/PageViewStream.scala
 
b/examples/src/main/scala/org/apache/spark/examples/streaming/clickstream/PageViewStream.scala
index d9b886e..55226c0 100644
--- 
a/examples/src/main/scala/org/apache/spark/examples/streaming/clickstream/PageViewStream.scala
+++ 
b/examples/src/main/scala/org/apache/spark/examples/streaming/clickstream/PageViewStream.scala
@@ -50,7 +50,7 @@ object PageViewStream {
     val ssc = new StreamingContext("local[2]", "PageViewStream", Seconds(1),
       System.getenv("SPARK_HOME"), 
StreamingContext.jarOfClass(this.getClass).toSeq)
 
-    // Create a NetworkInputDStream on target host:port and convert each line 
to a PageView
+    // Create a ReceiverInputDStream on target host:port and convert each line 
to a PageView
     val pageViews = ssc.socketTextStream(host, port)
                        .flatMap(_.split("\n"))
                        .map(PageView.fromString(_))

http://git-wip-us.apache.org/repos/asf/spark/blob/42b9d0d3/streaming/src/main/scala/org/apache/spark/streaming/StreamingContext.scala
----------------------------------------------------------------------
diff --git 
a/streaming/src/main/scala/org/apache/spark/streaming/StreamingContext.scala 
b/streaming/src/main/scala/org/apache/spark/streaming/StreamingContext.scala
index 54b2197..ec59221 100644
--- a/streaming/src/main/scala/org/apache/spark/streaming/StreamingContext.scala
+++ b/streaming/src/main/scala/org/apache/spark/streaming/StreamingContext.scala
@@ -187,7 +187,7 @@ class StreamingContext private[streaming] (
   /**
    * Set each DStreams in this context to remember RDDs it generated in the 
last given duration.
    * DStreams remember RDDs only for a limited duration of time and releases 
them for garbage
-   * collection. This method allows the developer to specify how to long to 
remember the RDDs (
+   * collection. This method allows the developer to specify how long to 
remember the RDDs (
    * if the developer wishes to query old data outside the DStream 
computation).
    * @param duration Minimum duration that each DStream should remember its 
RDDs
    */

http://git-wip-us.apache.org/repos/asf/spark/blob/42b9d0d3/streaming/src/main/scala/org/apache/spark/streaming/api/java/JavaStreamingContext.scala
----------------------------------------------------------------------
diff --git 
a/streaming/src/main/scala/org/apache/spark/streaming/api/java/JavaStreamingContext.scala
 
b/streaming/src/main/scala/org/apache/spark/streaming/api/java/JavaStreamingContext.scala
index 7db66c6..d8695b8 100644
--- 
a/streaming/src/main/scala/org/apache/spark/streaming/api/java/JavaStreamingContext.scala
+++ 
b/streaming/src/main/scala/org/apache/spark/streaming/api/java/JavaStreamingContext.scala
@@ -479,7 +479,7 @@ class JavaStreamingContext(val ssc: StreamingContext) 
extends Closeable {
   /**
    * Sets each DStreams in this context to remember RDDs it generated in the 
last given duration.
    * DStreams remember RDDs only for a limited duration of duration and 
releases them for garbage
-   * collection. This method allows the developer to specify how to long to 
remember the RDDs (
+   * collection. This method allows the developer to specify how long to 
remember the RDDs (
    * if the developer wishes to query old data outside the DStream 
computation).
    * @param duration Minimum duration that each DStream should remember its 
RDDs
    */

http://git-wip-us.apache.org/repos/asf/spark/blob/42b9d0d3/streaming/src/main/scala/org/apache/spark/streaming/dstream/ReceiverInputDStream.scala
----------------------------------------------------------------------
diff --git 
a/streaming/src/main/scala/org/apache/spark/streaming/dstream/ReceiverInputDStream.scala
 
b/streaming/src/main/scala/org/apache/spark/streaming/dstream/ReceiverInputDStream.scala
index 3e67161..c834744 100644
--- 
a/streaming/src/main/scala/org/apache/spark/streaming/dstream/ReceiverInputDStream.scala
+++ 
b/streaming/src/main/scala/org/apache/spark/streaming/dstream/ReceiverInputDStream.scala
@@ -29,7 +29,7 @@ import org.apache.spark.streaming.scheduler.ReceivedBlockInfo
 /**
  * Abstract class for defining any 
[[org.apache.spark.streaming.dstream.InputDStream]]
  * that has to start a receiver on worker nodes to receive external data.
- * Specific implementations of NetworkInputDStream must
+ * Specific implementations of ReceiverInputDStream must
  * define `the getReceiver()` function that gets the receiver object of type
  * [[org.apache.spark.streaming.receiver.Receiver]] that will be sent
  * to the workers to receive data.
@@ -39,17 +39,17 @@ import 
org.apache.spark.streaming.scheduler.ReceivedBlockInfo
 abstract class ReceiverInputDStream[T: ClassTag](@transient ssc_ : 
StreamingContext)
   extends InputDStream[T](ssc_) {
 
-  /** This is an unique identifier for the network input stream. */
+  /** This is an unique identifier for the receiver input stream. */
   val id = ssc.getNewReceiverStreamId()
 
   /**
    * Gets the receiver object that will be sent to the worker nodes
    * to receive data. This method needs to defined by any specific 
implementation
-   * of a NetworkInputDStream.
+   * of a ReceiverInputDStream.
    */
   def getReceiver(): Receiver[T]
 
-  // Nothing to start or stop as both taken care of by the 
ReceiverInputTracker.
+  // Nothing to start or stop as both taken care of by the ReceiverTracker.
   def start() {}
 
   def stop() {}

http://git-wip-us.apache.org/repos/asf/spark/blob/42b9d0d3/streaming/src/main/scala/org/apache/spark/streaming/scheduler/ReceiverTracker.scala
----------------------------------------------------------------------
diff --git 
a/streaming/src/main/scala/org/apache/spark/streaming/scheduler/ReceiverTracker.scala
 
b/streaming/src/main/scala/org/apache/spark/streaming/scheduler/ReceiverTracker.scala
index 1c3984d..32e481d 100644
--- 
a/streaming/src/main/scala/org/apache/spark/streaming/scheduler/ReceiverTracker.scala
+++ 
b/streaming/src/main/scala/org/apache/spark/streaming/scheduler/ReceiverTracker.scala
@@ -46,7 +46,7 @@ private[streaming] case class DeregisterReceiver(streamId: 
Int, msg: String, err
   extends ReceiverTrackerMessage
 
 /**
- * This class manages the execution of the receivers of NetworkInputDStreams. 
Instance of
+ * This class manages the execution of the receivers of ReceiverInputDStreams. 
Instance of
  * this class must be created after all input streams have been added and 
StreamingContext.start()
  * has been called because it needs the final set of input streams at the time 
of instantiation.
  *


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org

Reply via email to