Sean Owen commented on SPARK-17633:

The issue is more at the HDFS API level, which Spark uses to read data. It 
generally assumes data is immutable, and Spark assumes it doesn't need to check 
if underlying data changed. If it did change, when recomputed, the wrong thing 
might happen. Although I think you generally shouldn't do this, I'd be pleased 
to hear if there's a natural way to ensure that, if data is re-read, that it 
could conceivably deal with appends at least.

> texFile() and wholeTextFiles() count difference
> -----------------------------------------------
>                 Key: SPARK-17633
>                 URL: https://issues.apache.org/jira/browse/SPARK-17633
>             Project: Spark
>          Issue Type: Bug
>          Components: Input/Output
>    Affects Versions: 1.6.2
>         Environment: Unix/Linux
>            Reporter: Anshul
> sc.textFile() creates an RDD of string from a text file.
> After that when count is performed, the line count is correct, but if more 
> than one line is appended to the file manually and counting the same RDD of 
> string increments the output/result only by 1. 
> But in case of sc.wholeTextFiles() the output/result is correct.

This message was sent by Atlassian JIRA

To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to