Pls paste code and sample CSV
I m guessing it has to do with formatting time?
Kr

On 8 Sep 2016 12:38 am, "Daniel Lopes" <dan...@onematch.com.br> wrote:

> Hi,
>
> I'm* importing a few CSV*s with spark-csv package,
> Always when I give a select at each one looks ok
> But when i join then with sqlContext.sql give me this error
>
> all tables has fields timestamp
>
> joins are not with this dates
>
>
> *Py4JJavaError: An error occurred while calling o643.showString.*
> : org.apache.spark.SparkException: Job aborted due to stage failure: Task
> 54 in stage 92.0 failed 10 times, most recent failure: Lost task 54.9 in
> stage 92.0 (TID 6356, yp-spark-dal09-env5-0036):
> org.apache.spark.api.python.PythonException: Traceback (most recent call
> last):
>   File "/usr/local/src/spark160master/spark-1.6.0-
> bin-2.6.0/python/lib/pyspark.zip/pyspark/worker.py", line 111, in main
>     process()
>   File "/usr/local/src/spark160master/spark-1.6.0-
> bin-2.6.0/python/lib/pyspark.zip/pyspark/worker.py", line 106, in process
>     serializer.dump_stream(func(split_index, iterator), outfile)
>   File "/usr/local/src/spark160master/spark-1.6.0-
> bin-2.6.0/python/lib/pyspark.zip/pyspark/serializers.py", line 263, in
> dump_stream
>     vs = list(itertools.islice(iterator, batch))
>   File "/usr/local/src/spark160master/spark/python/pyspark/sql/functions.py",
> line 1563, in <lambda>
>     func = lambda _, it: map(lambda x: returnType.toInternal(f(*x)), it)
>   File "/usr/local/src/spark160master/spark-1.6.0-
> bin-2.6.0/python/lib/pyspark.zip/pyspark/sql/types.py", line 191, in
> toInternal
>     else time.mktime(dt.timetuple()))
> *ValueError: year out of range  *
>
> Any one knows this problem?
>
> Best,
>
> *Daniel Lopes*
> Chief Data and Analytics Officer | OneMatch
> c: +55 (18) 99764-2733 | https://www.linkedin.com/in/dslopes
>
> www.onematch.com.br
> <http://www.onematch.com.br/?utm_source=EmailSignature&utm_term=daniel-lopes>
>

Reply via email to