Hi, I'm* importing a few CSV*s with spark-csv package, Always when I give a select at each one looks ok But when i join then with sqlContext.sql give me this error
all tables has fields timestamp joins are not with this dates *Py4JJavaError: An error occurred while calling o643.showString.* : org.apache.spark.SparkException: Job aborted due to stage failure: Task 54 in stage 92.0 failed 10 times, most recent failure: Lost task 54.9 in stage 92.0 (TID 6356, yp-spark-dal09-env5-0036): org.apache.spark.api.python.PythonException: Traceback (most recent call last): File "/usr/local/src/spark160master/spark-1.6.0-bin-2.6.0/python/lib/pyspark.zip/pyspark/worker.py", line 111, in main process() File "/usr/local/src/spark160master/spark-1.6.0-bin-2.6.0/python/lib/pyspark.zip/pyspark/worker.py", line 106, in process serializer.dump_stream(func(split_index, iterator), outfile) File "/usr/local/src/spark160master/spark-1.6.0-bin-2.6.0/python/lib/pyspark.zip/pyspark/serializers.py", line 263, in dump_stream vs = list(itertools.islice(iterator, batch)) File "/usr/local/src/spark160master/spark/python/pyspark/sql/functions.py", line 1563, in <lambda> func = lambda _, it: map(lambda x: returnType.toInternal(f(*x)), it) File "/usr/local/src/spark160master/spark-1.6.0-bin-2.6.0/python/lib/pyspark.zip/pyspark/sql/types.py", line 191, in toInternal else time.mktime(dt.timetuple())) *ValueError: year out of range * Any one knows this problem? Best, *Daniel Lopes* Chief Data and Analytics Officer | OneMatch c: +55 (18) 99764-2733 | https://www.linkedin.com/in/dslopes www.onematch.com.br <http://www.onematch.com.br/?utm_source=EmailSignature&utm_term=daniel-lopes>