Github user maver1ck commented on the issue: https://github.com/apache/spark/pull/19234 Any idea why we're not using `datetime.datetime.fromtimestamp(ts / 100000.)` ? There is a comment about overflow. But if it exists ?
--- --------------------------------------------------------------------- To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org