Hi all, I encountered weird behavior for timestamp. It seems that when using lit to add it to column, the timestamp goes from milliseconds representation to seconds representation:
scala> spark.range(1).withColumn("a", lit(new java.sql.Timestamp(1485503350000L)).cast("long")).show() +---+----------+ | id| a| +---+----------+ | 0|1485503350| +---+----------+ scala> spark.range(1).withColumn("a", lit(1485503350000L).cast(org.apache.spark.sql.types.TimestampType).cast(org.apache.spark.sql.types.LongType)).show() +---+-------------+ | id| a| +---+-------------+ | 0|1485503350000| +---+-------------+ Is this a bug or am I missing something here? Thanks, Assaf -- View this message in context: http://apache-spark-developers-list.1001551.n3.nabble.com/Possible-bug-inconsistent-timestamp-behavior-tp22144.html Sent from the Apache Spark Developers List mailing list archive at Nabble.com.