How do I convert timestamp to unix format with pyspark

11,368

You can use the built in unix_timestamp the following ways:

from pyspark.sql.functions import unix_timestamp
df = df.withColumn('unix', unix_timestamp('timestamp'))

Or

df = df.selectExpr('unix_timestamp(timestamp)')
Share:
11,368
Ticoincoin
Author by

Ticoincoin

Updated on June 05, 2022

Comments

  • Ticoincoin
    Ticoincoin almost 2 years

    I have a dataframe with timestamp values, like this one: 2018-02-15T11:39:13.000Z I want to have it in UNIX format, using Pyspark.

    I tried something like data = datasample.withColumn('timestamp_cast', datasample['timestamp'].cast('date')) but I lose a lot of information, since I only get day/month/year when I have milliseconds information in my source.

    Result: 2018-02-15

    Any idea to get unix format and keep precision? Thank you!