Hi Nidhi, can you create a BigQuery table with a bignumeric and numeric column types, add a few lines and try to read into spark. through DF
and do df.printSchema() df.show(5,False) HTH view my Linkedin profile <https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/> https://en.everybodywiki.com/Mich_Talebzadeh *Disclaimer:* Use it at your own risk. Any and all responsibility for any loss, damage or destruction of data or any other property which may arise from relying on this email's technical content is explicitly disclaimed. The author will in no case be liable for any monetary damages arising from such loss, damage or destruction. On Fri, 24 Feb 2023 at 02:47, nidhi kher <kherni...@gmail.com> wrote: > Hello, > > > I am facing below issue in pyspark code: > > We are running spark code using dataproc serverless batch in google cloud > platform. Spark code is causing issue while writing the data to bigquery > table. In bigquery table , few of the columns have datatype as bignumeric > and spark code is changing the datatype from bignumeric to numeric while > writing the data. We need datatype to be kept as bignumeric only as we need > data of 38,20 precision. > > > Can we cast a column to bignumeric in spark sql dataframe like below code > for decimal: > > > df= spark.sql("""SELECT cast(col1 as decimal(38,20)) as col1 from > table1""") > > Spark version :3.3 > > Pyspark version : 1.1 > > > Regards, > > Nidhi >