Hi Rajnil, Just curious, what version of spark-bigquery-connector are you using?
Thanks, Atheeth On Sat, 25 Feb 2023 at 23:48, Mich Talebzadeh <mich.talebza...@gmail.com> wrote: > sounds like it is cosmetric. The important point is that if the data > stored in GBQ is valid? > > > THT > > > view my Linkedin profile > <https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/> > > > https://en.everybodywiki.com/Mich_Talebzadeh > > > > *Disclaimer:* Use it at your own risk. Any and all responsibility for any > loss, damage or destruction of data or any other property which may arise > from relying on this email's technical content is explicitly disclaimed. > The author will in no case be liable for any monetary damages arising from > such loss, damage or destruction. > > > > > On Sat, 25 Feb 2023 at 18:12, Rajnil Guha <rajnil94.g...@gmail.com> wrote: > >> Hi All, >> >> I had created an issue on Stackoverflow(linked below) a few months back >> about issues while handling bignumeric type values of BigQuery in Spark. >> >> link >> <https://stackoverflow.com/questions/74719503/getting-error-while-reading-bignumeric-data-type-from-a-bigquery-table-using-apa> >> >> On Fri, Feb 24, 2023 at 3:54 PM Mich Talebzadeh < >> mich.talebza...@gmail.com> wrote: >> >>> Hi Nidhi, >>> >>> can you create a BigQuery table with a bignumeric and numeric column >>> types, add a few lines and try to read into spark. through DF >>> >>> and do >>> >>> >>> df.printSchema() >>> >>> df.show(5,False) >>> >>> >>> HTH >>> >>> >>> view my Linkedin profile >>> <https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/> >>> >>> >>> https://en.everybodywiki.com/Mich_Talebzadeh >>> >>> >>> >>> *Disclaimer:* Use it at your own risk. Any and all responsibility for >>> any loss, damage or destruction of data or any other property which may >>> arise from relying on this email's technical content is explicitly >>> disclaimed. The author will in no case be liable for any monetary damages >>> arising from such loss, damage or destruction. >>> >>> >>> >>> >>> On Fri, 24 Feb 2023 at 02:47, nidhi kher <kherni...@gmail.com> wrote: >>> >>>> Hello, >>>> >>>> >>>> I am facing below issue in pyspark code: >>>> >>>> We are running spark code using dataproc serverless batch in google >>>> cloud platform. Spark code is causing issue while writing the data to >>>> bigquery table. In bigquery table , few of the columns have datatype as >>>> bignumeric and spark code is changing the datatype from bignumeric to >>>> numeric while writing the data. We need datatype to be kept as bignumeric >>>> only as we need data of 38,20 precision. >>>> >>>> >>>> Can we cast a column to bignumeric in spark sql dataframe like below >>>> code for decimal: >>>> >>>> >>>> df= spark.sql("""SELECT cast(col1 as decimal(38,20)) as col1 from >>>> table1""") >>>> >>>> Spark version :3.3 >>>> >>>> Pyspark version : 1.1 >>>> >>>> >>>> Regards, >>>> >>>> Nidhi >>>> >>>