hang on are you saving this as a new table?

Dr Mich Talebzadeh



LinkedIn * 
https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
<https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*



http://talebzadehmich.wordpress.com



On 3 June 2016 at 14:13, Benjamin Kim <bbuil...@gmail.com> wrote:

> Does anyone know how to save data in a DataFrame to a table partitioned
> using an existing column reformatted into a derived column?
>
>                 val partitionedDf = df.withColumn("dt",
> concat(substring($"timestamp", 1, 10), lit(" "), substring($"timestamp",
> 12, 2), lit(":00")))
>
>                 sqlContext.setConf("hive.exec.dynamic.partition", "true")
>                 sqlContext.setConf("hive.exec.dynamic.partition.mode",
> "nonstrict")
>                 partitionedDf.write
>                     .mode(SaveMode.Append)
>                     .partitionBy("dt")
>                     .saveAsTable("ds.amo_bi_events")
>
> I am getting an ArrayOutOfBounds error. There are 83 columns in the
> destination table. But after adding the derived column, then I get an 84
> error. I assumed that the column used for the partition would not be
> counted.
>
> Can someone please help.
>
> Thanks,
> Ben
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
>
>

Reply via email to