Hi Sowjany,

You can try and import data into delimited files on hdfs and create an
external table that can be used to insert data into the bucketed table.

That is usually a pattern I use to load data into Hive. Only use managed
tables for actual user access.

Regards,

Douglas

On Wednesday, March 28, 2018, Sowjanya Kakarala <sowja...@agrible.com>
wrote:

> Hi Guys,
>
> I am trying to import the data from my db to Hive using Sqoop.
>
> My usage is to get the data to Hive and later if necessary update the
> records from Hive. For ACID transactions I learnt that I definitely need
> bucketing tables.
>
> Now I created bucketed table and I am trying to import the data using
> Sqoop import where it is throwing a error as follows:
>
> ERROR tool.ImportTool: Encountered IOException running import job:
> org.apache.hive.hcatalog.common.HCatException : 2016 : Error operation
> not supported : *Store into a partition with bucket definition from
> Pig/Mapreduce is not supported*
>
> If it is not supported is there a work around?
>
> I created a partitioned table and import data into that table, can I
> insert these data from partitioned table to created bucketing table? If yes
> can someone provide a *best* example?
>
> I have tried with `insert into` statement for each partition which takes
> so long.
> Because my each year contains “241972735” records.
>
> Also, this might be not a good practice for production environment.
>
> Appreciate your help.
>
> Thanks
> Sowjanya
>

Reply via email to