Hey Ryan,

I was using spark merge into command  but its performance became slower as
the size of data in the iceberg table was growing.
A part of the reason may be due to I need to overwrite some of the older
partitions often as a side effect of delete and update.

Another reason may be that *MERGE INTO* command require reading all iceberg
data into spark to do join*.*



I wanted to know if there is a way by which I can acquire lock on iceberg
table.
After acquiring lock, I will  do delete, update and insert operations
explicitly and release the lock.

And If there were any issues then I will roll back to the snapshot id  that
was latest before applying my sql operations.



I require a lock on the iceberg table because no other process should write
to the iceberg table while my sql operations are in progress.

Thanks,

vivek



On 2021/05/02 18:10:19, Ryan Blue <b...@apache.org> wrote:

> Vivek,>

>

> Currently, Spark doesn't support any of the BEGIN/COMMIT statements for>

> transactions, so I don't think that it is possible right now. What are
you>

> trying to do? It may be that some of the newer commands, like MERGE
INTO,>

> would work for you instead.>

>

> On Thu, Apr 29, 2021 at 5:49 PM vivek B <vivekbalachan...@gmail.com>
wrote:>

>

> > Hey All,>

> > Is there a way to run multiple sql operations via spark as one single>

> > transaction ?>

> >>

> > Thanks,>

> > vivek>

> >>

>

>

> -- >

> Ryan Blue>

>

Reply via email to