Hi All,
1) in spark created a carbondata table
    create table temp(col1 string, col2 string) STORED BY
'org.apache.carbondata.format';
2) insert into table once
  insert into temp values ('test', 'test'), ('test2', 'test2');
3) drop table temp;

in this is above steps which flow will create below lock files
clean_files.lock
compaction.lock
droptable.lock
meta.lock
update.lock

when I repeat  above steps rarely  some files are not deleted, in hdfs
after drop table ;
<size 0 bytes> clean_files.lock
<size 0 bytes> compaction.lock
<size 0 bytes> droptable.lock
<size 0 bytes> meta.lock
<size 0 bytes> update.lock
                         metadata
<size 190 bytes> metadata/schema

thanks,
sandeep

Reply via email to