Andy Skelton created HIVE-9389:
----------------------------------
Summary: INSERT OVERWRITE DIRECTORY fails to delete old data files
Key: HIVE-9389
URL: https://issues.apache.org/jira/browse/HIVE-9389
Project: Hive
Issue Type: Bug
Affects Versions: 0.13.1
Environment: CDH 5.3.0, non-secure hdfs, perm checking off
Reporter: Andy Skelton
{code:sql}
FROM myview INSERT OVERWRITE DIRECTORY 'hdfs://nameservice/path/' SELECT
COUNT(DISTINCT mycol);
{code}
This always produces one row. Sometimes the output is two files, {{000000_0}}
and {{000000_1}}, one of which is empty. Sometimes we have seen new results in
{{000000_0}} while old results remain in {{000000_1}}.
We were alerted to this because Sqoop was exporting the output files in order
by filename, writing first the new value and then overwriting with the old
value, triggering an alert because the value stopped increasing in our database.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)