Hi Ankit,
While not a part of Spark, there is a project called 'WaggleDance' that can
federate multiple Hive metastores so that they are accessible via a single
URI: https://github.com/ExpediaGroup/waggle-dance
This may be useful or perhaps serve as inspiration.
Thanks,
Elliot.
On Mon, 17 Apr
Jörn, I'm interested in your point on coverage. Coverage has been a useful
tool for highlighting areas in the codebase that pose a source of potential
risk. However, generally speaking, I've found that traditional coverage
tools do not provide useful information when applied to distributed data
pro
Related to this, there exists an API in Hive to simplify the integrations
of other frameworks with Hive's ACID feature:
See:
https://cwiki.apache.org/confluence/display/Hive/HCatalog+Streaming+Mutation+API
It contains code for maintaining heartbeats, handling locks and
transactions, and submittin
, Elliot West wrote:
> Thanks for your response Jorge and apologies for my delay in replying. I
> took your advice with case 5 and declared the column names explicitly
> instead of the wildcard. This did the trick and I can now add partitions to
> an existing table. I also tried removing the
ot;xxx", 1), ("yyy", 2);
>
>
>
> hive (default)> insert into table new_record_source
> > values (3, "zzz");
>
>
>
> Regards
>
>
> On 11/01/2016, at 13:36, Elliot West wrote:
>
> Hello,
>
> I am in the process of
Hive's ACID feature (which introduces transactions) is not required for
inserts, only updates and deletes. Inserts should be supported on a vanilla
Hive shell. I'm not sure how Spark interacts with Hive in that regard but
perhaps the HiveSQLContext implementation is lacking support.
On a separate
Hello,
I am in the process of evaluating Spark (1.5.2) for a wide range of use
cases. In particular I'm keen to understand the depth of the integration
with HCatalog (aka the Hive Metastore). I am very encouraged when browsing
the source contained within the org.apache.spark.sql.hive package. My g