Hi,

Have you considered batching them up into a nicely defined directory
structure and use directory pruning as part of your queries?

I ask because our batch processes does that. Data is arranged into Hour,
Day, Month, Quarter, Years structures (which we then roll-up in different
ways, based on volume (from H->*->Y)).
We then use simple directory pruning to decide what data is applicable for
each query.

Hope this helps,
 -Stefán

On Thu, Dec 8, 2016 at 5:13 PM, Alexander Reshetov <
alexander.v.reshe...@gmail.com> wrote:

> By the way, is it possible to append data to parquet data source?
> I'm looking for possibility to update (append to) existing data new
> rows so every query execution will have new data rows.
>
> Surely it's possible with plain JSON, but I want more efficient binary
> format which will give quicker reads (and executions of queries).
>
> On Wed, Dec 7, 2016 at 4:08 PM, Alexander Reshetov
> <alexander.v.reshe...@gmail.com> wrote:
> > Hello,
> >
> > I want to load batches of unstructured data in Drill. Mostly JSON data.
> >
> > Is there any batch API or other options to do so?
> >
> >
> > Thanks.
>

Reply via email to