[ 
https://issues.apache.org/jira/browse/SDAP-317?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17373294#comment-17373294
 ] 

Antoine Queric commented on SDAP-317:
-------------------------------------

Dear [~tloubrieu] , [~skperez] , [~thuang], [~nchung] (others ?)

I think this may be an interesting conversation before deciding whether we 
should or should not try to implement such a feature.

 

Our aim is to be able to concatenate multiple time ranges into one tile (also 
geo-spatially sliced) from daily netcdf files (only one TIME step inside). We 
are interested into trying that because we thought that may enhance long 
timeseries queries on the nexus webapp (yet to be proved).

Below is the directory structure of the dataset we are working with :

```

2015
├── 001
│   └── 20150101-IFR-L4_GHRSST-SSTfnd-ODYSSEA-GLOB_010-v2.0-fv1.0.nc
├── 002
│   └── 20150102-IFR-L4_GHRSST-SSTfnd-ODYSSEA-GLOB_010-v2.0-fv1.0.nc
├── 003
│   └── 20150103-IFR-L4_GHRSST-SSTfnd-ODYSSEA-GLOB_010-v2.0-fv1.0.nc
├── 004
│   └── 20150104-IFR-L4_GHRSST-SSTfnd-ODYSSEA-GLOB_010-v2.0-fv1.0.nc
├── 005
│   └── 20150105-IFR-L4_GHRSST-SSTfnd-ODYSSEA-GLOB_010-v2.0-fv1.0.nc
├── 006
│   └── 20150106-IFR-L4_GHRSST-SSTfnd-ODYSSEA-GLOB_010-v2.0-fv1.0.nc
├── 007
│   └── 20150107-IFR-L4_GHRSST-SSTfnd-ODYSSEA-GLOB_010-v2.0-fv1.0.nc
├── 008
│   └── 20150108-IFR-L4_GHRSST-SSTfnd-ODYSSEA-GLOB_010-v2.0-fv1.0.nc
├── 009
│   └── 20150109-IFR-L4_GHRSST-SSTfnd-ODYSSEA-GLOB_010-v2.0-fv1.0.nc

```

Headers chunk for the dimensions size :

```

netcdf \20150101-IFR-L4_GHRSST-SSTfnd-ODYSSEA-GLOB_010-v2.0-fv1.0 {
dimensions:
 lat = 1600 ;
 lon = 3600 ;
 time = 1 ;

```

First of all, is this something we really want to add in SDAP ingester ? I'm 
not sure re-processing the netcdf files in order to re-chunk on time dimension 
is something we will do in some cases, so my first guess is that we would want 
granule_ingester to do that for us.

 

If we wanted this, how to handle the collection_manager for it to properly 
queue a specific number of contiguous files that one granule_ingester intsance 
will take care of ?

 

Maybe the more simple way to handle that would simply be to (for this speciic 
need) bypass the collection_manager & launch a granule_ingester instance with a 
list of files & configuration to process ?

 

Maybe I'm missing something, we started evaluating the new SDAP ingester module 
this week (and already comitting code for elasticsearch support) ; maybe what 
we want to do is alreadyy possible & I missed that :)

 

Best regards,

Antoine

> Open multiple netcdf files in order to generate granules with multiple time 
> steps
> ---------------------------------------------------------------------------------
>
>                 Key: SDAP-317
>                 URL: https://issues.apache.org/jira/browse/SDAP-317
>             Project: Apache Science Data Analytics Platform
>          Issue Type: Improvement
>          Components: granule-ingester
>            Reporter: Antoine Queric
>            Priority: Major
>
> When netcdf files only include one single time step, it may be interesting to 
> open multiple files & generate a data cube which contains :
>  * longitude slice
>  * latitude slice
>  * time slice
> We will develop & test such a feature in order to compare performance when 
> querying long timeseries.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to