Hi Punit,

JSON can be hard to parse in parallel due to its nested structure. It
depends on the schema and (textual) representation of the JSON whether and
how it can be done. The problem is that a parallel input format needs to be
able to identify record boundaries without context information. This can be
very easy, if your JSON data is a list of JSON objects which are separated
by a new line character. However, this is hard to generalize. That's why
Flink does not offer tooling for it (yet).

If your JSON objects are separated by new line characters, the easiest way
is to read it as text file, where each line results in a String and parse
each object using a standard JSON parser. This would look like:

ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment();

DataSet<String> text = env.readTextFile("/path/to/jsonfile");
DataSet<YourObject> json = text.map(new YourMapFunctionWhichParsesJSON());

Best, Fabian

2016-04-26 8:06 GMT+02:00 Punit Naik <naik.puni...@gmail.com>:

> Hi
>
> I am new to Flink. I was experimenting with the Dataset API and found out
> that there is no explicit method for loading a JSON file as input. Can
> anyone please suggest me a workaround?
>
> --
> Thank You
>
> Regards
>
> Punit Naik
>

Reply via email to