H François,

The TableEnvironment.connect() method can only be used if you provide
(quite a bit) more code.
It requires a TableSourceFactory and handling of all the properties that
are defined in the other builder methods. See [1].

I would recommend to either register the BatchTableSource directly
(tEnv.registerTableSource()) or get a DataSet (via env.createSource()) and
register the DataSet as a Table (tEnv.registerDataSet()).

Best, Fabian

[1]
https://ci.apache.org/projects/flink/flink-docs-release-1.7/dev/table/sourceSinks.html#define-a-tablefactory


Am Mo., 11. Feb. 2019 um 21:09 Uhr schrieb françois lacombe <
francois.laco...@dcbrain.com>:

> Hi Fabian,
>
> I've got issues for a custom InputFormat implementation with my existing
> code.
>
> Is this can be used in combination with a BatchTableSource custom source?
> As I understand your solution, I should move my source to implementations
> like :
>
> tableEnvironment
>   .connect(...)
>   .withFormat(...)
>   .withSchema(...)
>   .inAppendMode()
>   .registerTableSource("MyTable")
>
> right?
>
> I currently have a BatchTableSource class which produce a DataSet<Row>
> from a single geojson file.
> This doesn't sound compatible with a custom InputFormat, don't you?
>
> Thanks in advance for any addition hint, all the best
>
> François
>
> Le lun. 4 févr. 2019 à 12:10, Fabian Hueske <fhue...@gmail.com> a écrit :
>
>> Hi,
>>
>> The files will be read in a streaming fashion.
>> Typically files are broken down into processing splits that are
>> distributed to tasks for reading.
>> How a task reads a file split depends on the implementation, but usually
>> the format reads the split as a stream and does not read the split as a
>> whole before emitting records.
>>
>> Best,
>> Fabian
>>
>> Am Mo., 4. Feb. 2019 um 12:06 Uhr schrieb françois lacombe <
>> francois.laco...@dcbrain.com>:
>>
>>> Hi Fabian,
>>>
>>> Thank you for this input.
>>> This is interesting.
>>>
>>> With such an input format, will all the file will be loaded in memory
>>> before to be processed or will all be streamed?
>>>
>>> All the best
>>> François
>>>
>>> Le mar. 29 janv. 2019 à 22:20, Fabian Hueske <fhue...@gmail.com> a
>>> écrit :
>>>
>>>> Hi,
>>>>
>>>> You can point a file-based input format to a directory and the input
>>>> format should read all files in that directory.
>>>> That works as well for TableSources that are internally use file-based
>>>> input formats.
>>>> Is that what you are looking for?
>>>>
>>>> Best, Fabian
>>>>
>>>> Am Mo., 28. Jan. 2019 um 17:22 Uhr schrieb françois lacombe <
>>>> francois.laco...@dcbrain.com>:
>>>>
>>>>> Hi all,
>>>>>
>>>>> I'm wondering if it's possible and what's the best way to achieve the
>>>>> loading of multiple files with a Json source to a JDBC sink ?
>>>>> I'm running Flink 1.7.0
>>>>>
>>>>> Let's say I have about 1500 files with the same structure (same
>>>>> format, schema, everything) and I want to load them with a *batch* job
>>>>> Can Flink handle the loading of one and each file in a single source
>>>>> and send data to my JDBC sink?
>>>>> I wish I can provide the URL of the directory containing my thousand
>>>>> files to the batch source to make it load all of them sequentially.
>>>>> My sources and sinks are currently available for BatchTableSource, I
>>>>> guess the cost to make them available for streaming would be quite
>>>>> expensive for me for the moment.
>>>>>
>>>>> Have someone ever done this?
>>>>> Am I wrong to expect doing so with a batch job?
>>>>>
>>>>> All the best
>>>>>
>>>>> François Lacombe
>>>>>
>>>>>
>>>>> <http://www.dcbrain.com/>   <https://twitter.com/dcbrain_feed?lang=fr>
>>>>>    <https://www.linkedin.com/company/dcbrain>
>>>>> <https://www.youtube.com/channel/UCSJrWPBLQ58fHPN8lP_SEGw>
>>>>>
>>>>> [image: Arbre vert.jpg] Pensez à la planète, imprimer ce papier que
>>>>> si nécessaire
>>>>>
>>>>
>>>
>>> <http://www.dcbrain.com/>   <https://twitter.com/dcbrain_feed?lang=fr>
>>> <https://www.linkedin.com/company/dcbrain>
>>> <https://www.youtube.com/channel/UCSJrWPBLQ58fHPN8lP_SEGw>
>>>
>>> [image: Arbre vert.jpg] Pensez à la planète, imprimer ce papier que si
>>> nécessaire
>>>
>>
>
> <http://www.dcbrain.com/>   <https://twitter.com/dcbrain_feed?lang=fr>
> <https://www.linkedin.com/company/dcbrain>
> <https://www.youtube.com/channel/UCSJrWPBLQ58fHPN8lP_SEGw>
>
> [image: Arbre vert.jpg] Pensez à la planète, imprimer ce papier que si
> nécessaire
>

Reply via email to