Hi Fabian,

Thank you for the clarification.

Best,
Morven Huang

On Wed, Apr 10, 2019 at 9:57 PM Fabian Hueske <fhue...@gmail.com> wrote:

> Hi,
>
> Flink's Hadoop compatibility functions just wrap functions that were
> implemented against Hadoop's interfaces in wrapper functions that are
> implemented against Flink's interfaces.
> There is no Hadoop cluster started or MapReduce job being executed.
>
> Job is just a class of the Hadoop API. It does not imply that a Hadoop job
> is executed.
>
> Best,
> Fabian
>
>
>
> Am Mi., 10. Apr. 2019 um 15:12 Uhr schrieb Morven Huang <
> morven.hu...@gmail.com>:
>
>> Hi,
>>
>>
>>
>> I’d like to sink my data into hdfs using
>> SequenceFileAsBinaryOutputFormat with compression, and I find a way from
>> the link
>> https://ci.apache.org/projects/flink/flink-docs-stable/dev/batch/hadoop_compatibility.html,
>> the code works, but I’m curious to know, since it creates a mapreduce
>> Job instance here, would this Flink application creates and run a mapreduce
>> underneath? If so, will it kill performance?
>>
>>
>>
>> I tried to figure out by looking into log, but couldn’t get a clue, hope
>> people could shed some light here. Thank you.
>>
>>
>>
>> Job job = Job.getInstance();
>>
>> HadoopOutputFormat<BytesWritable, BytesWritable> hadoopOF = new
>> HadoopOutputFormat<BytesWritable, BytesWritable>(
>>
>>                     new SequenceFileAsBinaryOutputFormat(), job);
>>
>>
>>
>> hadoopOF.getConfiguration().set("mapreduce.output.fileoutputformat.compress",
>> "true");
>>
>> hadoopOF.getConfiguration().set("mapreduce.output.fileoutputformat.compress.type",
>> CompressionType.BLOCK.toString());
>>
>> TextOutputFormat.setOutputPath(job, new Path("hdfs://..."));
>>
>> dataset.output(hadoopOF);
>>
>

Reply via email to