I have successfully connected Azure blob storage to Flink-1.1.

Below are the steps necessary:
- Add hadoop-azure-2.7.2.jar (assuming you are using a Hadoop 2.7 Flink
binary) and azure-storage-4.3.0.jar to <flinkdir>/lib, and set file
permissions / ownership accordingly.
- Add the following to a file 'core-site.xml'

<property>
        <name>fs.wasb.impl</name>
        <value>org.apache.hadoop.fs.azure.NativeAzureFileSystem</value>
 </property>

<property>
        <name>fs.wasbs.impl</name>
        <value>org.apache.hadoop.fs.azure.NativeAzureFileSystem</value>
 </property>

<property>
        <name>fs.azure.account.key.STORAGEACCOUNTNAME.blob.core.windows.net
</name>
        <value>ACCOUNTKEY</value>
</property>

- Update the parameter fs.hdfs.hadoopconf: to the path for the directory
where core-site.xml is located.
- Restart Flink

It is now possible to read from blobs (block and page) by referencing
'wasb://containern...@storageaccountname.blob.core.windows.net/PATH' or
'wasbs://containern...@storageaccountname.blob.core.windows.net/PATH'

Regards,
Lau

On 16 August 2016 at 14:37, MIkkel Islay <my.inputstr...@gmail.com> wrote:
>
> Hello,
>
> I would like to access data in Azure blob storage from Flink, via the
Azure storage HDFS-compatibility interface.
> That is feasible from Apache Drill, and I am thinking something similar
should be doable from Flink. A documentation page on eternal storage
connectors for Flink exist, but it was written pre 1.0.
> Does anyone have experience with setting up a Azure blob connector?
>
> Mikkel




--
Lau Sennels
Founder, scaling biologist
https://dk.linkedin.com/pub/lau-sennels/a9/3b5/196

Reply via email to