[jira] [Comment Edited] (FLINK-34565) Enhance flink kubernetes configMap to accommodate additional configuration files

2024-04-03 Thread Surendra Singh Lilhore (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34565?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17833440#comment-17833440
 ] 

Surendra Singh Lilhore edited comment on FLINK-34565 at 4/3/24 9:21 AM:


[~zhuzh] If the user files are dynamic, then it is very useful to support them 
in ConfigMap, especially in an App Mode cluster.

 A similar use case was discussed on the user mailing list: [link to the 
mailing list 
thread.|https://lists.apache.org/thread/md2zq0dbvt2dxytdfxw16jbfh02yq0w9]

[~wangyang0918] , Any thought about this?


was (Author: surendrasingh):
[~zhuzh] If the user files are dynamic, then it is very useful to support them 
in ConfigMap, especially in an App Mode cluster.

 A similar use case was discussed on the user mailing list: [link to the 
mailing list 
thread.|https://lists.apache.org/thread/md2zq0dbvt2dxytdfxw16jbfh02yq0w9]

> Enhance flink kubernetes configMap to accommodate additional configuration 
> files
> 
>
> Key: FLINK-34565
> URL: https://issues.apache.org/jira/browse/FLINK-34565
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / Kubernetes
>Reporter: Surendra Singh Lilhore
>Priority: Major
>  Labels: pull-request-available
>
> Flink kubernetes client currently supports a fixed number of files 
> (flink-conf.yaml, logback-console.xml, log4j-console.properties) in the JM 
> and TM Pod ConfigMap. In certain scenarios, particularly in app mode, 
> additional configuration files are required for jobs to run successfully. 
> Presently, users must resort to workarounds to include dynamic configuration 
> files in the JM and TM. This proposed improvement allows users to easily add 
> extra files by configuring the 
> '{*}kubernetes.flink.configmap.additional.resources{*}' property. Users can 
> provide a semicolon-separated list of local files in the client Flink config 
> directory that should be included in the Flink ConfigMap.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-34565) Enhance flink kubernetes configMap to accommodate additional configuration files

2024-04-03 Thread Surendra Singh Lilhore (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34565?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17833440#comment-17833440
 ] 

Surendra Singh Lilhore commented on FLINK-34565:


[~zhuzh] If the user files are dynamic, then it is very useful to support them 
in ConfigMap, especially in an App Mode cluster.

 A similar use case was discussed on the user mailing list: [link to the 
mailing list 
thread.|https://lists.apache.org/thread/md2zq0dbvt2dxytdfxw16jbfh02yq0w9]

> Enhance flink kubernetes configMap to accommodate additional configuration 
> files
> 
>
> Key: FLINK-34565
> URL: https://issues.apache.org/jira/browse/FLINK-34565
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / Kubernetes
>Reporter: Surendra Singh Lilhore
>Priority: Major
>  Labels: pull-request-available
>
> Flink kubernetes client currently supports a fixed number of files 
> (flink-conf.yaml, logback-console.xml, log4j-console.properties) in the JM 
> and TM Pod ConfigMap. In certain scenarios, particularly in app mode, 
> additional configuration files are required for jobs to run successfully. 
> Presently, users must resort to workarounds to include dynamic configuration 
> files in the JM and TM. This proposed improvement allows users to easily add 
> extra files by configuring the 
> '{*}kubernetes.flink.configmap.additional.resources{*}' property. Users can 
> provide a semicolon-separated list of local files in the client Flink config 
> directory that should be included in the Flink ConfigMap.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-34565) Enhance flink kubernetes configMap to accommodate additional configuration files

2024-03-27 Thread Surendra Singh Lilhore (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34565?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17831263#comment-17831263
 ] 

Surendra Singh Lilhore commented on FLINK-34565:


[~zhuzh] can you review this ?

> Enhance flink kubernetes configMap to accommodate additional configuration 
> files
> 
>
> Key: FLINK-34565
> URL: https://issues.apache.org/jira/browse/FLINK-34565
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / Kubernetes
>Reporter: Surendra Singh Lilhore
>Priority: Major
>  Labels: pull-request-available
>
> Flink kubernetes client currently supports a fixed number of files 
> (flink-conf.yaml, logback-console.xml, log4j-console.properties) in the JM 
> and TM Pod ConfigMap. In certain scenarios, particularly in app mode, 
> additional configuration files are required for jobs to run successfully. 
> Presently, users must resort to workarounds to include dynamic configuration 
> files in the JM and TM. This proposed improvement allows users to easily add 
> extra files by configuring the 
> '{*}kubernetes.flink.configmap.additional.resources{*}' property. Users can 
> provide a semicolon-separated list of local files in the client Flink config 
> directory that should be included in the Flink ConfigMap.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-34565) Enhance flink kubernetes configMap to accommodate additional configuration files

2024-03-01 Thread Surendra Singh Lilhore (Jira)
Surendra Singh Lilhore created FLINK-34565:
--

 Summary: Enhance flink kubernetes configMap to accommodate 
additional configuration files
 Key: FLINK-34565
 URL: https://issues.apache.org/jira/browse/FLINK-34565
 Project: Flink
  Issue Type: Bug
  Components: Deployment / Kubernetes
Reporter: Surendra Singh Lilhore


Flink kubernetes client currently supports a fixed number of files 
(flink-conf.yaml, logback-console.xml, log4j-console.properties) in the JM and 
TM Pod ConfigMap. In certain scenarios, particularly in app mode, additional 
configuration files are required for jobs to run successfully. Presently, users 
must resort to workarounds to include dynamic configuration files in the JM and 
TM. This proposed improvement allows users to easily add extra files by 
configuring the '{*}kubernetes.flink.configmap.additional.resources{*}' 
property. Users can provide a semicolon-separated list of local files in the 
client Flink config directory that should be included in the Flink ConfigMap.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-24150) Support to configure cpu resource request and limit in pod template

2023-12-04 Thread Surendra Singh Lilhore (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-24150?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17793128#comment-17793128
 ] 

Surendra Singh Lilhore commented on FLINK-24150:


[~richard.su] , I have given [patch|https://github.com/apache/flink/pull/23768] 
for similar issue in Flink-33609. Check if this can help here.

> Support to configure cpu resource request and limit in pod template
> ---
>
> Key: FLINK-24150
> URL: https://issues.apache.org/jira/browse/FLINK-24150
> Project: Flink
>  Issue Type: New Feature
>  Components: Deployment / Kubernetes
>Reporter: Yang Wang
>Priority: Major
>
> Why Flink needs to overwrite memory resource defined in pod template?
> The major reason is that Flink need to ensure the consistency between Flink 
> configuration
> (\{{taskmanager.memory.process.size}} , {{kubernetes.taskmanager.cpu}}) and 
> pod template resource settings. Since users could specify the total process 
> memory or detailed memory[2], Flink will calculate the pod resource 
> internally.
>  
> For the CPU case the template’s requests/limits should have priority if they 
> are specified. The factor could still be used if the template doesn’t specify 
> anything. The logic could be something like this:
>  # To choose CPU request
>  # Read pod template first
>  # If template doesn’t have anything, read from {{kubernetes.taskmanager.cpu}}
>  # If configuration is not specified, fall back to default
>  # To choose CPU limit
>  # Read from template first
>  # If template doesn’t have anything, apply factor to what was chosen in step 
> 1, where the default factor is 1.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (FLINK-33609) Take into account the resource limit specified in the pod template.

2023-11-26 Thread Surendra Singh Lilhore (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33609?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Surendra Singh Lilhore resolved FLINK-33609.

Resolution: Duplicate

There is one more Jira to fix this issue in better way : FLINK-33548

> Take into account the resource limit specified in the pod template.
> ---
>
> Key: FLINK-33609
> URL: https://issues.apache.org/jira/browse/FLINK-33609
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / Kubernetes
>Affects Versions: 1.16.0
>Reporter: Surendra Singh Lilhore
>Priority: Major
>  Labels: pull-request-available
>
> Flink is currently not considering the pod template resource limits and is 
> only utilizing the limit obtained from the configured or default limit 
> factor. Flink should consider both the value obtained from the limit factor 
> and the pod template resource limits. It should take the maximum value of the 
> pod template resource limits and the value obtained from the limit factor 
> calculation.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-33609) Take into account the resource limit specified in the pod template.

2023-11-21 Thread Surendra Singh Lilhore (Jira)
Surendra Singh Lilhore created FLINK-33609:
--

 Summary: Take into account the resource limit specified in the pod 
template.
 Key: FLINK-33609
 URL: https://issues.apache.org/jira/browse/FLINK-33609
 Project: Flink
  Issue Type: Bug
  Components: Deployment / Kubernetes
Affects Versions: 1.16.0
Reporter: Surendra Singh Lilhore


Flink is currently not considering the pod template resource limits and is only 
utilizing the limit obtained from the configured or default limit factor. Flink 
should consider both the value obtained from the limit factor and the pod 
template resource limits. It should take the maximum value of the pod template 
resource limits and the value obtained from the limit factor calculation.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-33095) Job jar related issue should be reported as BAD_REQUEST instead of INTERNAL_SERVER_ERROR

2023-09-16 Thread Surendra Singh Lilhore (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33095?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Surendra Singh Lilhore updated FLINK-33095:
---
Attachment: image-2023-09-16-13-21-24-634.png

> Job jar related issue should be reported as BAD_REQUEST instead of 
> INTERNAL_SERVER_ERROR
> 
>
> Key: FLINK-33095
> URL: https://issues.apache.org/jira/browse/FLINK-33095
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / REST
>Affects Versions: 1.16.0
>Reporter: Surendra Singh Lilhore
>Priority: Major
> Attachments: image-2023-09-16-13-20-15-052.png, 
> image-2023-09-16-13-21-24-634.png
>
>
> When submitting a job with incorrect parameters, such as an invalid entry 
> class, the current response is an internal server error.
> To enhance the user experience and consistency, it is recommended to throw a 
> Rest exception and return a BAD_REQUEST response code in such cases.
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-33095) Job jar related issue should be reported as BAD_REQUEST instead of INTERNAL_SERVER_ERROR

2023-09-16 Thread Surendra Singh Lilhore (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33095?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17765922#comment-17765922
 ] 

Surendra Singh Lilhore commented on FLINK-33095:


Current response from server :

!image-2023-09-16-13-20-15-052.png!

==

It should be something like this :

!image-2023-09-16-13-21-24-634.png!

 

> Job jar related issue should be reported as BAD_REQUEST instead of 
> INTERNAL_SERVER_ERROR
> 
>
> Key: FLINK-33095
> URL: https://issues.apache.org/jira/browse/FLINK-33095
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / REST
>Affects Versions: 1.16.0
>Reporter: Surendra Singh Lilhore
>Priority: Major
> Attachments: image-2023-09-16-13-20-15-052.png, 
> image-2023-09-16-13-21-24-634.png
>
>
> When submitting a job with incorrect parameters, such as an invalid entry 
> class, the current response is an internal server error.
> To enhance the user experience and consistency, it is recommended to throw a 
> Rest exception and return a BAD_REQUEST response code in such cases.
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-33095) Job jar related issue should be reported as BAD_REQUEST instead of INTERNAL_SERVER_ERROR

2023-09-16 Thread Surendra Singh Lilhore (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33095?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Surendra Singh Lilhore updated FLINK-33095:
---
Attachment: image-2023-09-16-13-20-15-052.png

> Job jar related issue should be reported as BAD_REQUEST instead of 
> INTERNAL_SERVER_ERROR
> 
>
> Key: FLINK-33095
> URL: https://issues.apache.org/jira/browse/FLINK-33095
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / REST
>Affects Versions: 1.16.0
>Reporter: Surendra Singh Lilhore
>Priority: Major
> Attachments: image-2023-09-16-13-20-15-052.png
>
>
> When submitting a job with incorrect parameters, such as an invalid entry 
> class, the current response is an internal server error.
> To enhance the user experience and consistency, it is recommended to throw a 
> Rest exception and return a BAD_REQUEST response code in such cases.
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-33095) Job jar related issue should be reported as BAD_REQUEST instead of INTERNAL_SERVER_ERROR

2023-09-16 Thread Surendra Singh Lilhore (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33095?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Surendra Singh Lilhore updated FLINK-33095:
---
Attachment: (was: image-2023-09-16-13-17-14-869.png)

> Job jar related issue should be reported as BAD_REQUEST instead of 
> INTERNAL_SERVER_ERROR
> 
>
> Key: FLINK-33095
> URL: https://issues.apache.org/jira/browse/FLINK-33095
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / REST
>Affects Versions: 1.16.0
>Reporter: Surendra Singh Lilhore
>Priority: Major
>
> When submitting a job with incorrect parameters, such as an invalid entry 
> class, the current response is an internal server error.
> To enhance the user experience and consistency, it is recommended to throw a 
> Rest exception and return a BAD_REQUEST response code in such cases.
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-33095) Job jar related issue should be reported as BAD_REQUEST instead of INTERNAL_SERVER_ERROR

2023-09-16 Thread Surendra Singh Lilhore (Jira)
Surendra Singh Lilhore created FLINK-33095:
--

 Summary: Job jar related issue should be reported as BAD_REQUEST 
instead of INTERNAL_SERVER_ERROR
 Key: FLINK-33095
 URL: https://issues.apache.org/jira/browse/FLINK-33095
 Project: Flink
  Issue Type: Bug
  Components: Runtime / REST
Affects Versions: 1.16.0
Reporter: Surendra Singh Lilhore


When submitting a job with incorrect parameters, such as an invalid entry 
class, the current response is an internal server error.

To enhance the user experience and consistency, it is recommended to throw a 
Rest exception and return a BAD_REQUEST response code in such cases.

 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-33095) Job jar related issue should be reported as BAD_REQUEST instead of INTERNAL_SERVER_ERROR

2023-09-16 Thread Surendra Singh Lilhore (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33095?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Surendra Singh Lilhore updated FLINK-33095:
---
Attachment: (was: image-2023-09-16-13-17-49-945.png)

> Job jar related issue should be reported as BAD_REQUEST instead of 
> INTERNAL_SERVER_ERROR
> 
>
> Key: FLINK-33095
> URL: https://issues.apache.org/jira/browse/FLINK-33095
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / REST
>Affects Versions: 1.16.0
>Reporter: Surendra Singh Lilhore
>Priority: Major
>
> When submitting a job with incorrect parameters, such as an invalid entry 
> class, the current response is an internal server error.
> To enhance the user experience and consistency, it is recommended to throw a 
> Rest exception and return a BAD_REQUEST response code in such cases.
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-32122) Update the Azure Blob Storage document to assist in configuring the MSI provider with a shaded class name

2023-05-17 Thread Surendra Singh Lilhore (Jira)
Surendra Singh Lilhore created FLINK-32122:
--

 Summary: Update the Azure Blob Storage document to assist in 
configuring the MSI provider with a shaded class name
 Key: FLINK-32122
 URL: https://issues.apache.org/jira/browse/FLINK-32122
 Project: Flink
  Issue Type: Improvement
  Components: Connectors / FileSystem, Documentation
Reporter: Surendra Singh Lilhore


Many users have reported on the mailing list that they are unable to configure 
the ABFS filesystem as a checkpoint directory. This is often due to 
ClassNotFoundException errors for Hadoop classes that are configured in the 
configuration value. For instance, when using MsiTokenProvider for ABFS storage 
in Flink, it should be configured with the shaded class name. However, many 
users mistakenly use the Hadoop class name or package instead.

 

fs.azure.account.oauth.provider.type: 
*org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.oauth2.MsiTokenProvider*



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-30745) Check-pointing with Azure Data Lake Storage

2023-01-21 Thread Surendra Singh Lilhore (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-30745?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17679514#comment-17679514
 ] 

Surendra Singh Lilhore edited comment on FLINK-30745 at 1/22/23 5:55 AM:
-

[~dheerajpanangat] , sorry for late reply.

As mentioned in Flink doc ([Azure blob Flink configuration 
|https://nightlies.apache.org/flink/flink-docs-master/docs/deployment/filesystems/azure/]
 ), you need to configure abfs properties in *flink-conf.yaml.*

[HadoopConfigLoader 
|https://github.com/apache/flink/blob/master/flink-filesystems/flink-hadoop-fs/src/main/java/org/apache/flink/runtime/util/HadoopConfigLoader.java#L82]
 load this configuration from flink configuration. 
{quote}Provided the shaded classes instead of Hadoop classes
{quote}
You are correct

 

Please configure below properties in flink-conf.yaml in Kubernetes cluster and 
try. 
{noformat}
fs.azure.account.auth.type : OAuth

fs.azure.account.oauth.provider.type : 
org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.oauth2.ClientCredsTokenProvider

fs.azure.account.oauth2.client.id : 

fs.azure.account.oauth2.client.secret : 

fs.azure.account.oauth2.client.endpoint : 
https://XX.com/X/oauth2/token {noformat}
 

 


was (Author: surendrasingh):
[~dheerajpanangat] , sorry for late reply.

As mentioned in Flink doc ([Azure blob Flink configuration 
|https://nightlies.apache.org/flink/flink-docs-master/docs/deployment/filesystems/azure/]
 ), you need to configure abfs properties in *flink-conf.yaml.*

[HadoopConfigLoader 
|https://github.com/apache/flink/blob/master/flink-filesystems/flink-hadoop-fs/src/main/java/org/apache/flink/runtime/util/HadoopConfigLoader.java#L82]
 load this configuration from flink configuration. 
{quote}Provided the shaded classes instead of Hadoop classes
{quote}
You we correct

 

Please configure below properties in flink-conf.yaml in Kubernetes cluster and 
try. 
{noformat}
fs.azure.account.auth.type : OAuth

fs.azure.account.oauth.provider.type : 
org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.oauth2.ClientCredsTokenProvider

fs.azure.account.oauth2.client.id : 

fs.azure.account.oauth2.client.secret : 

fs.azure.account.oauth2.client.endpoint : 
https://XX.com/X/oauth2/token {noformat}
 

 

> Check-pointing with Azure Data Lake Storage
> ---
>
> Key: FLINK-30745
> URL: https://issues.apache.org/jira/browse/FLINK-30745
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / FileSystem
>Affects Versions: 1.15.2, 1.14.6
>Reporter: Dheeraj Panangat
>Priority: Major
>
> Hi,
> While checkpointing to Azure Blob Storage using Flink, we get the following 
> error :
> {code:java}
> Caused by: Configuration property .dfs.core.windows.net not 
> found.
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AbfsConfiguration.getStorageAccountKey(AbfsConfiguration.java:372)
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AzureBlobFileSystemStore.initializeClient(AzureBlobFileSystemStore.java:1133)
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AzureBlobFileSystemStore.(AzureBlobFileSystemStore.java:174)
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AzureBlobFileSystem.initialize(AzureBlobFileSystem.java:110)
>  {code}
> We have given the configurations in core-site.xml too for following
> {code:java}
> fs.hdfs.impl
> fs.abfs.impl -> org.apache.hadoop.fs.azurebfs.AzureBlobFileSystem
> fs.file.impl
> fs.azure.account.auth.type
> fs.azure.account.oauth.provider.type
> fs.azure.account.oauth2.client.id
> fs.azure.account.oauth2.client.secret
> fs.azure.account.oauth2.client.endpoint
> fs.azure.createRemoteFileSystemDuringInitialization -> true {code}
> On debugging found that flink reads from core-default-shaded.xml, but even if 
> the properties are specified there, the default configs are not loaded and we 
> get a different exception as :
> {code:java}
> Caused by: Unable to load key provider class.
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AbfsConfiguration.getTokenProvider(AbfsConfiguration.java:540)
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AzureBlobFileSystemStore.initializeClient(AzureBlobFileSystemStore.java:1136)
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AzureBlobFileSystemStore.(AzureBlobFileSystemStore.java:174)
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AzureBlobFileSystem.initialize(AzureBlobFileSystem.java:110)
>  {code}
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-30745) Check-pointing with Azure Data Lake Storage

2023-01-21 Thread Surendra Singh Lilhore (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-30745?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17679514#comment-17679514
 ] 

Surendra Singh Lilhore edited comment on FLINK-30745 at 1/22/23 5:54 AM:
-

[~dheerajpanangat] , sorry for late reply.

As mentioned in Flink doc ([Azure blob Flink configuration 
|https://nightlies.apache.org/flink/flink-docs-master/docs/deployment/filesystems/azure/]
 ), you need to configure abfs properties in *flink-conf.yaml.*

[HadoopConfigLoader 
|https://github.com/apache/flink/blob/master/flink-filesystems/flink-hadoop-fs/src/main/java/org/apache/flink/runtime/util/HadoopConfigLoader.java#L82]
 load this configuration from flink configuration. 
{quote}Provided the shaded classes instead of Hadoop classes
{quote}
You we correct

 

Please configure below properties in flink-conf.yaml in Kubernetes cluster and 
try. 
{noformat}
fs.azure.account.auth.type : OAuth

fs.azure.account.oauth.provider.type : 
org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.oauth2.ClientCredsTokenProvider

fs.azure.account.oauth2.client.id : 

fs.azure.account.oauth2.client.secret : 

fs.azure.account.oauth2.client.endpoint : 
https://XX.com/X/oauth2/token {noformat}
 

 


was (Author: surendrasingh):
[~dheerajpanangat] , sorry for late reply.

As mentioned in flink doc ([Azure Blob Storage | Apache 
Flink|https://nightlies.apache.org/flink/flink-docs-master/docs/deployment/filesystems/azure/]),
 you need to configure abfs properties in *flink-conf.yaml.*

[HadoopConfigLoader 
|https://github.com/apache/flink/blob/master/flink-filesystems/flink-hadoop-fs/src/main/java/org/apache/flink/runtime/util/HadoopConfigLoader.java#L82]
 load this configuration from flink configuration. 
{quote}Provided the shaded classes instead of Hadoop classes
{quote}
You we correct

 

Please configure below properties in flink-conf.yaml in Kubernetes cluster and 
try. 
{noformat}
fs.azure.account.auth.type : OAuth

fs.azure.account.oauth.provider.type : 
org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.oauth2.ClientCredsTokenProvider

fs.azure.account.oauth2.client.id : 

fs.azure.account.oauth2.client.secret : 

fs.azure.account.oauth2.client.endpoint : 
https://XX.com/X/oauth2/token {noformat}
 

 

> Check-pointing with Azure Data Lake Storage
> ---
>
> Key: FLINK-30745
> URL: https://issues.apache.org/jira/browse/FLINK-30745
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / FileSystem
>Affects Versions: 1.15.2, 1.14.6
>Reporter: Dheeraj Panangat
>Priority: Major
>
> Hi,
> While checkpointing to Azure Blob Storage using Flink, we get the following 
> error :
> {code:java}
> Caused by: Configuration property .dfs.core.windows.net not 
> found.
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AbfsConfiguration.getStorageAccountKey(AbfsConfiguration.java:372)
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AzureBlobFileSystemStore.initializeClient(AzureBlobFileSystemStore.java:1133)
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AzureBlobFileSystemStore.(AzureBlobFileSystemStore.java:174)
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AzureBlobFileSystem.initialize(AzureBlobFileSystem.java:110)
>  {code}
> We have given the configurations in core-site.xml too for following
> {code:java}
> fs.hdfs.impl
> fs.abfs.impl -> org.apache.hadoop.fs.azurebfs.AzureBlobFileSystem
> fs.file.impl
> fs.azure.account.auth.type
> fs.azure.account.oauth.provider.type
> fs.azure.account.oauth2.client.id
> fs.azure.account.oauth2.client.secret
> fs.azure.account.oauth2.client.endpoint
> fs.azure.createRemoteFileSystemDuringInitialization -> true {code}
> On debugging found that flink reads from core-default-shaded.xml, but even if 
> the properties are specified there, the default configs are not loaded and we 
> get a different exception as :
> {code:java}
> Caused by: Unable to load key provider class.
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AbfsConfiguration.getTokenProvider(AbfsConfiguration.java:540)
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AzureBlobFileSystemStore.initializeClient(AzureBlobFileSystemStore.java:1136)
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AzureBlobFileSystemStore.(AzureBlobFileSystemStore.java:174)
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AzureBlobFileSystem.initialize(AzureBlobFileSystem.java:110)
>  {code}
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-30745) Check-pointing with Azure Data Lake Storage

2023-01-21 Thread Surendra Singh Lilhore (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-30745?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17679514#comment-17679514
 ] 

Surendra Singh Lilhore commented on FLINK-30745:


[~dheerajpanangat] , sorry for late reply.

As mentioned in flink doc ([Azure Blob Storage | Apache 
Flink|https://nightlies.apache.org/flink/flink-docs-master/docs/deployment/filesystems/azure/]),
 you need to configure abfs properties in *flink-conf.yaml.*

[HadoopConfigLoader 
|https://github.com/apache/flink/blob/master/flink-filesystems/flink-hadoop-fs/src/main/java/org/apache/flink/runtime/util/HadoopConfigLoader.java#L82]
 load this configuration from flink configuration. 
{quote}Provided the shaded classes instead of Hadoop classes
{quote}
You we correct

 

Please configure below properties in flink-conf.yaml in Kubernetes cluster and 
try. 
{noformat}
fs.azure.account.auth.type : OAuth

fs.azure.account.oauth.provider.type : 
org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.oauth2.ClientCredsTokenProvider

fs.azure.account.oauth2.client.id : 

fs.azure.account.oauth2.client.secret : 

fs.azure.account.oauth2.client.endpoint : 
https://XX.com/X/oauth2/token {noformat}
 

 

> Check-pointing with Azure Data Lake Storage
> ---
>
> Key: FLINK-30745
> URL: https://issues.apache.org/jira/browse/FLINK-30745
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / FileSystem
>Affects Versions: 1.15.2, 1.14.6
>Reporter: Dheeraj Panangat
>Priority: Major
>
> Hi,
> While checkpointing to Azure Blob Storage using Flink, we get the following 
> error :
> {code:java}
> Caused by: Configuration property .dfs.core.windows.net not 
> found.
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AbfsConfiguration.getStorageAccountKey(AbfsConfiguration.java:372)
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AzureBlobFileSystemStore.initializeClient(AzureBlobFileSystemStore.java:1133)
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AzureBlobFileSystemStore.(AzureBlobFileSystemStore.java:174)
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AzureBlobFileSystem.initialize(AzureBlobFileSystem.java:110)
>  {code}
> We have given the configurations in core-site.xml too for following
> {code:java}
> fs.hdfs.impl
> fs.abfs.impl -> org.apache.hadoop.fs.azurebfs.AzureBlobFileSystem
> fs.file.impl
> fs.azure.account.auth.type
> fs.azure.account.oauth.provider.type
> fs.azure.account.oauth2.client.id
> fs.azure.account.oauth2.client.secret
> fs.azure.account.oauth2.client.endpoint
> fs.azure.createRemoteFileSystemDuringInitialization -> true {code}
> On debugging found that flink reads from core-default-shaded.xml, but even if 
> the properties are specified there, the default configs are not loaded and we 
> get a different exception as :
> {code:java}
> Caused by: Unable to load key provider class.
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AbfsConfiguration.getTokenProvider(AbfsConfiguration.java:540)
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AzureBlobFileSystemStore.initializeClient(AzureBlobFileSystemStore.java:1136)
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AzureBlobFileSystemStore.(AzureBlobFileSystemStore.java:174)
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AzureBlobFileSystem.initialize(AzureBlobFileSystem.java:110)
>  {code}
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-30745) Check-pointing with Azure Data Lake Storage

2023-01-18 Thread Surendra Singh Lilhore (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-30745?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17678329#comment-17678329
 ] 

Surendra Singh Lilhore commented on FLINK-30745:


Thanks [~dheerajpanangat] for reporting this issue.

What value you configured for property 
"fs.azure.account.keyprovider." ?

> Check-pointing with Azure Data Lake Storage
> ---
>
> Key: FLINK-30745
> URL: https://issues.apache.org/jira/browse/FLINK-30745
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / FileSystem
>Affects Versions: 1.15.2, 1.14.6
>Reporter: Dheeraj Panangat
>Priority: Major
>
> Hi,
> While checkpointing to Azure Blob Storage using Flink, we get the following 
> error :
> {code:java}
> Caused by: Configuration property .dfs.core.windows.net not 
> found.
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AbfsConfiguration.getStorageAccountKey(AbfsConfiguration.java:372)
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AzureBlobFileSystemStore.initializeClient(AzureBlobFileSystemStore.java:1133)
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AzureBlobFileSystemStore.(AzureBlobFileSystemStore.java:174)
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AzureBlobFileSystem.initialize(AzureBlobFileSystem.java:110)
>  {code}
> We have given the configurations in core-site.xml too for following
> {code:java}
> fs.hdfs.impl
> fs.abfs.impl -> org.apache.hadoop.fs.azurebfs.AzureBlobFileSystem
> fs.file.impl
> fs.azure.account.auth.type
> fs.azure.account.oauth.provider.type
> fs.azure.account.oauth2.client.id
> fs.azure.account.oauth2.client.secret
> fs.azure.account.oauth2.client.endpoint
> fs.azure.createRemoteFileSystemDuringInitialization -> true {code}
> On debugging found that flink reads from core-default-shaded.xml, but even if 
> the properties are specified there, the default configs are not loaded and we 
> get a different exception as :
> {code:java}
> Caused by: Unable to load key provider class.
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AbfsConfiguration.getTokenProvider(AbfsConfiguration.java:540)
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AzureBlobFileSystemStore.initializeClient(AzureBlobFileSystemStore.java:1136)
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AzureBlobFileSystemStore.(AzureBlobFileSystemStore.java:174)
> at 
> org.apache.flink.fs.shaded.hadoop3.org.apache.hadoop.fs.azurebfs.AzureBlobFileSystem.initialize(AzureBlobFileSystem.java:110)
>  {code}
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)