[ https://issues.apache.org/jira/browse/FLINK-32318?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Luís Costa updated FLINK-32318: ------------------------------- Description: Greetings, I'm trying to configure [Flink's Kubernetes HA services|https://nightlies.apache.org/flink/flink-docs-release-1.13/docs/deployment/ha/kubernetes_ha/] for flink operator jobs, but got an error regarding s3 plugin: _"Could not find a file system implementation for scheme 's3'. The scheme is directly supported by Flink through the following plugin(s): flink-s3-fs-hadoop, flink-s3-fs-presto"_ {code:java} 2023-06-12 10:05:16,981 INFO akka.remote.Remoting [] - Starting remoting 2023-06-12 10:05:17,194 INFO akka.remote.Remoting [] - Remoting started; listening on addresses :[akka.tcp://flink@10.4.125.209:6123] 2023-06-12 10:05:17,377 INFO org.apache.flink.runtime.rpc.akka.AkkaRpcServiceUtils [] - Actor system started at akka.tcp://flink@10.4.125.209:6123 2023-06-12 10:05:18,175 INFO org.apache.flink.runtime.entrypoint.ClusterEntrypoint [] - Shutting KubernetesApplicationClusterEntrypoint down with application status FAILED. Diagnostics org.apache.flink.util.FlinkException: Could not create the ha services from the instantiated HighAvailabilityServicesFactory org.apache.flink.kubernetes.highavailability.KubernetesHaServicesFactory. at org.apache.flink.runtime.highavailability.HighAvailabilityServicesUtils.createCustomHAServices(HighAvailabilityServicesUtils.java:299) at org.apache.flink.runtime.highavailability.HighAvailabilityServicesUtils.createCustomHAServices(HighAvailabilityServicesUtils.java:285) at org.apache.flink.runtime.highavailability.HighAvailabilityServicesUtils.createHighAvailabilityServices(HighAvailabilityServicesUtils.java:145) at org.apache.flink.runtime.entrypoint.ClusterEntrypoint.createHaServices(ClusterEntrypoint.java:439) at org.apache.flink.runtime.entrypoint.ClusterEntrypoint.initializeServices(ClusterEntrypoint.java:382) at org.apache.flink.runtime.entrypoint.ClusterEntrypoint.runCluster(ClusterEntrypoint.java:282) at org.apache.flink.runtime.entrypoint.ClusterEntrypoint.lambda$startCluster$1(ClusterEntrypoint.java:232) at org.apache.flink.runtime.security.contexts.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:28) at org.apache.flink.runtime.entrypoint.ClusterEntrypoint.startCluster(ClusterEntrypoint.java:229) at org.apache.flink.runtime.entrypoint.ClusterEntrypoint.runClusterEntrypoint(ClusterEntrypoint.java:729) at org.apache.flink.kubernetes.entrypoint.KubernetesApplicationClusterEntrypoint.main(KubernetesApplicationClusterEntrypoint.java:86) Caused by: java.io.IOException: Could not create FileSystem for highly available storage path (s3://td-infra-stg-us-east-1-s3-flinkoperator/flink-data/ha/flink-basic-example-xpto) at org.apache.flink.runtime.blob.BlobUtils.createFileSystemBlobStore(BlobUtils.java:102) at org.apache.flink.runtime.blob.BlobUtils.createBlobStoreFromConfig(BlobUtils.java:86) at org.apache.flink.kubernetes.highavailability.KubernetesHaServicesFactory.createHAServices(KubernetesHaServicesFactory.java:41) at org.apache.flink.runtime.highavailability.HighAvailabilityServicesUtils.createCustomHAServices(HighAvailabilityServicesUtils.java:296) ... 10 more Caused by: org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Could not find a file system implementation for scheme 's3'. The scheme is directly supported by Flink through the following plugin(s): flink-s3-fs-hadoop, flink-s3-fs-presto. Please ensure that each plugin resides within its own subfolder within the plugins directory. See https://nightlies.apache.org/flink/flink-docs-stable/docs/deployment/filesystems/plugins/ for more information. If you want to use a Hadoop file system for that scheme, please add the scheme to the configuration fs.allowed-fallback-filesystems. For a full list of supported file systems, please see https://nightlies.apache.org/flink/flink-docs-stable/ops/filesystems/. at org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:515) at org.apache.flink.core.fs.FileSystem.get(FileSystem.java:409) at org.apache.flink.core.fs.Path.getFileSystem(Path.java:274) at org.apache.flink.runtime.blob.BlobUtils.createFileSystemBlobStore(BlobUtils.java:99) ... 13 more {code} Looking into the job container, can see that s3 plugins are in folder _/opt/flink/opt_ instead of {_}/opt/flink/plugins/s3{_}, as mentioned [here|https://nightlies.apache.org/flink/flink-docs-stable/docs/deployment/filesystems/plugins/] {code:java} root@flink-basic-example-xpto1-86bb9b9d44-hksq8:/opt/flink# cd plugins/ root@flink-basic-example-xpto1-86bb9b9d44-hksq8:/opt/flink/plugins# ls -la total 4 drwxr-xr-x 10 flink flink 210 May 18 06:07 . drwxr-xr-x 1 flink flink 37 Jun 11 20:17 .. drwxr-xr-x 2 flink flink 114 May 18 06:07 external-resource-gpu drwxr-xr-x 2 flink flink 46 May 18 06:07 metrics-datadog drwxr-xr-x 2 flink flink 47 May 18 06:07 metrics-graphite drwxr-xr-x 2 flink flink 47 May 18 06:07 metrics-influx drwxr-xr-x 2 flink flink 42 May 18 06:07 metrics-jmx drwxr-xr-x 2 flink flink 49 May 18 06:07 metrics-prometheus drwxr-xr-x 2 flink flink 44 May 18 06:07 metrics-slf4j drwxr-xr-x 2 flink flink 45 May 18 06:07 metrics-statsd -rwxr-xr-x 1 flink flink 654 May 17 09:19 README.txt root@flink-basic-example-xpto1-86bb9b9d44-hksq8:/opt/flink/plugins# cd .. root@flink-basic-example-xpto1-86bb9b9d44-hksq8:/opt/flink# cd opt/ root@flink-basic-example-xpto1-86bb9b9d44-hksq8:/opt/flink/opt# ls -la | grep s3 -rw-r--r-- 1 flink flink 30515842 May 18 06:00 flink-s3-fs-hadoop-1.16.2.jar -rw-r--r-- 1 flink flink 96171268 May 18 06:00 flink-s3-fs-presto-1.16.2.jar {code} Also, looking into container flink-operator, did not find those s3 plugins Best regards, Luís Costa was: Greetings, I'm trying to configure [Flink's Kubernetes HA services|https://nightlies.apache.org/flink/flink-docs-release-1.13/docs/deployment/ha/kubernetes_ha/] for flink operator jobs, but got an error regarding s3 plugin: _"Could not find a file system implementation for scheme 's3'. The scheme is directly supported by Flink through the following plugin(s): flink-s3-fs-hadoop, flink-s3-fs-presto"_ {code:java} 2023-06-12 10:05:16,981 INFO akka.remote.Remoting [] - Starting remoting 2023-06-12 10:05:17,194 INFO akka.remote.Remoting [] - Remoting started; listening on addresses :[akka.tcp://flink@10.4.125.209:6123] 2023-06-12 10:05:17,377 INFO org.apache.flink.runtime.rpc.akka.AkkaRpcServiceUtils [] - Actor system started at akka.tcp://flink@10.4.125.209:6123 2023-06-12 10:05:18,175 INFO org.apache.flink.runtime.entrypoint.ClusterEntrypoint [] - Shutting KubernetesApplicationClusterEntrypoint down with application status FAILED. Diagnostics org.apache.flink.util.FlinkException: Could not create the ha services from the instantiated HighAvailabilityServicesFactory org.apache.flink.kubernetes.highavailability.KubernetesHaServicesFactory. at org.apache.flink.runtime.highavailability.HighAvailabilityServicesUtils.createCustomHAServices(HighAvailabilityServicesUtils.java:299) at org.apache.flink.runtime.highavailability.HighAvailabilityServicesUtils.createCustomHAServices(HighAvailabilityServicesUtils.java:285) at org.apache.flink.runtime.highavailability.HighAvailabilityServicesUtils.createHighAvailabilityServices(HighAvailabilityServicesUtils.java:145) at org.apache.flink.runtime.entrypoint.ClusterEntrypoint.createHaServices(ClusterEntrypoint.java:439) at org.apache.flink.runtime.entrypoint.ClusterEntrypoint.initializeServices(ClusterEntrypoint.java:382) at org.apache.flink.runtime.entrypoint.ClusterEntrypoint.runCluster(ClusterEntrypoint.java:282) at org.apache.flink.runtime.entrypoint.ClusterEntrypoint.lambda$startCluster$1(ClusterEntrypoint.java:232) at org.apache.flink.runtime.security.contexts.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:28) at org.apache.flink.runtime.entrypoint.ClusterEntrypoint.startCluster(ClusterEntrypoint.java:229) at org.apache.flink.runtime.entrypoint.ClusterEntrypoint.runClusterEntrypoint(ClusterEntrypoint.java:729) at org.apache.flink.kubernetes.entrypoint.KubernetesApplicationClusterEntrypoint.main(KubernetesApplicationClusterEntrypoint.java:86) Caused by: java.io.IOException: Could not create FileSystem for highly available storage path (s3://td-infra-stg-us-east-1-s3-flinkoperator/flink-data/ha/flink-basic-example-xpto) at org.apache.flink.runtime.blob.BlobUtils.createFileSystemBlobStore(BlobUtils.java:102) at org.apache.flink.runtime.blob.BlobUtils.createBlobStoreFromConfig(BlobUtils.java:86) at org.apache.flink.kubernetes.highavailability.KubernetesHaServicesFactory.createHAServices(KubernetesHaServicesFactory.java:41) at org.apache.flink.runtime.highavailability.HighAvailabilityServicesUtils.createCustomHAServices(HighAvailabilityServicesUtils.java:296) ... 10 more Caused by: org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Could not find a file system implementation for scheme 's3'. The scheme is directly supported by Flink through the following plugin(s): flink-s3-fs-hadoop, flink-s3-fs-presto. Please ensure that each plugin resides within its own subfolder within the plugins directory. See https://nightlies.apache.org/flink/flink-docs-stable/docs/deployment/filesystems/plugins/ for more information. If you want to use a Hadoop file system for that scheme, please add the scheme to the configuration fs.allowed-fallback-filesystems. For a full list of supported file systems, please see https://nightlies.apache.org/flink/flink-docs-stable/ops/filesystems/. at org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:515) at org.apache.flink.core.fs.FileSystem.get(FileSystem.java:409) at org.apache.flink.core.fs.Path.getFileSystem(Path.java:274) at org.apache.flink.runtime.blob.BlobUtils.createFileSystemBlobStore(BlobUtils.java:99) ... 13 more {code} Looking into the job container, can see that s3 plugins are in folder _/opt/flink/opt_ instead of {_}/opt/flink/plugins/s3{_}, as mentioned [here|https://nightlies.apache.org/flink/flink-docs-stable/docs/deployment/filesystems/plugins/] {code:java} root@flink-basic-example-xpto1-86bb9b9d44-hksq8:/opt/flink# cd plugins/ root@flink-basic-example-xpto1-86bb9b9d44-hksq8:/opt/flink/plugins# ls -la total 4 drwxr-xr-x 10 flink flink 210 May 18 06:07 . drwxr-xr-x 1 flink flink 37 Jun 11 20:17 .. drwxr-xr-x 2 flink flink 114 May 18 06:07 external-resource-gpu drwxr-xr-x 2 flink flink 46 May 18 06:07 metrics-datadog drwxr-xr-x 2 flink flink 47 May 18 06:07 metrics-graphite drwxr-xr-x 2 flink flink 47 May 18 06:07 metrics-influx drwxr-xr-x 2 flink flink 42 May 18 06:07 metrics-jmx drwxr-xr-x 2 flink flink 49 May 18 06:07 metrics-prometheus drwxr-xr-x 2 flink flink 44 May 18 06:07 metrics-slf4j drwxr-xr-x 2 flink flink 45 May 18 06:07 metrics-statsd -rwxr-xr-x 1 flink flink 654 May 17 09:19 README.txt root@flink-basic-example-xpto1-86bb9b9d44-hksq8:/opt/flink/plugins# cd .. root@flink-basic-example-xpto1-86bb9b9d44-hksq8:/opt/flink# cd opt/ root@flink-basic-example-xpto1-86bb9b9d44-hksq8:/opt/flink/opt# ls -la | grep s3 -rw-r--r-- 1 flink flink 30515842 May 18 06:00 flink-s3-fs-hadoop-1.16.2.jar -rw-r--r-- 1 flink flink 96171268 May 18 06:00 flink-s3-fs-presto-1.16.2.jar {code} Also, looking into container flink-operator, did not find those s3 plugins Best regards, Luís Costa > [flink-operator] missing s3 plugin in folder plugins > ---------------------------------------------------- > > Key: FLINK-32318 > URL: https://issues.apache.org/jira/browse/FLINK-32318 > Project: Flink > Issue Type: Bug > Components: Kubernetes Operator > Affects Versions: kubernetes-operator-1.5.0 > Reporter: Luís Costa > Priority: Minor > > Greetings, > I'm trying to configure [Flink's Kubernetes HA > services|https://nightlies.apache.org/flink/flink-docs-release-1.13/docs/deployment/ha/kubernetes_ha/] > for flink operator jobs, but got an error regarding s3 plugin: _"Could not > find a file system implementation for scheme 's3'. The scheme is directly > supported by Flink through the following plugin(s): flink-s3-fs-hadoop, > flink-s3-fs-presto"_ > {code:java} > 2023-06-12 10:05:16,981 INFO akka.remote.Remoting > [] - Starting remoting > 2023-06-12 10:05:17,194 INFO akka.remote.Remoting > [] - Remoting started; listening on addresses > :[akka.tcp://flink@10.4.125.209:6123] > 2023-06-12 10:05:17,377 INFO > org.apache.flink.runtime.rpc.akka.AkkaRpcServiceUtils [] - Actor > system started at akka.tcp://flink@10.4.125.209:6123 > 2023-06-12 10:05:18,175 INFO > org.apache.flink.runtime.entrypoint.ClusterEntrypoint [] - Shutting > KubernetesApplicationClusterEntrypoint down with application status FAILED. > Diagnostics org.apache.flink.util.FlinkException: Could not create the ha > services from the instantiated HighAvailabilityServicesFactory > org.apache.flink.kubernetes.highavailability.KubernetesHaServicesFactory. > at > org.apache.flink.runtime.highavailability.HighAvailabilityServicesUtils.createCustomHAServices(HighAvailabilityServicesUtils.java:299) > at > org.apache.flink.runtime.highavailability.HighAvailabilityServicesUtils.createCustomHAServices(HighAvailabilityServicesUtils.java:285) > at > org.apache.flink.runtime.highavailability.HighAvailabilityServicesUtils.createHighAvailabilityServices(HighAvailabilityServicesUtils.java:145) > at > org.apache.flink.runtime.entrypoint.ClusterEntrypoint.createHaServices(ClusterEntrypoint.java:439) > at > org.apache.flink.runtime.entrypoint.ClusterEntrypoint.initializeServices(ClusterEntrypoint.java:382) > at > org.apache.flink.runtime.entrypoint.ClusterEntrypoint.runCluster(ClusterEntrypoint.java:282) > at > org.apache.flink.runtime.entrypoint.ClusterEntrypoint.lambda$startCluster$1(ClusterEntrypoint.java:232) > at > org.apache.flink.runtime.security.contexts.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:28) > at > org.apache.flink.runtime.entrypoint.ClusterEntrypoint.startCluster(ClusterEntrypoint.java:229) > at > org.apache.flink.runtime.entrypoint.ClusterEntrypoint.runClusterEntrypoint(ClusterEntrypoint.java:729) > at > org.apache.flink.kubernetes.entrypoint.KubernetesApplicationClusterEntrypoint.main(KubernetesApplicationClusterEntrypoint.java:86) > Caused by: java.io.IOException: Could not create FileSystem for highly > available storage path > (s3://td-infra-stg-us-east-1-s3-flinkoperator/flink-data/ha/flink-basic-example-xpto) > at > org.apache.flink.runtime.blob.BlobUtils.createFileSystemBlobStore(BlobUtils.java:102) > at > org.apache.flink.runtime.blob.BlobUtils.createBlobStoreFromConfig(BlobUtils.java:86) > at > org.apache.flink.kubernetes.highavailability.KubernetesHaServicesFactory.createHAServices(KubernetesHaServicesFactory.java:41) > at > org.apache.flink.runtime.highavailability.HighAvailabilityServicesUtils.createCustomHAServices(HighAvailabilityServicesUtils.java:296) > ... 10 more > Caused by: org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: > Could not find a file system implementation for scheme 's3'. The scheme is > directly supported by Flink through the following plugin(s): > flink-s3-fs-hadoop, flink-s3-fs-presto. Please ensure that each plugin > resides within its own subfolder within the plugins directory. See > https://nightlies.apache.org/flink/flink-docs-stable/docs/deployment/filesystems/plugins/ > for more information. If you want to use a Hadoop file system for that > scheme, please add the scheme to the configuration > fs.allowed-fallback-filesystems. For a full list of supported file systems, > please see > https://nightlies.apache.org/flink/flink-docs-stable/ops/filesystems/. > at > org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:515) > at org.apache.flink.core.fs.FileSystem.get(FileSystem.java:409) > at org.apache.flink.core.fs.Path.getFileSystem(Path.java:274) > at > org.apache.flink.runtime.blob.BlobUtils.createFileSystemBlobStore(BlobUtils.java:99) > ... 13 more {code} > Looking into the job container, can see that s3 plugins are in folder > _/opt/flink/opt_ instead of {_}/opt/flink/plugins/s3{_}, as mentioned > [here|https://nightlies.apache.org/flink/flink-docs-stable/docs/deployment/filesystems/plugins/] > {code:java} > root@flink-basic-example-xpto1-86bb9b9d44-hksq8:/opt/flink# cd plugins/ > root@flink-basic-example-xpto1-86bb9b9d44-hksq8:/opt/flink/plugins# ls -la > total 4 > drwxr-xr-x 10 flink flink 210 May 18 06:07 . > drwxr-xr-x 1 flink flink 37 Jun 11 20:17 .. > drwxr-xr-x 2 flink flink 114 May 18 06:07 external-resource-gpu > drwxr-xr-x 2 flink flink 46 May 18 06:07 metrics-datadog > drwxr-xr-x 2 flink flink 47 May 18 06:07 metrics-graphite > drwxr-xr-x 2 flink flink 47 May 18 06:07 metrics-influx > drwxr-xr-x 2 flink flink 42 May 18 06:07 metrics-jmx > drwxr-xr-x 2 flink flink 49 May 18 06:07 metrics-prometheus > drwxr-xr-x 2 flink flink 44 May 18 06:07 metrics-slf4j > drwxr-xr-x 2 flink flink 45 May 18 06:07 metrics-statsd > -rwxr-xr-x 1 flink flink 654 May 17 09:19 README.txt > root@flink-basic-example-xpto1-86bb9b9d44-hksq8:/opt/flink/plugins# cd .. > root@flink-basic-example-xpto1-86bb9b9d44-hksq8:/opt/flink# cd opt/ > root@flink-basic-example-xpto1-86bb9b9d44-hksq8:/opt/flink/opt# ls -la | grep > s3 > -rw-r--r-- 1 flink flink 30515842 May 18 06:00 flink-s3-fs-hadoop-1.16.2.jar > -rw-r--r-- 1 flink flink 96171268 May 18 06:00 flink-s3-fs-presto-1.16.2.jar > {code} > Also, looking into container flink-operator, did not find those s3 plugins > Best regards, > Luís Costa > -- This message was sent by Atlassian Jira (v8.20.10#820010)