Thanks!
I'm using now the *flink:1.7.0-hadoop24-scala_2.12* image.
The Hadoop related error is gone, but I have a new error:

Starting Task Manager
config file:
jobmanager.rpc.address: flink-jobmanager
jobmanager.rpc.port: 6123
jobmanager.heap.size: 1024m
taskmanager.heap.size: 1024m
taskmanager.numberOfTaskSlots: 2
parallelism.default: 1
rest.port: 8081
blob.server.port: 6124
query.server.port: 6125
Starting taskexecutor as a console application on host
flink-taskmanager-54b679f8bb-22b4r.
2018-12-19 13:09:38,469 INFO
org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -
--------------------------------------------------------------------------------
2018-12-19 13:09:38,470 INFO
org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -  Starting
TaskManager (Version: 1.7.0, Rev:49da9f9, Date:28.11.2018 @ 17:59:06 UTC)
2018-12-19 13:09:38,470 INFO
org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -  OS current
user: flink
2018-12-19 13:09:38,921 WARN
org.apache.hadoop.util.NativeCodeLoader                       - Unable to
load native-hadoop library for your platform... using builtin-java classes
where applicable
2018-12-19 13:09:39,307 INFO
org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -  Current
Hadoop/Kerberos user: flink
2018-12-19 13:09:39,307 INFO
org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -  JVM:
OpenJDK 64-Bit Server VM - Oracle Corporation - 1.8/25.181-b13
2018-12-19 13:09:39,307 INFO
org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -  Maximum
heap size: 922 MiBytes
2018-12-19 13:09:39,307 INFO
org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -  JAVA_HOME:
/docker-java-home/jre
2018-12-19 13:09:39,318 INFO
org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -  Hadoop
version: 2.4.1
2018-12-19 13:09:39,318 INFO
org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -  JVM
Options:
2018-12-19 13:09:39,319 INFO
org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -
-XX:+UseG1GC
2018-12-19 13:09:39,319 INFO
org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -     -Xms922M
2018-12-19 13:09:39,320 INFO
org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -     -Xmx922M
2018-12-19 13:09:39,320 INFO
org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -
-XX:MaxDirectMemorySize=8388607T
2018-12-19 13:09:39,320 INFO
org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -
-Dlog4j.configuration=file:/opt/flink/conf/log4j-console.properties
2018-12-19 13:09:39,320 INFO
org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -
-Dlogback.configurationFile=file:/opt/flink/conf/logback-console.xml
2018-12-19 13:09:39,320 INFO
org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -  Program
Arguments:
2018-12-19 13:09:39,321 INFO
org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -
--configDir
2018-12-19 13:09:39,321 INFO
org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -
/opt/flink/conf
2018-12-19 13:09:39,321 INFO
org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -  Classpath:
/opt/flink/lib/flink-python_2.12-1.7.0.jar:/opt/flink/lib/flink-shaded-hadoop2-uber-1.7.0.jar:/opt/flink/lib/log4j-1.2.17.jar:/opt/flink/lib/slf4j-log4j12-1.7.15.jar:/opt/flink/lib/flink-dist_2.12-1.7.0.jar:::
2018-12-19 13:09:39,321 INFO
org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -
--------------------------------------------------------------------------------
2018-12-19 13:09:39,323 INFO
org.apache.flink.runtime.taskexecutor.TaskManagerRunner       - Registered
UNIX signal handlers for [TERM, HUP, INT]
2018-12-19 13:09:39,329 INFO
org.apache.flink.runtime.taskexecutor.TaskManagerRunner       - Maximum
number of open file descriptors is 1048576.
2018-12-19 13:09:39,366 INFO
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: jobmanager.rpc.address, flink-jobmanager
2018-12-19 13:09:39,367 INFO
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: jobmanager.rpc.port, 6123
2018-12-19 13:09:39,367 INFO
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: jobmanager.heap.size, 1024m
2018-12-19 13:09:39,367 INFO
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: taskmanager.heap.size, 1024m
2018-12-19 13:09:39,368 INFO
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: taskmanager.numberOfTaskSlots, 2
2018-12-19 13:09:39,369 INFO
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: parallelism.default, 1
2018-12-19 13:09:39,370 INFO
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: rest.port, 8081
2018-12-19 13:09:39,372 INFO
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: blob.server.port, 6124
2018-12-19 13:09:39,374 INFO
org.apache.flink.configuration.GlobalConfiguration            - Loading
configuration property: query.server.port, 6125
2018-12-19 13:09:39,511 INFO
org.apache.flink.runtime.security.modules.HadoopModule        - Hadoop user
set to flink (auth:SIMPLE)
2018-12-19 13:10:00,708 ERROR
org.apache.flink.runtime.taskexecutor.TaskManagerRunner       - TaskManager
initialization failed.



















*java.net.UnknownHostException: flink-jobmanager: Temporary failure in name
resolution    at java.net.Inet4AddressImpl.lookupAllHostAddr(Native
Method)    at
java.net.InetAddress$2.lookupAllHostAddr(InetAddress.java:929)    at
java.net.InetAddress.getAddressesFromNameService(InetAddress.java:1324)
at java.net.InetAddress.getAllByName0(InetAddress.java:1277)    at
java.net.InetAddress.getAllByName(InetAddress.java:1193)    at
java.net.InetAddress.getAllByName(InetAddress.java:1127)    at
java.net.InetAddress.getByName(InetAddress.java:1077)    at
org.apache.flink.runtime.rpc.akka.AkkaRpcServiceUtils.getRpcUrl(AkkaRpcServiceUtils.java:167)
at
org.apache.flink.runtime.rpc.akka.AkkaRpcServiceUtils.getRpcUrl(AkkaRpcServiceUtils.java:133)
at
org.apache.flink.runtime.highavailability.HighAvailabilityServicesUtils.createHighAvailabilityServices(HighAvailabilityServicesUtils.java:89)
at
org.apache.flink.runtime.taskexecutor.TaskManagerRunner.<init>(TaskManagerRunner.java:127)
at
org.apache.flink.runtime.taskexecutor.TaskManagerRunner.runTaskManager(TaskManagerRunner.java:330)
at
org.apache.flink.runtime.taskexecutor.TaskManagerRunner$1.call(TaskManagerRunner.java:301)
at
org.apache.flink.runtime.taskexecutor.TaskManagerRunner$1.call(TaskManagerRunner.java:298)
at java.security.AccessController.doPrivileged(Native Method)    at
javax.security.auth.Subject.doAs(Subject.java:422)    at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1556)
at
org.apache.flink.runtime.security.HadoopSecurityContext.runSecured(HadoopSecurityContext.java:41)
at
org.apache.flink.runtime.taskexecutor.TaskManagerRunner.main(TaskManagerRunner.java:298)*

Is the minikube ssh... command supposed to mitigate this? (*minikube ssh
'sudo ip link set docker0 promisc on')*
Maybe the minikube ssh.. command didn't get executed properly? Is there a
way to check if it had executed correctly?
Or is it another type of issue?


*Thank you!*

On Wed, 19 Dec 2018 at 12:12, Dawid Wysakowicz <dwysakow...@apache.org>
wrote:

> Hi,
>
> You used a hadoopless docker image, therefore it cannot find hadoop
> dependencies. It is ok if you don't need to use any, the bolded messages
> are just INFO, those are not errors.
>
> Best,
>
> Dawid
> On 19/12/2018 12:58, Alexandru Gutan wrote:
>
> Dear all,
>
> I followed the instructions found here:
> https://ci.apache.org/projects/flink/flink-docs-release-1.7/ops/deployment/kubernetes.html
> Minikube version 0.31-01
> Kubernetes version 1.10
> Flink Docker image: flink:latest (1.7.0-scala_2.12)
>
> I ran the following commands:
>
> minikube start
> minikube ssh 'sudo ip link set docker0 promisc on'
> kubectl create -f jobmanager-deployment.yaml
> kubectl create -f taskmanager-deployment.yaml
> kubectl create -f jobmanager-service.yaml
>
> The 2 taskmanagers fail.
> Output:
>
> Starting Task Manager
> config file:
> jobmanager.rpc.address: flink-jobmanager
> jobmanager.rpc.port: 6123
> jobmanager.heap.size: 1024m
> taskmanager.heap.size: 1024m
> taskmanager.numberOfTaskSlots: 2
> parallelism.default: 1
> rest.port: 8081
> blob.server.port: 6124
> query.server.port: 6125
> Starting taskexecutor as a console application on host
> flink-taskmanager-7679c9d55d-n2trk.
> 2018-12-19 11:42:45,216 INFO
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -
> --------------------------------------------------------------------------------
> 2018-12-19 11:42:45,218 INFO
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -  Starting
> TaskManager (Version: 1.7.0, Rev:49da9f9, Date:28.11.2018 @ 17:59:06 UTC)
> 2018-12-19 11:42:45,218 INFO
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -  OS current
> user: flink
> 2018-12-19 11:42:45,219 INFO
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -  *Current
> Hadoop/Kerberos user: <no hadoop dependency found>*
> 2018-12-19 11:42:45,219 INFO
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -  JVM:
> OpenJDK 64-Bit Server VM - Oracle Corporation - 1.8/25.181-b13
> 2018-12-19 11:42:45,219 INFO
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -  Maximum
> heap size: 922 MiBytes
> 2018-12-19 11:42:45,220 INFO
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -  JAVA_HOME:
> /docker-java-home/jre
> 2018-12-19 11:42:45,220 INFO
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -  No Hadoop
> Dependency available
> 2018-12-19 11:42:45,221 INFO
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -  JVM
> Options:
> 2018-12-19 11:42:45,221 INFO
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -
> -XX:+UseG1GC
> 2018-12-19 11:42:45,221 INFO
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -     -Xms922M
> 2018-12-19 11:42:45,221 INFO
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -     -Xmx922M
> 2018-12-19 11:42:45,221 INFO
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -
> -XX:MaxDirectMemorySize=8388607T
> 2018-12-19 11:42:45,223 INFO
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -
> -Dlog4j.configuration=file:/opt/flink/conf/log4j-console.properties
> 2018-12-19 11:42:45,223 INFO
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -
> -Dlogback.configurationFile=file:/opt/flink/conf/logback-console.xml
> 2018-12-19 11:42:45,223 INFO
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -  Program
> Arguments:
> 2018-12-19 11:42:45,223 INFO
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -
> --configDir
> 2018-12-19 11:42:45,224 INFO
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -
> /opt/flink/conf
> 2018-12-19 11:42:45,224 INFO
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -  Classpath:
> /opt/flink/lib/flink-python_2.12-1.7.0.jar:/opt/flink/lib/log4j-1.2.17.jar:/opt/flink/lib/slf4j-log4j12-1.7.15.jar:/opt/flink/lib/flink-dist_2.12-1.7.0.jar:::
> 2018-12-19 11:42:45,224 INFO
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner       -
> --------------------------------------------------------------------------------
> 2018-12-19 11:42:45,228 INFO
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner       - Registered
> UNIX signal handlers for [TERM, HUP, INT]
> 2018-12-19 11:42:45,233 INFO
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner       - Maximum
> number of open file descriptors is 1048576.
> 2018-12-19 11:42:45,249 INFO
> org.apache.flink.configuration.GlobalConfiguration            - Loading
> configuration property: jobmanager.rpc.address, flink-jobmanager
> 2018-12-19 11:42:45,250 INFO
> org.apache.flink.configuration.GlobalConfiguration            - Loading
> configuration property: jobmanager.rpc.port, 6123
> 2018-12-19 11:42:45,251 INFO
> org.apache.flink.configuration.GlobalConfiguration            - Loading
> configuration property: jobmanager.heap.size, 1024m
> 2018-12-19 11:42:45,251 INFO
> org.apache.flink.configuration.GlobalConfiguration            - Loading
> configuration property: taskmanager.heap.size, 1024m
> 2018-12-19 11:42:45,251 INFO
> org.apache.flink.configuration.GlobalConfiguration            - Loading
> configuration property: taskmanager.numberOfTaskSlots, 2
> 2018-12-19 11:42:45,252 INFO
> org.apache.flink.configuration.GlobalConfiguration            - Loading
> configuration property: parallelism.default, 1
> 2018-12-19 11:42:45,252 INFO
> org.apache.flink.configuration.GlobalConfiguration            - Loading
> configuration property: rest.port, 8081
> 2018-12-19 11:42:45,254 INFO
> org.apache.flink.configuration.GlobalConfiguration            - Loading
> configuration property: blob.server.port, 6124
> 2018-12-19 11:42:45,254 INFO
> org.apache.flink.configuration.GlobalConfiguration            - Loading
> configuration property: query.server.port, 6125
>
>
> *2018-12-19 11:42:45,261 INFO
> org.apache.flink.core.fs.FileSystem                           - Hadoop is
> not in the classpath/dependencies. The extended set of supported File
> Systems via Hadoop is not available. 2018-12-19 11:42:45,282 INFO
> org.apache.flink.runtime.security.modules.HadoopModuleFactory  - Cannot
> create Hadoop Security Module because Hadoop cannot be found in the
> Classpath. 2018-12-19 11:42:45,311 INFO
> org.apache.flink.runtime.security.SecurityUtils               - Cannot
> install HadoopSecurityContext because Hadoop cannot be found in the
> Classpath.*
>
> Any suggestions? Should I try maybe the Hadoop images? (I'm not planning
> to integrate with Hadoop)
>
> Thank you!
>
>

Reply via email to