Hello Zabid, On Wed, Mar 4, 2020 at 1:47 PM Zahid Rahman <zahidr1...@gmail.com> wrote:
> Hi, > > I found the problem was because on my Linux Operating System the > /etc/hostname was blank. > > *STEP 1* > I searched on google the error message and there was an answer suggesting > I should add to /etc/hostname > > 127.0.0.1 [hostname] localhost. > I believe you've confused /etc/hostname and /etc/hosts -- > > I did that but there was still an error, this time the spark log in > standard output was concatenating the text content > of etc/hostname like so , 127.0.0.1[hostname]localhost. > > *STEP 2* > My second attempt was to change the /etc/hostname to 127.0.0.1 > This time I was getting a warning with information about "using loop > back" rather than an error. > > *STEP 3* > I wasn't happy with that so then I changed the /etc/hostname to (see > below) , > then the warning message disappeared. my guess is that it is the act of > creating spark session as to the cause of error, > in SparkConf() API. > > SparkConf sparkConf = new SparkConf() > .setAppName("Simple Application") > .setMaster("local") > .set("spark.executor.memory","2g"); > > $ cat /etc/hostname > # hosts This file describes a number of hostname-to-address > # mappings for the TCP/IP subsystem. It is mostly > # used at boot time, when no name servers are running. > # On small systems, this file can be used instead of a > # "named" name server. > # Syntax: > # > # IP-Address Full-Qualified-Hostname Short-Hostname > # > > 192.168.0.42 > > > > > > > > > > > > > zahid@localhost > :~/Downloads/apachespark/Apache-Spark-Example/Java-Code-Geek> > mvn exec:java -Dexec.mainClass=com.javacodegeek.examples.SparkExampleRDD > -Dexec.args="input.txt" > [INFO] Scanning for projects... > [WARNING] > [WARNING] Some problems were encountered while building the effective > model for javacodegeek:examples:jar:1.0-SNAPSHOT > [WARNING] 'build.plugins.plugin.version' for > org.apache.maven.plugins:maven-compiler-plugin is missing. @ line 12, > column 21 > [WARNING] > [WARNING] It is highly recommended to fix these problems because they > threaten the stability of your build. > [WARNING] > [WARNING] For this reason, future Maven versions might no longer support > building such malformed projects. > [WARNING] > [INFO] > [INFO] -----------------------< javacodegeek:examples > >------------------------ > [INFO] Building examples 1.0-SNAPSHOT > [INFO] --------------------------------[ jar > ]--------------------------------- > [INFO] > [INFO] --- exec-maven-plugin:1.6.0:java (default-cli) @ examples --- > WARNING: An illegal reflective access operation has occurred > WARNING: Illegal reflective access by org.apache.spark.unsafe.Platform > (file:/home/zahid/.m2/repository/org/apache/spark/spark-unsafe_2.12/2.4.5/spark-unsafe_2.12-2.4.5.jar) > to method java.nio.Bits.unaligned() > WARNING: Please consider reporting this to the maintainers of > org.apache.spark.unsafe.Platform > WARNING: Use --illegal-access=warn to enable warnings of further illegal > reflective access operations > WARNING: All illegal access operations will be denied in a future release > Using Spark's default log4j profile: > org/apache/spark/log4j-defaults.properties > 20/02/29 17:20:40 INFO SparkContext: Running Spark version 2.4.5 > 20/02/29 17:20:40 WARN NativeCodeLoader: Unable to load native-hadoop > library for your platform... using builtin-java classes where applicable > 20/02/29 17:20:41 INFO SparkContext: Submitted application: Word Count > 20/02/29 17:20:41 INFO SecurityManager: Changing view acls to: zahid > 20/02/29 17:20:41 INFO SecurityManager: Changing modify acls to: zahid > 20/02/29 17:20:41 INFO SecurityManager: Changing view acls groups to: > 20/02/29 17:20:41 INFO SecurityManager: Changing modify acls groups to: > 20/02/29 17:20:41 INFO SecurityManager: SecurityManager: authentication > disabled; ui acls disabled; users with view permissions: Set(zahid); > groups with view permissions: Set(); users with modify permissions: > Set(zahid); groups with modify permissions: Set() > 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:20:41 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:20:41 ERROR SparkContext: Error initializing SparkContext. > java.net.BindException: Cannot assign requested address: Service > 'sparkDriver' failed after 16 retries (on a random free port)! Consider > explicitly setting the appropriate binding address for the service > 'sparkDriver' (for example spark.driver.bindAddress for SparkDriver) to the > correct binding address. > at java.base/sun.nio.ch.Net.bind0(Native Method) > at java.base/sun.nio.ch.Net.bind(Net.java:469) > at java.base/sun.nio.ch.Net.bind(Net.java:458) > at > java.base/sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:220) > at > io.netty.channel.socket.nio.NioServerSocketChannel.doBind(NioServerSocketChannel.java:132) > at > io.netty.channel.AbstractChannel$AbstractUnsafe.bind(AbstractChannel.java:551) > at > io.netty.channel.DefaultChannelPipeline$HeadContext.bind(DefaultChannelPipeline.java:1346) > at > io.netty.channel.AbstractChannelHandlerContext.invokeBind(AbstractChannelHandlerContext.java:503) > at > io.netty.channel.AbstractChannelHandlerContext.bind(AbstractChannelHandlerContext.java:488) > at > io.netty.channel.DefaultChannelPipeline.bind(DefaultChannelPipeline.java:985) > at io.netty.channel.AbstractChannel.bind(AbstractChannel.java:247) > at > io.netty.bootstrap.AbstractBootstrap$2.run(AbstractBootstrap.java:344) > at > io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:163) > at > io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:510) > at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:518) > at > io.netty.util.concurrent.SingleThreadEventExecutor$6.run(SingleThreadEventExecutor.java:1044) > at > io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) > at > io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) > at java.base/java.lang.Thread.run(Thread.java:830) > 20/02/29 17:20:41 INFO SparkContext: Successfully stopped SparkContext > [WARNING] > java.net.BindException: Cannot assign requested address: Service > 'sparkDriver' failed after 16 retries (on a random free port)! Consider > explicitly setting the appropriate binding address for the service > 'sparkDriver' (for example spark.driver.bindAddress for SparkDriver) to the > correct binding address. > at sun.nio.ch.Net.bind0 (Native Method) > at sun.nio.ch.Net.bind (Net.java:469) > at sun.nio.ch.Net.bind (Net.java:458) > at sun.nio.ch.ServerSocketChannelImpl.bind > (ServerSocketChannelImpl.java:220) > at io.netty.channel.socket.nio.NioServerSocketChannel.doBind > (NioServerSocketChannel.java:132) > at io.netty.channel.AbstractChannel$AbstractUnsafe.bind > (AbstractChannel.java:551) > at io.netty.channel.DefaultChannelPipeline$HeadContext.bind > (DefaultChannelPipeline.java:1346) > at io.netty.channel.AbstractChannelHandlerContext.invokeBind > (AbstractChannelHandlerContext.java:503) > at io.netty.channel.AbstractChannelHandlerContext.bind > (AbstractChannelHandlerContext.java:488) > at io.netty.channel.DefaultChannelPipeline.bind > (DefaultChannelPipeline.java:985) > at io.netty.channel.AbstractChannel.bind (AbstractChannel.java:247) > at io.netty.bootstrap.AbstractBootstrap$2.run > (AbstractBootstrap.java:344) > at io.netty.util.concurrent.AbstractEventExecutor.safeExecute > (AbstractEventExecutor.java:163) > at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks > (SingleThreadEventExecutor.java:510) > at io.netty.channel.nio.NioEventLoop.run (NioEventLoop.java:518) > at io.netty.util.concurrent.SingleThreadEventExecutor$6.run > (SingleThreadEventExecutor.java:1044) > at io.netty.util.internal.ThreadExecutorMap$2.run > (ThreadExecutorMap.java:74) > at io.netty.util.concurrent.FastThreadLocalRunnable.run > (FastThreadLocalRunnable.java:30) > at java.lang.Thread.run (Thread.java:830) > [INFO] > ------------------------------------------------------------------------ > [INFO] BUILD FAILURE > [INFO] > ------------------------------------------------------------------------ > [INFO] Total time: 3.977 s > [INFO] Finished at: 2020-02-29T17:20:43Z > [INFO] > ------------------------------------------------------------------------ > [ERROR] Failed to execute goal > org.codehaus.mojo:exec-maven-plugin:1.6.0:java (default-cli) on project > examples: An exception occured while executing the Java class. Cannot > assign requested address: Service 'sparkDriver' failed after 16 retries (on > a random free port)! Consider explicitly setting the appropriate binding > address for the service 'sparkDriver' (for example spark.driver.bindAddress > for SparkDriver) to the correct binding address. -> [Help 1] > [ERROR] > [ERROR] To see the full stack trace of the errors, re-run Maven with the > -e switch. > [ERROR] Re-run Maven using the -X switch to enable full debug logging. > [ERROR] > [ERROR] For more information about the errors and possible solutions, > please read the following articles: > [ERROR] [Help 1] > http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException > zahid@localhost:~/Downloads/apachespark/Apache-Spark-Example/Java-Code-Geek> > mvn exec:java -Dexec.mainClass=com.javacodegeek.examples.SparkExampleRDD > -Dexec.args="input.txt" > log.out > WARNING: An illegal reflective access operation has occurred > WARNING: Illegal reflective access by org.apache.spark.unsafe.Platform > (file:/home/zahid/.m2/repository/org/apache/spark/spark-unsafe_2.12/2.4.5/spark-unsafe_2.12-2.4.5.jar) > to method java.nio.Bits.unaligned() > WARNING: Please consider reporting this to the maintainers of > org.apache.spark.unsafe.Platform > WARNING: Use --illegal-access=warn to enable warnings of further illegal > reflective access operations > WARNING: All illegal access operations will be denied in a future release > Using Spark's default log4j profile: > org/apache/spark/log4j-defaults.properties > 20/02/29 17:21:12 INFO SparkContext: Running Spark version 2.4.5 > 20/02/29 17:21:12 WARN NativeCodeLoader: Unable to load native-hadoop > library for your platform... using builtin-java classes where applicable > 20/02/29 17:21:12 INFO SparkContext: Submitted application: Word Count > 20/02/29 17:21:12 INFO SecurityManager: Changing view acls to: zahid > 20/02/29 17:21:12 INFO SecurityManager: Changing modify acls to: zahid > 20/02/29 17:21:12 INFO SecurityManager: Changing view acls groups to: > 20/02/29 17:21:12 INFO SecurityManager: Changing modify acls groups to: > 20/02/29 17:21:12 INFO SecurityManager: SecurityManager: authentication > disabled; ui acls disabled; users with view permissions: Set(zahid); > groups with view permissions: Set(); users with modify permissions: > Set(zahid); groups with modify permissions: Set() > 20/02/29 17:21:12 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:21:12 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:21:12 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:21:12 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:21:12 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:21:13 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:21:13 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:21:13 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:21:13 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:21:13 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:21:13 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:21:13 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:21:13 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:21:13 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:21:13 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:21:13 WARN Utils: Service 'sparkDriver' could not bind on a > random free port. You may check whether configuring an appropriate binding > address. > 20/02/29 17:21:13 ERROR SparkContext: Error initializing SparkContext. > java.net.BindException: Cannot assign requested address: Service > 'sparkDriver' failed after 16 retries (on a random free port)! Consider > explicitly setting the appropriate binding address for the service > 'sparkDriver' (for example spark.driver.bindAddress for SparkDriver) to the > correct binding address. > at java.base/sun.nio.ch.Net.bind0(Native Method) > at java.base/sun.nio.ch.Net.bind(Net.java:469) > at java.base/sun.nio.ch.Net.bind(Net.java:458) > at > java.base/sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:220) > at > io.netty.channel.socket.nio.NioServerSocketChannel.doBind(NioServerSocketChannel.java:132) > at > io.netty.channel.AbstractChannel$AbstractUnsafe.bind(AbstractChannel.java:551) > at > io.netty.channel.DefaultChannelPipeline$HeadContext.bind(DefaultChannelPipeline.java:1346) > at > io.netty.channel.AbstractChannelHandlerContext.invokeBind(AbstractChannelHandlerContext.java:503) > at > io.netty.channel.AbstractChannelHandlerContext.bind(AbstractChannelHandlerContext.java:488) > at > io.netty.channel.DefaultChannelPipeline.bind(DefaultChannelPipeline.java:985) > at io.netty.channel.AbstractChannel.bind(AbstractChannel.java:247) > at > io.netty.bootstrap.AbstractBootstrap$2.run(AbstractBootstrap.java:344) > at > io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:163) > at > io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:510) > at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:518) > at > io.netty.util.concurrent.SingleThreadEventExecutor$6.run(SingleThreadEventExecutor.java:1044) > at > io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) > at > io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) > at java.base/java.lang.Thread.run(Thread.java:830) > 20/02/29 17:21:13 INFO SparkContext: Successfully stopped SparkContext > zahid@localhost:~/Downloads/apachespark/Apache-Spark-Example/Java-Code-Geek> > ls > input.txt log.out pom.xml src target > zahid@localhost:~/Downloads/apachespark/Apache-Spark-Example/Java-Code-Geek> > cat log.out > [INFO] Scanning for projects... > [WARNING] > [WARNING] Some problems were encountered while building the effective > model for javacodegeek:examples:jar:1.0-SNAPSHOT > [WARNING] 'build.plugins.plugin.version' for > org.apache.maven.plugins:maven-compiler-plugin is missing. @ line 12, > column 21 > [WARNING] > [WARNING] It is highly recommended to fix these problems because they > threaten the stability of your build. > [WARNING] > [WARNING] For this reason, future Maven versions might no longer support > building such malformed projects. > [WARNING] > [INFO] > [INFO] -----------------------< javacodegeek:examples > >------------------------ > [INFO] Building examples 1.0-SNAPSHOT > [INFO] --------------------------------[ jar > ]--------------------------------- > [INFO] > [INFO] --- exec-maven-plugin:1.6.0:java (default-cli) @ examples --- > [WARNING] > java.net.BindException: Cannot assign requested address: Service > 'sparkDriver' failed after 16 retries (on a random free port)! Consider > explicitly setting the appropriate binding address for the service > 'sparkDriver' (for example spark.driver.bindAddress for SparkDriver) to the > correct binding address. > at sun.nio.ch.Net.bind0 (Native Method) > at sun.nio.ch.Net.bind (Net.java:469) > at sun.nio.ch.Net.bind (Net.java:458) > at sun.nio.ch.ServerSocketChannelImpl.bind > (ServerSocketChannelImpl.java:220) > at io.netty.channel.socket.nio.NioServerSocketChannel.doBind > (NioServerSocketChannel.java:132) > at io.netty.channel.AbstractChannel$AbstractUnsafe.bind > (AbstractChannel.java:551) > at io.netty.channel.DefaultChannelPipeline$HeadContext.bind > (DefaultChannelPipeline.java:1346) > at io.netty.channel.AbstractChannelHandlerContext.invokeBind > (AbstractChannelHandlerContext.java:503) > at io.netty.channel.AbstractChannelHandlerContext.bind > (AbstractChannelHandlerContext.java:488) > at io.netty.channel.DefaultChannelPipeline.bind > (DefaultChannelPipeline.java:985) > at io.netty.channel.AbstractChannel.bind (AbstractChannel.java:247) > at io.netty.bootstrap.AbstractBootstrap$2.run > (AbstractBootstrap.java:344) > at io.netty.util.concurrent.AbstractEventExecutor.safeExecute > (AbstractEventExecutor.java:163) > at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks > (SingleThreadEventExecutor.java:510) > at io.netty.channel.nio.NioEventLoop.run (NioEventLoop.java:518) > at io.netty.util.concurrent.SingleThreadEventExecutor$6.run > (SingleThreadEventExecutor.java:1044) > at io.netty.util.internal.ThreadExecutorMap$2.run > (ThreadExecutorMap.java:74) > at io.netty.util.concurrent.FastThreadLocalRunnable.run > (FastThreadLocalRunnable.java:30) > at java.lang.Thread.run (Thread.java:830) > [INFO] > ------------------------------------------------------------------------ > [INFO] BUILD FAILURE > [INFO] > ------------------------------------------------------------------------ > [INFO] Total time: 4.949 s > [INFO] Finished at: 2020-02-29T17:21:16Z > [INFO] > ------------------------------------------------------------------------ > [ERROR] Failed to execute goal > org.codehaus.mojo:exec-maven-plugin:1.6.0:java (default-cli) on project > examples: An exception occured while executing the Java class. Cannot > assign requested address: Service 'sparkDriver' failed after 16 retries (on > a random free port)! Consider explicitly setting the appropriate binding > address for the service 'sparkDriver' (for example spark.driver.bindAddress > for SparkDriver) to the correct binding address. -> [Help 1] > [ERROR] > [ERROR] To see the full stack trace of the errors, re-run Maven with the > -e switch. > [ERROR] Re-run Maven using the -X switch to enable full debug logging. > [ERROR] > [ERROR] For more information about the errors and possible solutions, > please read the following articles: > [ERROR] [Help 1] > http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionExc: > BUGeption > <http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException> > zahid@localhost:~/Downloads/apachespark/Apache-Spark-Example/Java-Code-Geek> > > > > Backbutton.co.uk > ¯\_(ツ)_/¯ > ♡۶Java♡۶RMI ♡۶ > Make Use Method {MUM} > makeuse.org > <http://www.backbutton.co.uk> >