Hello! I guess you should add VM options:
--add-exports=java.base/jdk.internal.misc=ALL-UNNAMED--add-exports=java.base/sun.nio.ch=ALL-UNNAMED As per running under Java 9. In case of your IDE, please specify JVM as noted. Regard, -- Ilya Kasnacheev сб, 5 янв. 2019 г. в 11:59, mehdi sey <[email protected]>: > hi, i have a code for writing into ignite rdd. this program read data from > spark rdd and catch it on ignite rdd. i run it with command line in Linux > Ubuntu but in the middle of execution i have encounter with below error. i > checked in spark UI for watching if job complete or not but the job is not > complete and failed. why? i have attached piece of code that i have wrote > and run with command. > > $SPARK_HOME/bin/spark-submit --class "com.gridgain.RDDWriter" --master > spark://linux-client:7077 ~/spark\ and\ ignite\ > issue/ignite-and-spark-integration-master/ignite-rdd/ignite-spark-scala/target/ignite-spark-scala-1.0.jar > > 2019-01-05 11:47:02 WARN Utils:66 - Your hostname, linux-client resolves > to > a loopback address: 127.0.1.1, but we couldn't find any external IP > address! > 2019-01-05 11:47:02 WARN Utils:66 - Set SPARK_LOCAL_IP if you need to bind > to another address > 2019-01-05 11:47:03 WARN NativeCodeLoader:62 - Unable to load > native-hadoop > library for your platform... using builtin-java classes where applicable > 2019-01-05 11:47:03 INFO SparkContext:54 - Running Spark version 2.4.0 > 2019-01-05 11:47:03 INFO SparkContext:54 - Submitted application: > RDDWriter > 2019-01-05 11:47:03 INFO SecurityManager:54 - Changing view acls to: mehdi > 2019-01-05 11:47:03 INFO SecurityManager:54 - Changing modify acls to: > mehdi > 2019-01-05 11:47:03 INFO SecurityManager:54 - Changing view acls groups > to: > 2019-01-05 11:47:03 INFO SecurityManager:54 - Changing modify acls groups > to: > 2019-01-05 11:47:03 INFO SecurityManager:54 - SecurityManager: > authentication disabled; ui acls disabled; users with view permissions: > Set(mehdi); groups with view permissions: Set(); users with modify > permissions: Set(mehdi); groups with modify permissions: Set() > 2019-01-05 11:47:03 WARN MacAddressUtil:136 - Failed to find a usable > hardware address from the network interfaces; using random bytes: > 88:26:00:23:5d:50:a0:61 > 2019-01-05 11:47:03 INFO Utils:54 - Successfully started service > 'sparkDriver' on port 36233. > 2019-01-05 11:47:03 INFO SparkEnv:54 - Registering MapOutputTracker > 2019-01-05 11:47:03 INFO SparkEnv:54 - Registering BlockManagerMaster > 2019-01-05 11:47:03 INFO BlockManagerMasterEndpoint:54 - Using > org.apache.spark.storage.DefaultTopologyMapper for getting topology > information > 2019-01-05 11:47:03 INFO BlockManagerMasterEndpoint:54 - > BlockManagerMasterEndpoint up > 2019-01-05 11:47:03 INFO DiskBlockManager:54 - Created local directory at > /tmp/blockmgr-6e47832e-855a-4305-a293-662379733b7f > 2019-01-05 11:47:03 INFO MemoryStore:54 - MemoryStore started with > capacity > 366.3 MB > 2019-01-05 11:47:03 INFO SparkEnv:54 - Registering OutputCommitCoordinator > 2019-01-05 11:47:03 INFO log:192 - Logging initialized @2024ms > 2019-01-05 11:47:04 INFO Server:351 - jetty-9.3.z-SNAPSHOT, build > timestamp: unknown, git hash: unknown > 2019-01-05 11:47:04 INFO Server:419 - Started @2108ms > 2019-01-05 11:47:04 INFO AbstractConnector:278 - Started > ServerConnector@5ba745bc{HTTP/1.1,[http/1.1]}{0.0.0.0:4040} > 2019-01-05 11:47:04 INFO Utils:54 - Successfully started service 'SparkUI' > on port 4040. > 2019-01-05 11:47:04 INFO ContextHandler:781 - Started > o.s.j.s.ServletContextHandler@606fc505{/jobs,null,AVAILABLE,@Spark} > 2019-01-05 11:47:04 INFO ContextHandler:781 - Started > o.s.j.s.ServletContextHandler@2c30b71f{/jobs/json,null,AVAILABLE,@Spark} > 2019-01-05 11:47:04 INFO ContextHandler:781 - Started > o.s.j.s.ServletContextHandler@1d81e101{/jobs/job,null,AVAILABLE,@Spark} > 2019-01-05 11:47:04 INFO ContextHandler:781 - Started > o.s.j.s.ServletContextHandler@bf71cec > {/jobs/job/json,null,AVAILABLE,@Spark} > 2019-01-05 11:47:04 INFO ContextHandler:781 - Started > o.s.j.s.ServletContextHandler@22d6cac2{/stages,null,AVAILABLE,@Spark} > 2019-01-05 11:47:04 INFO ContextHandler:781 - Started > o.s.j.s.ServletContextHandler@30cdae70{/stages/json,null,AVAILABLE,@Spark} > 2019-01-05 11:47:04 INFO ContextHandler:781 - Started > o.s.j.s.ServletContextHandler@1654a892 > {/stages/stage,null,AVAILABLE,@Spark} > 2019-01-05 11:47:04 INFO ContextHandler:781 - Started > o.s.j.s.ServletContextHandler@6c000e0c > {/stages/stage/json,null,AVAILABLE,@Spark} > 2019-01-05 11:47:04 INFO ContextHandler:781 - Started > o.s.j.s.ServletContextHandler@5f233b26{/stages/pool,null,AVAILABLE,@Spark} > 2019-01-05 11:47:04 INFO ContextHandler:781 - Started > o.s.j.s.ServletContextHandler@44f9779c > {/stages/pool/json,null,AVAILABLE,@Spark} > 2019-01-05 11:47:04 INFO ContextHandler:781 - Started > o.s.j.s.ServletContextHandler@6974a715{/storage,null,AVAILABLE,@Spark} > 2019-01-05 11:47:04 INFO ContextHandler:781 - Started > o.s.j.s.ServletContextHandler@5e8a459{/storage/json,null,AVAILABLE,@Spark} > 2019-01-05 11:47:04 INFO ContextHandler:781 - Started > o.s.j.s.ServletContextHandler@43d455c9{/storage/rdd,null,AVAILABLE,@Spark} > 2019-01-05 11:47:04 INFO ContextHandler:781 - Started > o.s.j.s.ServletContextHandler@4c9e9fb8 > {/storage/rdd/json,null,AVAILABLE,@Spark} > 2019-01-05 11:47:04 INFO ContextHandler:781 - Started > o.s.j.s.ServletContextHandler@9ec531{/environment,null,AVAILABLE,@Spark} > 2019-01-05 11:47:04 INFO ContextHandler:781 - Started > o.s.j.s.ServletContextHandler@40147317 > {/environment/json,null,AVAILABLE,@Spark} > 2019-01-05 11:47:04 INFO ContextHandler:781 - Started > o.s.j.s.ServletContextHandler@210f0cc1{/executors,null,AVAILABLE,@Spark} > 2019-01-05 11:47:04 INFO ContextHandler:781 - Started > o.s.j.s.ServletContextHandler@19542407 > {/executors/json,null,AVAILABLE,@Spark} > 2019-01-05 11:47:04 INFO ContextHandler:781 - Started > o.s.j.s.ServletContextHandler@6f95cd51 > {/executors/threadDump,null,AVAILABLE,@Spark} > 2019-01-05 11:47:04 INFO ContextHandler:781 - Started > o.s.j.s.ServletContextHandler@c7a977f > {/executors/threadDump/json,null,AVAILABLE,@Spark} > 2019-01-05 11:47:04 INFO ContextHandler:781 - Started > o.s.j.s.ServletContextHandler@55caeb35{/static,null,AVAILABLE,@Spark} > 2019-01-05 11:47:04 INFO ContextHandler:781 - Started > o.s.j.s.ServletContextHandler@3ae66c85{/,null,AVAILABLE,@Spark} > 2019-01-05 11:47:04 INFO ContextHandler:781 - Started > o.s.j.s.ServletContextHandler@16943e88{/api,null,AVAILABLE,@Spark} > 2019-01-05 11:47:04 INFO ContextHandler:781 - Started > o.s.j.s.ServletContextHandler@3961a41a > {/jobs/job/kill,null,AVAILABLE,@Spark} > 2019-01-05 11:47:04 INFO ContextHandler:781 - Started > o.s.j.s.ServletContextHandler@5a4ed68f > {/stages/stage/kill,null,AVAILABLE,@Spark} > 2019-01-05 11:47:04 INFO SparkUI:54 - Bound SparkUI to 0.0.0.0, and > started > at http://linux-client:4040 > 2019-01-05 11:47:04 INFO SparkContext:54 - Added JAR > > file:/home/mehdi/spark%20and%20ignite%20issue/ignite-and-spark-integration-master/ignite-rdd/ignite-spark-scala/target/ignite-spark-scala-1.0.jar > at spark://linux-client:36233/jars/ignite-spark-scala-1.0.jar with > timestamp > 1546676224149 > 2019-01-05 11:47:04 INFO StandaloneAppClient$ClientEndpoint:54 - > Connecting > to master spark://linux-client:7077... > 2019-01-05 11:47:04 INFO TransportClientFactory:267 - Successfully created > connection to linux-client/127.0.1.1:7077 after 40 ms (0 ms spent in > bootstraps) > 2019-01-05 11:47:04 INFO StandaloneSchedulerBackend:54 - Connected to > Spark > cluster with app ID app-20190105114704-0003 > 2019-01-05 11:47:04 INFO StandaloneAppClient$ClientEndpoint:54 - Executor > added: app-20190105114704-0003/0 on worker-20190105103259-127.0.1.1-43911 > (127.0.1.1:43911) with 2 core(s) > 2019-01-05 11:47:04 INFO StandaloneSchedulerBackend:54 - Granted executor > ID app-20190105114704-0003/0 on hostPort 127.0.1.1:43911 with 2 core(s), > 512.0 MB RAM > 2019-01-05 11:47:04 INFO StandaloneAppClient$ClientEndpoint:54 - Executor > added: app-20190105114704-0003/1 on worker-20190105103304-127.0.1.1-44569 > (127.0.1.1:44569) with 2 core(s) > 2019-01-05 11:47:04 INFO StandaloneSchedulerBackend:54 - Granted executor > ID app-20190105114704-0003/1 on hostPort 127.0.1.1:44569 with 2 core(s), > 512.0 MB RAM > 2019-01-05 11:47:04 INFO StandaloneAppClient$ClientEndpoint:54 - Executor > added: app-20190105114704-0003/2 on worker-20190105103301-127.0.1.1-34465 > (127.0.1.1:34465) with 2 core(s) > 2019-01-05 11:47:04 INFO StandaloneSchedulerBackend:54 - Granted executor > ID app-20190105114704-0003/2 on hostPort 127.0.1.1:34465 with 2 core(s), > 512.0 MB RAM > 2019-01-05 11:47:04 INFO StandaloneAppClient$ClientEndpoint:54 - Executor > added: app-20190105114704-0003/3 on worker-20190105103256-127.0.1.1-46653 > (127.0.1.1:46653) with 2 core(s) > 2019-01-05 11:47:04 INFO StandaloneSchedulerBackend:54 - Granted executor > ID app-20190105114704-0003/3 on hostPort 127.0.1.1:46653 with 2 core(s), > 512.0 MB RAM > 2019-01-05 11:47:04 INFO Utils:54 - Successfully started service > 'org.apache.spark.network.netty.NettyBlockTransferService' on port 41343. > 2019-01-05 11:47:04 INFO NettyBlockTransferService:54 - Server created on > linux-client:41343 > 2019-01-05 11:47:04 INFO BlockManager:54 - Using > org.apache.spark.storage.RandomBlockReplicationPolicy for block replication > policy > 2019-01-05 11:47:04 INFO StandaloneAppClient$ClientEndpoint:54 - Executor > updated: app-20190105114704-0003/2 is now RUNNING > 2019-01-05 11:47:04 INFO StandaloneAppClient$ClientEndpoint:54 - Executor > updated: app-20190105114704-0003/1 is now RUNNING > 2019-01-05 11:47:04 INFO StandaloneAppClient$ClientEndpoint:54 - Executor > updated: app-20190105114704-0003/3 is now RUNNING > 2019-01-05 11:47:04 INFO StandaloneAppClient$ClientEndpoint:54 - Executor > updated: app-20190105114704-0003/0 is now RUNNING > 2019-01-05 11:47:04 INFO BlockManagerMaster:54 - Registering BlockManager > BlockManagerId(driver, linux-client, 41343, None) > 2019-01-05 11:47:04 INFO BlockManagerMasterEndpoint:54 - Registering block > manager linux-client:41343 with 366.3 MB RAM, BlockManagerId(driver, > linux-client, 41343, None) > 2019-01-05 11:47:04 INFO BlockManagerMaster:54 - Registered BlockManager > BlockManagerId(driver, linux-client, 41343, None) > 2019-01-05 11:47:04 INFO BlockManager:54 - Initialized BlockManager: > BlockManagerId(driver, linux-client, 41343, None) > 2019-01-05 11:47:04 INFO ContextHandler:781 - Started > o.s.j.s.ServletContextHandler@70e02081 > {/metrics/json,null,AVAILABLE,@Spark} > 2019-01-05 11:47:05 INFO EventLoggingListener:54 - Logging events to > file:/tmp/spark-events/app-20190105114704-0003 > 2019-01-05 11:47:05 INFO StandaloneSchedulerBackend:54 - SchedulerBackend > is ready for scheduling beginning after reached > minRegisteredResourcesRatio: > 0.0 > 2019-01-05 11:47:06 INFO XmlBeanDefinitionReader:317 - Loading XML bean > definitions from URL > > [file:/usr/local/apache-ignite-fabric-2.6.0-bin/examples/config/spark/example-shared-rdd.xml] > 2019-01-05 11:47:06 INFO GenericApplicationContext:583 - Refreshing > org.springframework.context.support.GenericApplicationContext@2ec3633f: > startup date [Sat Jan 05 11:47:06 IRST 2019]; root of context hierarchy > Can't load log handler > "org.apache.ignite.logger.java.JavaLoggerFileHandler" > java.lang.ClassNotFoundException: > org.apache.ignite.logger.java.JavaLoggerFileHandler > java.lang.ClassNotFoundException: > org.apache.ignite.logger.java.JavaLoggerFileHandler > at java.net.URLClassLoader.findClass(URLClassLoader.java:381) > at java.lang.ClassLoader.loadClass(ClassLoader.java:424) > at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:349) > at java.lang.ClassLoader.loadClass(ClassLoader.java:357) > at java.util.logging.LogManager$5.run(LogManager.java:965) > at java.security.AccessController.doPrivileged(Native Method) > at > java.util.logging.LogManager.loadLoggerHandlers(LogManager.java:958) > at > java.util.logging.LogManager.initializeGlobalHandlers(LogManager.java:1578) > at java.util.logging.LogManager.access$1500(LogManager.java:145) > at > > java.util.logging.LogManager$RootLogger.accessCheckedHandlers(LogManager.java:1667) > at java.util.logging.Logger.getHandlers(Logger.java:1777) > at > org.apache.ignite.logger.java.JavaLogger.findHandler(JavaLogger.java:411) > at > org.apache.ignite.logger.java.JavaLogger.configure(JavaLogger.java:241) > at > org.apache.ignite.logger.java.JavaLogger.<init>(JavaLogger.java:181) > at > org.apache.ignite.logger.java.JavaLogger.<init>(JavaLogger.java:135) > at > > org.apache.ignite.internal.LongJVMPauseDetector.<clinit>(LongJVMPauseDetector.java:44) > at > org.apache.ignite.internal.IgniteKernal.<clinit>(IgniteKernal.java:300) > at > > org.apache.ignite.internal.IgnitionEx$IgniteNamedInstance.start0(IgnitionEx.java:2009) > at > > org.apache.ignite.internal.IgnitionEx$IgniteNamedInstance.start(IgnitionEx.java:1723) > at > org.apache.ignite.internal.IgnitionEx.start0(IgnitionEx.java:1151) > at org.apache.ignite.internal.IgnitionEx.start(IgnitionEx.java:671) > at org.apache.ignite.internal.IgnitionEx.start(IgnitionEx.java:611) > at org.apache.ignite.Ignition.getOrStart(Ignition.java:419) > at > org.apache.ignite.spark.IgniteContext.ignite(IgniteContext.scala:150) > at > org.apache.ignite.spark.IgniteContext.<init>(IgniteContext.scala:63) > at > org.apache.ignite.spark.IgniteContext.<init>(IgniteContext.scala:99) > at > > com.gridgain.RDDWriter$.delayedEndpoint$com$gridgain$RDDWriter$1(SparkIgniteTest.scala:26) > at > com.gridgain.RDDWriter$delayedInit$body.apply(SparkIgniteTest.scala:23) > at scala.Function0$class.apply$mcV$sp(Function0.scala:34) > at > scala.runtime.AbstractFunction0.apply$mcV$sp(AbstractFunction0.scala:12) > at scala.App$$anonfun$main$1.apply(App.scala:76) > at scala.App$$anonfun$main$1.apply(App.scala:76) > at scala.collection.immutable.List.foreach(List.scala:392) > at > > scala.collection.generic.TraversableForwarder$class.foreach(TraversableForwarder.scala:35) > at scala.App$class.main(App.scala:76) > at com.gridgain.RDDWriter$.main(SparkIgniteTest.scala:23) > at com.gridgain.RDDWriter.main(SparkIgniteTest.scala) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at > > org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52) > at > org.apache.spark.deploy.SparkSubmit.org > $apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:849) > at > org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:167) > at > org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:195) > at > org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:86) > at > org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:924) > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:933) > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > 2019-01-05 11:47:07 INFO IgniteKernal:566 - > > >>> __________ ________________ > >>> / _/ ___/ |/ / _/_ __/ __/ > >>> _/ // (7 7 // / / / / _/ > >>> /___/\___/_/|_/___/ /_/ /___/ > >>> > >>> ver. 2.6.0#19700101-sha1:DEV > >>> 2018 Copyright(C) Apache Software Foundation > >>> > >>> Ignite documentation: http://ignite.apache.org > > 2019-01-05 11:47:07 INFO IgniteKernal:566 - Config URL: n/a > 2019-01-05 11:47:07 INFO IgniteKernal:566 - IgniteConfiguration > [igniteInstanceName=null, pubPoolSize=8, svcPoolSize=8, callbackPoolSize=8, > stripedPoolSize=8, sysPoolSize=8, mgmtPoolSize=4, igfsPoolSize=8, > dataStreamerPoolSize=8, utilityCachePoolSize=8, > utilityCacheKeepAliveTime=60000, p2pPoolSize=2, qryPoolSize=8, > igniteHome=/usr/local/apache-ignite-fabric-2.6.0-bin/, > igniteWorkDir=/usr/local/apache-ignite-fabric-2.6.0-bin/work, > mbeanSrv=com.sun.jmx.mbeanserver.JmxMBeanServer@76b47204, > nodeId=f512a374-f6b6-4918-9790-6183a43f8064, > marsh=org.apache.ignite.internal.binary.BinaryMarshaller@1f1cddf3, > marshLocJobs=false, daemon=false, p2pEnabled=false, netTimeout=5000, > sndRetryDelay=1000, sndRetryCnt=3, metricsHistSize=10000, > metricsUpdateFreq=2000, metricsExpTime=9223372036854775807, > discoSpi=TcpDiscoverySpi [addrRslvr=null, sockTimeout=0, ackTimeout=0, > marsh=null, reconCnt=10, reconDelay=2000, maxAckTimeout=600000, > forceSrvMode=false, clientReconnectDisabled=false, internalLsnr=null], > segPlc=STOP, segResolveAttempts=2, waitForSegOnStart=true, > allResolversPassReq=true, segChkFreq=10000, commSpi=TcpCommunicationSpi > [connectGate=null, connPlc=null, enableForcibleNodeKill=false, > enableTroubleshootingLog=false, > > srvLsnr=org.apache.ignite.spi.communication.tcp.TcpCommunicationSpi$2@3fd05b3e > , > locAddr=null, locHost=null, locPort=47100, locPortRange=100, shmemPort=-1, > directBuf=true, directSndBuf=false, idleConnTimeout=600000, > connTimeout=5000, maxConnTimeout=600000, reconCnt=10, sockSndBuf=32768, > sockRcvBuf=32768, msgQueueLimit=0, slowClientQueueLimit=0, nioSrvr=null, > shmemSrv=null, usePairedConnections=false, connectionsPerNode=1, > tcpNoDelay=true, filterReachableAddresses=false, ackSndThreshold=32, > unackedMsgsBufSize=0, sockWriteTimeout=2000, lsnr=null, boundTcpPort=-1, > boundTcpShmemPort=-1, selectorsCnt=4, selectorSpins=0, addrRslvr=null, > ctxInitLatch=java.util.concurrent.CountDownLatch@4eb9f2af[Count = 1], > stopping=false, > > metricsLsnr=org.apache.ignite.spi.communication.tcp.TcpCommunicationMetricsListener@4a481728 > ], > evtSpi=org.apache.ignite.spi.eventstorage.NoopEventStorageSpi@4ace284d, > colSpi=NoopCollisionSpi [], deploySpi=LocalDeploymentSpi [lsnr=null], > indexingSpi=org.apache.ignite.spi.indexing.noop.NoopIndexingSpi@64deb58f, > addrRslvr=null, clientMode=true, rebalanceThreadPoolSize=1, > txCfg=org.apache.ignite.configuration.TransactionConfiguration@2b1cd7bc, > cacheSanityCheckEnabled=true, discoStartupDelay=60000, deployMode=SHARED, > p2pMissedCacheSize=100, locHost=null, timeSrvPortBase=31100, > timeSrvPortRange=100, failureDetectionTimeout=10000, > clientFailureDetectionTimeout=30000, metricsLogFreq=60000, hadoopCfg=null, > > connectorCfg=org.apache.ignite.configuration.ConnectorConfiguration@33379242 > , > odbcCfg=null, warmupClos=null, atomicCfg=AtomicConfiguration > [seqReserveSize=1000, cacheMode=PARTITIONED, backups=1, aff=null, > grpName=null], classLdr=null, sslCtxFactory=null, platformCfg=null, > binaryCfg=null, memCfg=null, pstCfg=null, dsCfg=null, activeOnStart=true, > autoActivation=true, longQryWarnTimeout=3000, sqlConnCfg=null, > cliConnCfg=ClientConnectorConfiguration [host=null, port=10800, > portRange=100, sockSndBufSize=0, sockRcvBufSize=0, tcpNoDelay=true, > maxOpenCursorsPerConn=128, threadPoolSize=8, idleTimeout=0, > jdbcEnabled=true, odbcEnabled=true, thinCliEnabled=true, sslEnabled=false, > useIgniteSslCtxFactory=true, sslClientAuth=false, sslCtxFactory=null], > authEnabled=false, failureHnd=null, commFailureRslvr=null] > 2019-01-05 11:47:07 INFO IgniteKernal:566 - Daemon mode: off > 2019-01-05 11:47:07 INFO IgniteKernal:566 - OS: Linux 4.15.0-43-generic > amd64 > 2019-01-05 11:47:07 INFO IgniteKernal:566 - OS user: mehdi > 2019-01-05 11:47:07 INFO IgniteKernal:566 - PID: 7165 > 2019-01-05 11:47:07 INFO IgniteKernal:566 - Language runtime: Scala ver. > 2.11.12 > 2019-01-05 11:47:07 INFO IgniteKernal:566 - VM information: Java(TM) SE > Runtime Environment 1.8.0_192-ea-b04 Oracle Corporation Java HotSpot(TM) > 64-Bit Server VM 25.192-b04 > 2019-01-05 11:47:07 INFO IgniteKernal:566 - VM total memory: 0.89GB > 2019-01-05 11:47:07 INFO IgniteKernal:566 - Remote Management [restart: > off, REST: off, JMX (remote: off)] > 2019-01-05 11:47:07 INFO IgniteKernal:566 - Logger: Log4JLogger > [quiet=false, config=null] > 2019-01-05 11:47:07 INFO IgniteKernal:566 - > IGNITE_HOME=/usr/local/apache-ignite-fabric-2.6.0-bin/ > 2019-01-05 11:47:07 INFO IgniteKernal:566 - VM arguments: [-Xmx1g] > 2019-01-05 11:47:07 INFO IgniteKernal:566 - Configured caches [in > 'sysMemPlc' dataRegion: ['ignite-sys-cache'], in 'null' dataRegion: > ['sharedRDD']] > 2019-01-05 11:47:07 WARN GridDiagnostic:571 - Default local host is a > loopback address. This can be a sign of potential network configuration > problem. > 2019-01-05 11:47:07 INFO IgniteKernal:566 - 3-rd party licenses can be > found at: /usr/local/apache-ignite-fabric-2.6.0-bin//libs/licenses > 2019-01-05 11:47:07 WARN GridDiagnostic:571 - No live network interfaces > detected. If IP-multicast discovery is used - make sure to add 127.0.0.1 as > a local address. > 2019-01-05 11:47:07 WARN GridDiagnostic:571 - Initial heap size is 126MB > (should be no less than 512MB, use -Xms512m -Xmx512m). > 2019-01-05 11:47:08 INFO IgnitePluginProcessor:566 - Configured plugins: > 2019-01-05 11:47:08 INFO IgnitePluginProcessor:566 - ^-- None > 2019-01-05 11:47:08 INFO IgnitePluginProcessor:566 - > 2019-01-05 11:47:08 INFO FailureProcessor:566 - Configured failure > handler: > [hnd=StopNodeOrHaltFailureHandler [tryStop=false, timeout=0]] > 2019-01-05 11:47:08 INFO TcpCommunicationSpi:566 - Successfully bound > communication NIO server to TCP port [port=47101, locHost=0.0.0.0/0.0.0.0, > selectorsCnt=4, selectorSpins=0, pairedConn=false] > 2019-01-05 11:47:08 WARN TcpCommunicationSpi:571 - Message queue limit is > set to 0 which may lead to potential OOMEs when running cache operations in > FULL_ASYNC or PRIMARY_SYNC modes due to message queues growth on sender and > receiver sides. > 2019-01-05 11:47:08 WARN NoopCheckpointSpi:571 - Checkpoints are disabled > (to enable configure any GridCheckpointSpi implementation) > 2019-01-05 11:47:08 WARN GridCollisionManager:571 - Collision resolution > is > disabled (all jobs will be activated upon arrival). > 2019-01-05 11:47:08 INFO IgniteKernal:566 - Security status > [authentication=off, tls/ssl=off] > 2019-01-05 11:47:09 INFO ClientListenerProcessor:566 - Client connector > processor has started on TCP port 10801 > 2019-01-05 11:47:09 INFO GridRestProcessor:566 - REST protocols do not > start on client node. To start the protocols on client node set > '-DIGNITE_REST_START_ON_CLIENT=true' system property. > 2019-01-05 11:47:09 INFO IgniteKernal:566 - Non-loopback local IPs: N/A > 2019-01-05 11:47:09 INFO IgniteKernal:566 - Enabled local MACs: N/A > 2019-01-05 11:47:09 WARN IgniteKernal:571 - Ignite is starting on loopback > address... Only nodes on the same physical computer can participate in > topology. > 2019-01-05 11:47:09 WARN TcpDiscoveryMulticastIpFinder:571 - Failed to > send > multicast message (is multicast enabled on this node?). > 2019-01-05 11:47:10 INFO time:566 - Started exchange init > [topVer=AffinityTopologyVersion [topVer=4, minorTopVer=0], crd=false, > evt=NODE_JOINED, evtNode=f512a374-f6b6-4918-9790-6183a43f8064, > customEvt=null, allowMerge=true] > 2019-01-05 11:47:11 INFO GridCacheProcessor:566 - Started cache > [name=ignite-sys-cache, id=-2100569601, memoryPolicyName=sysMemPlc, > mode=REPLICATED, atomicity=TRANSACTIONAL, backups=2147483647] > 2019-01-05 11:47:11 INFO TcpCommunicationSpi:566 - Established outgoing > communication connection [locAddr=/0:0:0:0:0:0:0:1:43032, > rmtAddr=/0:0:0:0:0:0:0:1%lo:47100] > 2019-01-05 11:47:11 INFO GridCacheProcessor:566 - Started cache > [name=sharedRDD, id=-1581581875, memoryPolicyName=null, mode=PARTITIONED, > atomicity=ATOMIC, backups=1] > 2019-01-05 11:47:11 INFO time:566 - Finished exchange init > [topVer=AffinityTopologyVersion [topVer=4, minorTopVer=0], crd=false] > 2019-01-05 11:47:11 INFO CoarseGrainedSchedulerBackend$DriverEndpoint:54 - > Registered executor NettyRpcEndpointRef(spark-client://Executor) > (127.0.0.1:49300) with ID 0 > 2019-01-05 11:47:11 INFO GridDhtPartitionsExchangeFuture:566 - Received > full message, will finish exchange > [node=0a01ccfc-6d3a-4490-bdd8-90cf3b71928d, resVer=AffinityTopologyVersion > [topVer=4, minorTopVer=0]] > 2019-01-05 11:47:11 INFO GridDhtPartitionsExchangeFuture:566 - Finish > exchange future [startVer=AffinityTopologyVersion [topVer=4, > minorTopVer=0], > resVer=AffinityTopologyVersion [topVer=4, minorTopVer=0], err=null] > 2019-01-05 11:47:11 INFO CoarseGrainedSchedulerBackend$DriverEndpoint:54 - > Registered executor NettyRpcEndpointRef(spark-client://Executor) > (127.0.0.1:49296) with ID 3 > 2019-01-05 11:47:11 INFO CoarseGrainedSchedulerBackend$DriverEndpoint:54 - > Registered executor NettyRpcEndpointRef(spark-client://Executor) > (127.0.0.1:49294) with ID 2 > 2019-01-05 11:47:11 INFO IgniteKernal:566 - Performance suggestions for > grid (fix if possible) > 2019-01-05 11:47:11 INFO IgniteKernal:566 - To disable, set > -DIGNITE_PERFORMANCE_SUGGESTIONS_DISABLED=true > 2019-01-05 11:47:11 INFO IgniteKernal:566 - ^-- Enable G1 Garbage > Collector (add '-XX:+UseG1GC' to JVM options) > 2019-01-05 11:47:11 INFO IgniteKernal:566 - ^-- Set max direct memory > size if getting 'OOME: Direct buffer memory' (add > '-XX:MaxDirectMemorySize=<size>[g|G|m|M|k|K]' to JVM options) > 2019-01-05 11:47:11 INFO IgniteKernal:566 - ^-- Disable processing of > calls to System.gc() (add '-XX:+DisableExplicitGC' to JVM options) > 2019-01-05 11:47:11 INFO IgniteKernal:566 - ^-- Decrease number of > backups (set 'backups' to 0) > 2019-01-05 11:47:11 INFO IgniteKernal:566 - Refer to this page for more > performance suggestions: > https://apacheignite.readme.io/docs/jvm-and-system-tuning > 2019-01-05 11:47:11 INFO IgniteKernal:566 - > 2019-01-05 11:47:11 INFO IgniteKernal:566 - To start Console Management & > Monitoring run ignitevisorcmd.{sh|bat} > 2019-01-05 11:47:11 INFO IgniteKernal:566 - > 2019-01-05 11:47:11 INFO IgniteKernal:566 - > > >>> +---------------------------------+ > >>> Ignite ver. 2.6.0#19700101-sha1:DEV > >>> +---------------------------------+ > >>> OS name: Linux 4.15.0-43-generic amd64 > >>> CPU(s): 8 > >>> Heap: 0.89GB > >>> VM name: 7165@linux-client > >>> Local node [ID=F512A374-F6B6-4918-9790-6183A43F8064, order=4, > >>> clientMode=true] > >>> Local node addresses: [0:0:0:0:0:0:0:1%lo, 127.0.0.1] > >>> Local ports: TCP:10801 TCP:47101 UDP:47400 > > 2019-01-05 11:47:11 INFO GridDiscoveryManager:566 - Topology snapshot > [ver=4, servers=1, clients=1, CPUs=8, offheap=1.6GB, heap=1.9GB] > 2019-01-05 11:47:11 INFO GridDiscoveryManager:566 - ^-- Node > [id=F512A374-F6B6-4918-9790-6183A43F8064, clusterState=ACTIVE] > 2019-01-05 11:47:11 INFO CoarseGrainedSchedulerBackend$DriverEndpoint:54 - > Registered executor NettyRpcEndpointRef(spark-client://Executor) > (127.0.0.1:49298) with ID 1 > 2019-01-05 11:47:12 INFO BlockManagerMasterEndpoint:54 - Registering block > manager 127.0.1.1:38343 with 127.2 MB RAM, BlockManagerId(1, 127.0.1.1, > 38343, None) > 2019-01-05 11:47:12 INFO BlockManagerMasterEndpoint:54 - Registering block > manager 127.0.1.1:42897 with 127.2 MB RAM, BlockManagerId(2, 127.0.1.1, > 42897, None) > 2019-01-05 11:47:12 INFO BlockManagerMasterEndpoint:54 - Registering block > manager 127.0.1.1:32861 with 127.2 MB RAM, BlockManagerId(3, 127.0.1.1, > 32861, None) > 2019-01-05 11:47:12 INFO BlockManagerMasterEndpoint:54 - Registering block > manager 127.0.1.1:32913 with 127.2 MB RAM, BlockManagerId(0, 127.0.1.1, > 32913, None) > 2019-01-05 11:47:12 INFO SparkContext:54 - Starting job: foreachPartition > at IgniteRDD.scala:233 > 2019-01-05 11:47:12 INFO DAGScheduler:54 - Got job 0 (foreachPartition at > IgniteRDD.scala:233) with 10 output partitions > 2019-01-05 11:47:12 INFO DAGScheduler:54 - Final stage: ResultStage 0 > (foreachPartition at IgniteRDD.scala:233) > 2019-01-05 11:47:12 INFO DAGScheduler:54 - Parents of final stage: List() > 2019-01-05 11:47:12 INFO DAGScheduler:54 - Missing parents: List() > 2019-01-05 11:47:12 INFO DAGScheduler:54 - Submitting ResultStage 0 > (MapPartitionsRDD[2] at map at SparkIgniteTest.scala:28), which has no > missing parents > 2019-01-05 11:47:12 INFO MemoryStore:54 - Block broadcast_0 stored as > values in memory (estimated size 4.6 KB, free 366.3 MB) > 2019-01-05 11:47:12 INFO MemoryStore:54 - Block broadcast_0_piece0 stored > as bytes in memory (estimated size 2.6 KB, free 366.3 MB) > 2019-01-05 11:47:12 INFO BlockManagerInfo:54 - Added broadcast_0_piece0 in > memory on linux-client:41343 (size: 2.6 KB, free: 366.3 MB) > 2019-01-05 11:47:12 INFO SparkContext:54 - Created broadcast 0 from > broadcast at DAGScheduler.scala:1161 > 2019-01-05 11:47:12 INFO DAGScheduler:54 - Submitting 10 missing tasks > from > ResultStage 0 (MapPartitionsRDD[2] at map at SparkIgniteTest.scala:28) > (first 15 tasks are for partitions Vector(0, 1, 2, 3, 4, 5, 6, 7, 8, 9)) > 2019-01-05 11:47:12 INFO TaskSchedulerImpl:54 - Adding task set 0.0 with > 10 > tasks > 2019-01-05 11:47:12 INFO TaskSetManager:54 - Starting task 0.0 in stage > 0.0 > (TID 0, 127.0.1.1, executor 1, partition 0, PROCESS_LOCAL, 7870 bytes) > 2019-01-05 11:47:12 INFO TaskSetManager:54 - Starting task 1.0 in stage > 0.0 > (TID 1, 127.0.1.1, executor 0, partition 1, PROCESS_LOCAL, 7870 bytes) > 2019-01-05 11:47:12 INFO TaskSetManager:54 - Starting task 2.0 in stage > 0.0 > (TID 2, 127.0.1.1, executor 2, partition 2, PROCESS_LOCAL, 7870 bytes) > 2019-01-05 11:47:12 INFO TaskSetManager:54 - Starting task 3.0 in stage > 0.0 > (TID 3, 127.0.1.1, executor 3, partition 3, PROCESS_LOCAL, 7870 bytes) > 2019-01-05 11:47:12 INFO TaskSetManager:54 - Starting task 4.0 in stage > 0.0 > (TID 4, 127.0.1.1, executor 1, partition 4, PROCESS_LOCAL, 7870 bytes) > 2019-01-05 11:47:12 INFO TaskSetManager:54 - Starting task 5.0 in stage > 0.0 > (TID 5, 127.0.1.1, executor 0, partition 5, PROCESS_LOCAL, 7870 bytes) > 2019-01-05 11:47:12 INFO TaskSetManager:54 - Starting task 6.0 in stage > 0.0 > (TID 6, 127.0.1.1, executor 2, partition 6, PROCESS_LOCAL, 7870 bytes) > 2019-01-05 11:47:12 INFO TaskSetManager:54 - Starting task 7.0 in stage > 0.0 > (TID 7, 127.0.1.1, executor 3, partition 7, PROCESS_LOCAL, 7870 bytes) > 2019-01-05 11:47:13 INFO BlockManagerInfo:54 - Added broadcast_0_piece0 in > memory on 127.0.1.1:38343 (size: 2.6 KB, free: 127.2 MB) > 2019-01-05 11:47:13 INFO BlockManagerInfo:54 - Added broadcast_0_piece0 in > memory on 127.0.1.1:32913 (size: 2.6 KB, free: 127.2 MB) > 2019-01-05 11:47:13 INFO BlockManagerInfo:54 - Added broadcast_0_piece0 in > memory on 127.0.1.1:32861 (size: 2.6 KB, free: 127.2 MB) > 2019-01-05 11:47:13 INFO BlockManagerInfo:54 - Added broadcast_0_piece0 in > memory on 127.0.1.1:42897 (size: 2.6 KB, free: 127.2 MB) > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Starting task 8.0 in stage > 0.0 > (TID 8, 127.0.1.1, executor 1, partition 8, PROCESS_LOCAL, 7870 bytes) > 2019-01-05 11:47:14 WARN TaskSetManager:66 - Lost task 0.0 in stage 0.0 > (TID 0, 127.0.1.1, executor 1): java.lang.NoClassDefFoundError: Could not > initialize class org.apache.ignite.internal.util.IgniteUtils > at > > org.apache.ignite.spark.IgniteContext$.setIgniteHome(IgniteContext.scala:195) > at > org.apache.ignite.spark.IgniteContext.ignite(IgniteContext.scala:142) > at > > org.apache.ignite.spark.IgniteRDD$$anonfun$savePairs$1.apply(IgniteRDD.scala:234) > at > > org.apache.ignite.spark.IgniteRDD$$anonfun$savePairs$1.apply(IgniteRDD.scala:233) > at > > org.apache.spark.rdd.RDD$$anonfun$foreachPartition$1$$anonfun$apply$28.apply(RDD.scala:935) > at > > org.apache.spark.rdd.RDD$$anonfun$foreachPartition$1$$anonfun$apply$28.apply(RDD.scala:935) > at > > org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2101) > at > > org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2101) > at > org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90) > at org.apache.spark.scheduler.Task.run(Task.scala:121) > at > > org.apache.spark.executor.Executor$TaskRunner$$anonfun$10.apply(Executor.scala:402) > at > org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1360) > at > org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:408) > at > > java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1135) > at > > java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) > at java.base/java.lang.Thread.run(Thread.java:844) > > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Starting task 0.1 in stage > 0.0 > (TID 9, 127.0.1.1, executor 1, partition 0, PROCESS_LOCAL, 7870 bytes) > 2019-01-05 11:47:14 WARN TaskSetManager:66 - Lost task 4.0 in stage 0.0 > (TID 4, 127.0.1.1, executor 1): java.lang.ExceptionInInitializerError > at > org.apache.ignite.internal.util.IgniteUtils.<clinit>(IgniteUtils.java:769) > at > > org.apache.ignite.spark.IgniteContext$.setIgniteHome(IgniteContext.scala:195) > at > org.apache.ignite.spark.IgniteContext.ignite(IgniteContext.scala:142) > at > > org.apache.ignite.spark.IgniteRDD$$anonfun$savePairs$1.apply(IgniteRDD.scala:234) > at > > org.apache.ignite.spark.IgniteRDD$$anonfun$savePairs$1.apply(IgniteRDD.scala:233) > at > > org.apache.spark.rdd.RDD$$anonfun$foreachPartition$1$$anonfun$apply$28.apply(RDD.scala:935) > at > > org.apache.spark.rdd.RDD$$anonfun$foreachPartition$1$$anonfun$apply$28.apply(RDD.scala:935) > at > > org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2101) > at > > org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2101) > at > org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90) > at org.apache.spark.scheduler.Task.run(Task.scala:121) > at > > org.apache.spark.executor.Executor$TaskRunner$$anonfun$10.apply(Executor.scala:402) > at > org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1360) > at > org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:408) > at > > java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1135) > at > > java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) > at java.base/java.lang.Thread.run(Thread.java:844) > Caused by: java.lang.RuntimeException: jdk.internal.misc.JavaNioAccess > class > is unavailable. > at > > org.apache.ignite.internal.util.GridUnsafe.javaNioAccessObject(GridUnsafe.java:1453) > at > org.apache.ignite.internal.util.GridUnsafe.<clinit>(GridUnsafe.java:112) > ... 17 more > Caused by: java.lang.IllegalAccessException: class > org.apache.ignite.internal.util.GridUnsafe cannot access class > jdk.internal.misc.SharedSecrets (in module java.base) because module > java.base does not export jdk.internal.misc to unnamed module @2c5781b6 > at > > java.base/jdk.internal.reflect.Reflection.newIllegalAccessException(Reflection.java:360) > at > > java.base/java.lang.reflect.AccessibleObject.checkAccess(AccessibleObject.java:589) > at java.base/java.lang.reflect.Method.invoke(Method.java:556) > at > > org.apache.ignite.internal.util.GridUnsafe.javaNioAccessObject(GridUnsafe.java:1450) > ... 18 more > > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Starting task 4.1 in stage > 0.0 > (TID 10, 127.0.1.1, executor 3, partition 4, PROCESS_LOCAL, 7870 bytes) > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Lost task 3.0 in stage 0.0 > (TID 3) on 127.0.1.1, executor 3: java.lang.NoClassDefFoundError (Could not > initialize class org.apache.ignite.internal.util.IgniteUtils) [duplicate 1] > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Starting task 3.1 in stage > 0.0 > (TID 11, 127.0.1.1, executor 3, partition 3, PROCESS_LOCAL, 7870 bytes) > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Lost task 7.0 in stage 0.0 > (TID 7) on 127.0.1.1, executor 3: java.lang.ExceptionInInitializerError > (null) [duplicate 1] > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Starting task 7.1 in stage > 0.0 > (TID 12, 127.0.1.1, executor 2, partition 7, PROCESS_LOCAL, 7870 bytes) > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Lost task 2.0 in stage 0.0 > (TID 2) on 127.0.1.1, executor 2: java.lang.NoClassDefFoundError (Could not > initialize class org.apache.ignite.internal.util.IgniteUtils) [duplicate 2] > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Starting task 2.1 in stage > 0.0 > (TID 13, 127.0.1.1, executor 1, partition 2, PROCESS_LOCAL, 7870 bytes) > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Starting task 9.0 in stage > 0.0 > (TID 14, 127.0.1.1, executor 1, partition 9, PROCESS_LOCAL, 7927 bytes) > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Lost task 0.1 in stage 0.0 > (TID 9) on 127.0.1.1, executor 1: java.lang.NoClassDefFoundError (Could not > initialize class org.apache.ignite.internal.util.IgniteUtils) [duplicate 3] > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Lost task 8.0 in stage 0.0 > (TID 8) on 127.0.1.1, executor 1: java.lang.NoClassDefFoundError (Could not > initialize class org.apache.ignite.internal.util.IgniteUtils) [duplicate 4] > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Starting task 8.1 in stage > 0.0 > (TID 15, 127.0.1.1, executor 0, partition 8, PROCESS_LOCAL, 7870 bytes) > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Lost task 1.0 in stage 0.0 > (TID 1) on 127.0.1.1, executor 0: java.lang.NoClassDefFoundError (Could not > initialize class org.apache.ignite.internal.util.IgniteUtils) [duplicate 5] > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Starting task 1.1 in stage > 0.0 > (TID 16, 127.0.1.1, executor 3, partition 1, PROCESS_LOCAL, 7870 bytes) > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Starting task 0.2 in stage > 0.0 > (TID 17, 127.0.1.1, executor 2, partition 0, PROCESS_LOCAL, 7870 bytes) > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Lost task 4.1 in stage 0.0 > (TID 10) on 127.0.1.1, executor 3: java.lang.NoClassDefFoundError (Could > not > initialize class org.apache.ignite.internal.util.IgniteUtils) [duplicate 6] > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Lost task 5.0 in stage 0.0 > (TID 5) on 127.0.1.1, executor 0: java.lang.ExceptionInInitializerError > (null) [duplicate 2] > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Starting task 5.1 in stage > 0.0 > (TID 18, 127.0.1.1, executor 0, partition 5, PROCESS_LOCAL, 7870 bytes) > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Lost task 6.0 in stage 0.0 > (TID 6) on 127.0.1.1, executor 2: java.lang.ExceptionInInitializerError > (null) [duplicate 3] > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Starting task 6.1 in stage > 0.0 > (TID 19, 127.0.1.1, executor 2, partition 6, PROCESS_LOCAL, 7870 bytes) > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Lost task 7.1 in stage 0.0 > (TID 12) on 127.0.1.1, executor 2: java.lang.NoClassDefFoundError (Could > not > initialize class org.apache.ignite.internal.util.IgniteUtils) [duplicate 7] > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Lost task 9.0 in stage 0.0 > (TID 14) on 127.0.1.1, executor 1: java.lang.NoClassDefFoundError (Could > not > initialize class org.apache.ignite.internal.util.IgniteUtils) [duplicate 8] > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Starting task 9.1 in stage > 0.0 > (TID 20, 127.0.1.1, executor 1, partition 9, PROCESS_LOCAL, 7927 bytes) > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Starting task 7.2 in stage > 0.0 > (TID 21, 127.0.1.1, executor 1, partition 7, PROCESS_LOCAL, 7870 bytes) > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Lost task 2.1 in stage 0.0 > (TID 13) on 127.0.1.1, executor 1: java.lang.NoClassDefFoundError (Could > not > initialize class org.apache.ignite.internal.util.IgniteUtils) [duplicate 9] > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Starting task 2.2 in stage > 0.0 > (TID 22, 127.0.1.1, executor 2, partition 2, PROCESS_LOCAL, 7870 bytes) > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Lost task 0.2 in stage 0.0 > (TID 17) on 127.0.1.1, executor 2: java.lang.NoClassDefFoundError (Could > not > initialize class org.apache.ignite.internal.util.IgniteUtils) [duplicate > 10] > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Starting task 0.3 in stage > 0.0 > (TID 23, 127.0.1.1, executor 3, partition 0, PROCESS_LOCAL, 7870 bytes) > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Lost task 3.1 in stage 0.0 > (TID 11) on 127.0.1.1, executor 3: java.lang.NoClassDefFoundError (Could > not > initialize class org.apache.ignite.internal.util.IgniteUtils) [duplicate > 11] > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Lost task 5.1 in stage 0.0 > (TID 18) on 127.0.1.1, executor 0: java.lang.NoClassDefFoundError (Could > not > initialize class org.apache.ignite.internal.util.IgniteUtils) [duplicate > 12] > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Starting task 5.2 in stage > 0.0 > (TID 24, 127.0.1.1, executor 0, partition 5, PROCESS_LOCAL, 7870 bytes) > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Starting task 3.2 in stage > 0.0 > (TID 25, 127.0.1.1, executor 1, partition 3, PROCESS_LOCAL, 7870 bytes) > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Lost task 9.1 in stage 0.0 > (TID 20) on 127.0.1.1, executor 1: java.lang.NoClassDefFoundError (Could > not > initialize class org.apache.ignite.internal.util.IgniteUtils) [duplicate > 13] > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Starting task 9.2 in stage > 0.0 > (TID 26, 127.0.1.1, executor 3, partition 9, PROCESS_LOCAL, 7927 bytes) > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Starting task 4.2 in stage > 0.0 > (TID 27, 127.0.1.1, executor 3, partition 4, PROCESS_LOCAL, 7870 bytes) > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Lost task 0.3 in stage 0.0 > (TID 23) on 127.0.1.1, executor 3: java.lang.NoClassDefFoundError (Could > not > initialize class org.apache.ignite.internal.util.IgniteUtils) [duplicate > 14] > 2019-01-05 11:47:14 ERROR TaskSetManager:70 - Task 0 in stage 0.0 failed 4 > times; aborting job > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Lost task 1.1 in stage 0.0 > (TID 16) on 127.0.1.1, executor 3: java.lang.NoClassDefFoundError (Could > not > initialize class org.apache.ignite.internal.util.IgniteUtils) [duplicate > 15] > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Lost task 7.2 in stage 0.0 > (TID 21) on 127.0.1.1, executor 1: java.lang.NoClassDefFoundError (Could > not > initialize class org.apache.ignite.internal.util.IgniteUtils) [duplicate > 16] > 2019-01-05 11:47:14 INFO TaskSchedulerImpl:54 - Cancelling stage 0 > 2019-01-05 11:47:14 INFO TaskSchedulerImpl:54 - Killing all running tasks > in stage 0: Stage cancelled > 2019-01-05 11:47:14 INFO TaskSchedulerImpl:54 - Stage 0 was cancelled > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Lost task 8.1 in stage 0.0 > (TID 15) on 127.0.1.1, executor 0: java.lang.NoClassDefFoundError (Could > not > initialize class org.apache.ignite.internal.util.IgniteUtils) [duplicate > 17] > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Lost task 2.2 in stage 0.0 > (TID 22) on 127.0.1.1, executor 2: java.lang.NoClassDefFoundError (Could > not > initialize class org.apache.ignite.internal.util.IgniteUtils) [duplicate > 18] > 2019-01-05 11:47:14 INFO DAGScheduler:54 - ResultStage 0 (foreachPartition > at IgniteRDD.scala:233) failed in 2.358 s due to Job aborted due to stage > failure: Task 0 in stage 0.0 failed 4 times, most recent failure: Lost task > 0.3 in stage 0.0 (TID 23, 127.0.1.1, executor 3): > java.lang.NoClassDefFoundError: Could not initialize class > org.apache.ignite.internal.util.IgniteUtils > at > > org.apache.ignite.spark.IgniteContext$.setIgniteHome(IgniteContext.scala:195) > at > org.apache.ignite.spark.IgniteContext.ignite(IgniteContext.scala:142) > at > > org.apache.ignite.spark.IgniteRDD$$anonfun$savePairs$1.apply(IgniteRDD.scala:234) > at > > org.apache.ignite.spark.IgniteRDD$$anonfun$savePairs$1.apply(IgniteRDD.scala:233) > at > > org.apache.spark.rdd.RDD$$anonfun$foreachPartition$1$$anonfun$apply$28.apply(RDD.scala:935) > at > > org.apache.spark.rdd.RDD$$anonfun$foreachPartition$1$$anonfun$apply$28.apply(RDD.scala:935) > at > > org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2101) > at > > org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2101) > at > org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90) > at org.apache.spark.scheduler.Task.run(Task.scala:121) > at > > org.apache.spark.executor.Executor$TaskRunner$$anonfun$10.apply(Executor.scala:402) > at > org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1360) > at > org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:408) > at > > java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1135) > at > > java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) > at java.base/java.lang.Thread.run(Thread.java:844) > > Driver stacktrace: > 2019-01-05 11:47:14 INFO TaskSetManager:54 - Lost task 3.2 in stage 0.0 > (TID 25) on 127.0.1.1, executor 1: java.lang.NoClassDefFoundError (Could > not > initialize class org.apache.ignite.internal.util.IgniteUtils) [duplicate > 19] > 2019-01-05 11:47:15 INFO DAGScheduler:54 - Job 0 failed: foreachPartition > at IgniteRDD.scala:233, took 2.443962 s > Exception in thread "main" org.apache.spark.SparkException: Job aborted due > to stage failure: Task 0 in stage 0.0 failed 4 times, most recent failure: > Lost task 0.3 in stage 0.0 (TID 23, 127.0.1.1, executor 3): > java.lang.NoClassDefFoundError: Could not initialize class > org.apache.ignite.internal.util.IgniteUtils > at > > org.apache.ignite.spark.IgniteContext$.setIgniteHome(IgniteContext.scala:195) > at > org.apache.ignite.spark.IgniteContext.ignite(IgniteContext.scala:142) > at > > org.apache.ignite.spark.IgniteRDD$$anonfun$savePairs$1.apply(IgniteRDD.scala:234) > at > > org.apache.ignite.spark.IgniteRDD$$anonfun$savePairs$1.apply(IgniteRDD.scala:233) > at > > org.apache.spark.rdd.RDD$$anonfun$foreachPartition$1$$anonfun$apply$28.apply(RDD.scala:935) > at > > org.apache.spark.rdd.RDD$$anonfun$foreachPartition$1$$anonfun$apply$28.apply(RDD.scala:935) > at > > org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2101) > at > > org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2101) > at > org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90) > at org.apache.spark.scheduler.Task.run(Task.scala:121) > at > > org.apache.spark.executor.Executor$TaskRunner$$anonfun$10.apply(Executor.scala:402) > at > org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1360) > at > org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:408) > at > > java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1135) > at > > java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) > at java.base/java.lang.Thread.run(Thread.java:844) > > Driver stacktrace: > at > org.apache.spark.scheduler.DAGScheduler.org > $apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1887) > at > > org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1875) > at > > org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1874) > at > > scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) > at > scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) > at > org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1874) > at > > org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:926) > at > > org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:926) > at scala.Option.foreach(Option.scala:257) > at > > org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:926) > at > > org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:2108) > at > > org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2057) > at > > org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2046) > at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:49) > at > org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:737) > at org.apache.spark.SparkContext.runJob(SparkContext.scala:2061) > at org.apache.spark.SparkContext.runJob(SparkContext.scala:2082) > at org.apache.spark.SparkContext.runJob(SparkContext.scala:2101) > at org.apache.spark.SparkContext.runJob(SparkContext.scala:2126) > at > org.apache.spark.rdd.RDD$$anonfun$foreachPartition$1.apply(RDD.scala:935) > at > org.apache.spark.rdd.RDD$$anonfun$foreachPartition$1.apply(RDD.scala:933) > at > > org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) > at > > org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) > at org.apache.spark.rdd.RDD.withScope(RDD.scala:363) > at org.apache.spark.rdd.RDD.foreachPartition(RDD.scala:933) > at org.apache.ignite.spark.IgniteRDD.savePairs(IgniteRDD.scala:233) > at > > com.gridgain.RDDWriter$.delayedEndpoint$com$gridgain$RDDWriter$1(SparkIgniteTest.scala:28) > at > com.gridgain.RDDWriter$delayedInit$body.apply(SparkIgniteTest.scala:23) > at scala.Function0$class.apply$mcV$sp(Function0.scala:34) > at > scala.runtime.AbstractFunction0.apply$mcV$sp(AbstractFunction0.scala:12) > at scala.App$$anonfun$main$1.apply(App.scala:76) > at scala.App$$anonfun$main$1.apply(App.scala:76) > at scala.collection.immutable.List.foreach(List.scala:392) > at > > scala.collection.generic.TraversableForwarder$class.foreach(TraversableForwarder.scala:35) > at scala.App$class.main(App.scala:76) > at com.gridgain.RDDWriter$.main(SparkIgniteTest.scala:23) > at com.gridgain.RDDWriter.main(SparkIgniteTest.scala) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at > > org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52) > at > org.apache.spark.deploy.SparkSubmit.org > $apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:849) > at > org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:167) > at > org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:195) > at > org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:86) > at > org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:924) > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:933) > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > Caused by: java.lang.NoClassDefFoundError: Could not initialize class > org.apache.ignite.internal.util.IgniteUtils > at > > org.apache.ignite.spark.IgniteContext$.setIgniteHome(IgniteContext.scala:195) > at > org.apache.ignite.spark.IgniteContext.ignite(IgniteContext.scala:142) > at > > org.apache.ignite.spark.IgniteRDD$$anonfun$savePairs$1.apply(IgniteRDD.scala:234) > at > > org.apache.ignite.spark.IgniteRDD$$anonfun$savePairs$1.apply(IgniteRDD.scala:233) > at > > org.apache.spark.rdd.RDD$$anonfun$foreachPartition$1$$anonfun$apply$28.apply(RDD.scala:935) > at > > org.apache.spark.rdd.RDD$$anonfun$foreachPartition$1$$anonfun$apply$28.apply(RDD.scala:935) > at > > org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2101) > at > > org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2101) > at > org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90) > at org.apache.spark.scheduler.Task.run(Task.scala:121) > at > > org.apache.spark.executor.Executor$TaskRunner$$anonfun$10.apply(Executor.scala:402) > at > org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1360) > at > org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:408) > at > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1135) > at > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) > at java.lang.Thread.run(Thread.java:844) > 2019-01-05 11:47:15 INFO TaskSetManager:54 - Lost task 6.1 in stage 0.0 > (TID 19) on 127.0.1.1, executor 2: java.lang.NoClassDefFoundError (Could > not > initialize class org.apache.ignite.internal.util.IgniteUtils) [duplicate > 20] > 2019-01-05 11:47:15 INFO TaskSetManager:54 - Lost task 4.2 in stage 0.0 > (TID 27) on 127.0.1.1, executor 3: java.lang.NoClassDefFoundError (Could > not > initialize class org.apache.ignite.internal.util.IgniteUtils) [duplicate > 21] > 2019-01-05 11:47:15 INFO TaskSchedulerImpl:54 - Removed TaskSet 0.0, whose > tasks have all completed, from pool > 2019-01-05 11:47:15 INFO TaskSetManager:54 - Lost task 9.2 in stage 0.0 > (TID 26) on 127.0.1.1, executor 3: java.lang.NoClassDefFoundError (Could > not > initialize class org.apache.ignite.internal.util.IgniteUtils) [duplicate > 22] > 2019-01-05 11:47:15 INFO TaskSchedulerImpl:54 - Removed TaskSet 0.0, whose > tasks have all completed, from pool > 2019-01-05 11:47:15 INFO TaskSetManager:54 - Lost task 5.2 in stage 0.0 > (TID 24) on 127.0.1.1, executor 0: java.lang.NoClassDefFoundError (Could > not > initialize class org.apache.ignite.internal.util.IgniteUtils) [duplicate > 23] > 2019-01-05 11:47:15 INFO TaskSchedulerImpl:54 - Removed TaskSet 0.0, whose > tasks have all completed, from pool > 2019-01-05 11:48:11 INFO IgniteKernal:566 - > Metrics for local node (to disable set 'metricsLogFrequency' to 0) > ^-- Node [id=f512a374, uptime=00:01:00.027] > ^-- H/N/C [hosts=1, nodes=2, CPUs=8] > ^-- CPU [cur=0.43%, avg=0.85%, GC=0%] > ^-- PageMemory [pages=0] > ^-- Heap [used=392MB, free=56.85%, comm=602MB] > ^-- Non heap [used=83MB, free=-1%, comm=85MB] > ^-- Outbound messages queue [size=0] > ^-- Public thread pool [active=0, idle=0, qSize=0] > ^-- System thread pool [active=0, idle=0, qSize=0] > 2019-01-05 11:48:20 INFO GridUpdateNotifier:566 - Update status is not > available. > > < > http://apache-ignite-users.70518.x6.nabble.com/file/t2160/Screenshot_from_2019-01-05_12-21-06.png> > > > > > > > > > -- > Sent from: http://apache-ignite-users.70518.x6.nabble.com/ >
