Re: Fail to launch spark-shell on windows 2008 R2
Hi Ningjun, I have been working with Spark 1.2 on Windows 7 and Windows 2008 R2 (purely for development purposes). I had most recently installed them utilizing Java 1.8, Scala 2.10.4, and Spark 1.2 Precompiled for Hadoop 2.4+. A handy thread concerning the null\bin\winutils issue is addressed in an earlier thread at: http://apache-spark-user-list.1001560.n3.nabble.com/Run-spark-unit-test-on-Windows-7-td8656.html Hope this helps a little bit! Denny On Tue Feb 03 2015 at 8:24:24 AM Wang, Ningjun (LNG-NPV) < ningjun.w...@lexisnexis.com> wrote: > Hi Gen > > > > Thanks for your feedback. We do have a business reason to run spark on > windows. We have an existing application that is built on C# .NET running > on windows. We are considering adding spark to the application for parallel > processing of large data. We want spark to run on windows so it integrate > with our existing app easily. > > > > Has anybody use spark on windows for production system? Is spark reliable > on windows? > > > > Ningjun > > > > *From:* gen tang [mailto:gen.tan...@gmail.com] > *Sent:* Thursday, January 29, 2015 12:53 PM > > > *To:* Wang, Ningjun (LNG-NPV) > *Cc:* user@spark.apache.org > *Subject:* Re: Fail to launch spark-shell on windows 2008 R2 > > > > Hi, > > > > Using spark under windows is a really bad idea, because even you solve the > problems about hadoop, you probably will meet the problem of > java.net.SocketException. connection reset by peer. It is caused by the > fact we ask socket port too frequently under windows. In my knowledge, it > is really difficult to solve. And you will find something really funny: the > same code sometimes works and sometimes not, even in the shell mode. > > > > And I am sorry but I don't see the interest to run spark under windows and > moreover using local file system in a business environment. Do you have a > cluster in windows? > > > > FYI, I have used spark prebuilt on hadoop 1 under windows 7 and there is > no problem to launch, but have problem of java.net.SocketException. If you > are using spark prebuilt on hadoop 2, you should consider follow the > solution provided by https://issues.apache.org/jira/browse/SPARK-2356 > > > > Cheers > > Gen > > > > > > > > On Thu, Jan 29, 2015 at 5:54 PM, Wang, Ningjun (LNG-NPV) < > ningjun.w...@lexisnexis.com> wrote: > > Install virtual box which run Linux? That does not help us. We have > business reason to run it on Windows operating system, e.g. Windows 2008 R2. > > > > If anybody have done that, please give some advise on what version of > spark, which version of Hadoop do you built spark against, etc…. Note that > we only use local file system and do not have any hdfs file system at all. > I don’t understand why spark generate so many error on Hadoop while we > don’t even need hdfs. > > > > Ningjun > > > > > > *From:* gen tang [mailto:gen.tan...@gmail.com] > *Sent:* Thursday, January 29, 2015 10:45 AM > *To:* Wang, Ningjun (LNG-NPV) > *Cc:* user@spark.apache.org > *Subject:* Re: Fail to launch spark-shell on windows 2008 R2 > > > > Hi, > > > > I tried to use spark under windows once. However the only solution that I > found is to install virtualbox > > > > Hope this can help you. > > Best > > Gen > > > > > > On Thu, Jan 29, 2015 at 4:18 PM, Wang, Ningjun (LNG-NPV) < > ningjun.w...@lexisnexis.com> wrote: > > I deployed spark-1.1.0 on Windows 7 and was albe to launch the > spark-shell. I then deploy it to windows 2008 R2 and launch the > spark-shell, I got the error > > > > java.lang.RuntimeException: Error while running command to get file > permissions : java.io.IOExceptio > > n: Cannot run program "ls": CreateProcess error=2, The system cannot find > the file specified > > at java.lang.ProcessBuilder.start(Unknown Source) > > at org.apache.hadoop.util.Shell.runCommand(Shell.java:200) > > at org.apache.hadoop.util.Shell.run(Shell.java:182) > > at > org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:375) > > at org.apache.hadoop.util.Shell.execCommand(Shell.java:461) > > at org.apache.hadoop.util.Shell.execCommand(Shell.java:444) > > at org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:710) > > at > org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.loadPermissionInfo(RawLocalFil > > eSystem.java:443) > > at > org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.getPermission(RawLocalFileSyst > > em.java:418) > > > > &g
RE: Fail to launch spark-shell on windows 2008 R2
Hi Gen Thanks for your feedback. We do have a business reason to run spark on windows. We have an existing application that is built on C# .NET running on windows. We are considering adding spark to the application for parallel processing of large data. We want spark to run on windows so it integrate with our existing app easily. Has anybody use spark on windows for production system? Is spark reliable on windows? Ningjun From: gen tang [mailto:gen.tan...@gmail.com] Sent: Thursday, January 29, 2015 12:53 PM To: Wang, Ningjun (LNG-NPV) Cc: user@spark.apache.org Subject: Re: Fail to launch spark-shell on windows 2008 R2 Hi, Using spark under windows is a really bad idea, because even you solve the problems about hadoop, you probably will meet the problem of java.net.SocketException. connection reset by peer. It is caused by the fact we ask socket port too frequently under windows. In my knowledge, it is really difficult to solve. And you will find something really funny: the same code sometimes works and sometimes not, even in the shell mode. And I am sorry but I don't see the interest to run spark under windows and moreover using local file system in a business environment. Do you have a cluster in windows? FYI, I have used spark prebuilt on hadoop 1 under windows 7 and there is no problem to launch, but have problem of java.net.SocketException. If you are using spark prebuilt on hadoop 2, you should consider follow the solution provided by https://issues.apache.org/jira/browse/SPARK-2356 Cheers Gen On Thu, Jan 29, 2015 at 5:54 PM, Wang, Ningjun (LNG-NPV) mailto:ningjun.w...@lexisnexis.com>> wrote: Install virtual box which run Linux? That does not help us. We have business reason to run it on Windows operating system, e.g. Windows 2008 R2. If anybody have done that, please give some advise on what version of spark, which version of Hadoop do you built spark against, etc…. Note that we only use local file system and do not have any hdfs file system at all. I don’t understand why spark generate so many error on Hadoop while we don’t even need hdfs. Ningjun From: gen tang [mailto:gen.tan...@gmail.com<mailto:gen.tan...@gmail.com>] Sent: Thursday, January 29, 2015 10:45 AM To: Wang, Ningjun (LNG-NPV) Cc: user@spark.apache.org<mailto:user@spark.apache.org> Subject: Re: Fail to launch spark-shell on windows 2008 R2 Hi, I tried to use spark under windows once. However the only solution that I found is to install virtualbox Hope this can help you. Best Gen On Thu, Jan 29, 2015 at 4:18 PM, Wang, Ningjun (LNG-NPV) mailto:ningjun.w...@lexisnexis.com>> wrote: I deployed spark-1.1.0 on Windows 7 and was albe to launch the spark-shell. I then deploy it to windows 2008 R2 and launch the spark-shell, I got the error java.lang.RuntimeException: Error while running command to get file permissions : java.io.IOExceptio n: Cannot run program "ls": CreateProcess error=2, The system cannot find the file specified at java.lang.ProcessBuilder.start(Unknown Source) at org.apache.hadoop.util.Shell.runCommand(Shell.java:200) at org.apache.hadoop.util.Shell.run(Shell.java:182) at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:375) at org.apache.hadoop.util.Shell.execCommand(Shell.java:461) at org.apache.hadoop.util.Shell.execCommand(Shell.java:444) at org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:710) at org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.loadPermissionInfo(RawLocalFil eSystem.java:443) at org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.getPermission(RawLocalFileSyst em.java:418) Here is the detail output C:\spark-1.1.0\bin> spark-shell 15/01/29 10:13:13 INFO SecurityManager: Changing view acls to: ningjun.wang, 15/01/29 10:13:13 INFO SecurityManager: Changing modify acls to: ningjun.wang, 15/01/29 10:13:13 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(ningjun.wang, ); users with modify permissions: Set(ningjun.wang, ) 15/01/29 10:13:13 INFO HttpServer: Starting HTTP Server 15/01/29 10:13:14 INFO Server: jetty-8.y.z-SNAPSHOT 15/01/29 10:13:14 INFO AbstractConnector: Started SocketConnector@0.0.0.0:53692<http://SocketConnector@0.0.0.0:53692> 15/01/29 10:13:14 INFO Utils: Successfully started service 'HTTP class server' on port 53692. Failed to created SparkJLineReader: java.lang.NoClassDefFoundError: Could not initialize class org.f usesource.jansi.internal.Kernel32 Falling back to SimpleReader. Welcome to __ / __/__ ___ _/ /__ _\ \/ _ \/ _ `/ __/ '_/ /___/ .__/\_,_/_/ /_/\_\ version 1.1.0 /_/ Using Scala version 2.10.4 (Java HotSpot(TM) 64-Bit Server VM, Java 1.7.0_71) Type in expressions to have them evaluated. Type :help for more information
Re: Fail to launch spark-shell on windows 2008 R2
Hi, Using spark under windows is a really bad idea, because even you solve the problems about hadoop, you probably will meet the problem of java.net.SocketException. connection reset by peer. It is caused by the fact we ask socket port too frequently under windows. In my knowledge, it is really difficult to solve. And you will find something really funny: the same code sometimes works and sometimes not, even in the shell mode. And I am sorry but I don't see the interest to run spark under windows and moreover using local file system in a business environment. Do you have a cluster in windows? FYI, I have used spark prebuilt on hadoop 1 under windows 7 and there is no problem to launch, but have problem of java.net.SocketException. If you are using spark prebuilt on hadoop 2, you should consider follow the solution provided by https://issues.apache.org/jira/browse/SPARK-2356 Cheers Gen On Thu, Jan 29, 2015 at 5:54 PM, Wang, Ningjun (LNG-NPV) < ningjun.w...@lexisnexis.com> wrote: > Install virtual box which run Linux? That does not help us. We have > business reason to run it on Windows operating system, e.g. Windows 2008 R2. > > > > If anybody have done that, please give some advise on what version of > spark, which version of Hadoop do you built spark against, etc…. Note that > we only use local file system and do not have any hdfs file system at all. > I don’t understand why spark generate so many error on Hadoop while we > don’t even need hdfs. > > > > Ningjun > > > > > > *From:* gen tang [mailto:gen.tan...@gmail.com] > *Sent:* Thursday, January 29, 2015 10:45 AM > *To:* Wang, Ningjun (LNG-NPV) > *Cc:* user@spark.apache.org > *Subject:* Re: Fail to launch spark-shell on windows 2008 R2 > > > > Hi, > > > > I tried to use spark under windows once. However the only solution that I > found is to install virtualbox > > > > Hope this can help you. > > Best > > Gen > > > > > > On Thu, Jan 29, 2015 at 4:18 PM, Wang, Ningjun (LNG-NPV) < > ningjun.w...@lexisnexis.com> wrote: > > I deployed spark-1.1.0 on Windows 7 and was albe to launch the > spark-shell. I then deploy it to windows 2008 R2 and launch the > spark-shell, I got the error > > > > java.lang.RuntimeException: Error while running command to get file > permissions : java.io.IOExceptio > > n: Cannot run program "ls": CreateProcess error=2, The system cannot find > the file specified > > at java.lang.ProcessBuilder.start(Unknown Source) > > at org.apache.hadoop.util.Shell.runCommand(Shell.java:200) > > at org.apache.hadoop.util.Shell.run(Shell.java:182) > > at > org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:375) > > at org.apache.hadoop.util.Shell.execCommand(Shell.java:461) > > at org.apache.hadoop.util.Shell.execCommand(Shell.java:444) > > at org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:710) > > at > org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.loadPermissionInfo(RawLocalFil > > eSystem.java:443) > > at > org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.getPermission(RawLocalFileSyst > > em.java:418) > > > > > > > > Here is the detail output > > > > > > C:\spark-1.1.0\bin> spark-shell > > 15/01/29 10:13:13 INFO SecurityManager: Changing view acls to: > ningjun.wang, > > 15/01/29 10:13:13 INFO SecurityManager: Changing modify acls to: > ningjun.wang, > > 15/01/29 10:13:13 INFO SecurityManager: SecurityManager: authentication > disabled; ui acls disabled; > > users with view permissions: Set(ningjun.wang, ); users with modify > permissions: Set(ningjun.wang, ) > > > > 15/01/29 10:13:13 INFO HttpServer: Starting HTTP Server > > 15/01/29 10:13:14 INFO Server: jetty-8.y.z-SNAPSHOT > > 15/01/29 10:13:14 INFO AbstractConnector: Started > SocketConnector@0.0.0.0:53692 > > 15/01/29 10:13:14 INFO Utils: Successfully started service 'HTTP class > server' on port 53692. > > Failed to created SparkJLineReader: java.lang.NoClassDefFoundError: Could > not initialize class org.f > > usesource.jansi.internal.Kernel32 > > Falling back to SimpleReader. > > Welcome to > > __ > > / __/__ ___ _/ /__ > > _\ \/ _ \/ _ `/ __/ '_/ > >/___/ .__/\_,_/_/ /_/\_\ version 1.1.0 > > /_/ > > > > Using Scala version 2.10.4 (Java HotSpot(TM) 64-Bit Server VM, Java > 1.7.0_71) > > Type in expressions to have them evaluated. > > Type :help for more information. > > 15/01/29 10:13:18 INFO SecurityManage
RE: Fail to launch spark-shell on windows 2008 R2
Install virtual box which run Linux? That does not help us. We have business reason to run it on Windows operating system, e.g. Windows 2008 R2. If anybody have done that, please give some advise on what version of spark, which version of Hadoop do you built spark against, etc…. Note that we only use local file system and do not have any hdfs file system at all. I don’t understand why spark generate so many error on Hadoop while we don’t even need hdfs. Ningjun From: gen tang [mailto:gen.tan...@gmail.com] Sent: Thursday, January 29, 2015 10:45 AM To: Wang, Ningjun (LNG-NPV) Cc: user@spark.apache.org Subject: Re: Fail to launch spark-shell on windows 2008 R2 Hi, I tried to use spark under windows once. However the only solution that I found is to install virtualbox Hope this can help you. Best Gen On Thu, Jan 29, 2015 at 4:18 PM, Wang, Ningjun (LNG-NPV) mailto:ningjun.w...@lexisnexis.com>> wrote: I deployed spark-1.1.0 on Windows 7 and was albe to launch the spark-shell. I then deploy it to windows 2008 R2 and launch the spark-shell, I got the error java.lang.RuntimeException: Error while running command to get file permissions : java.io.IOExceptio n: Cannot run program "ls": CreateProcess error=2, The system cannot find the file specified at java.lang.ProcessBuilder.start(Unknown Source) at org.apache.hadoop.util.Shell.runCommand(Shell.java:200) at org.apache.hadoop.util.Shell.run(Shell.java:182) at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:375) at org.apache.hadoop.util.Shell.execCommand(Shell.java:461) at org.apache.hadoop.util.Shell.execCommand(Shell.java:444) at org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:710) at org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.loadPermissionInfo(RawLocalFil eSystem.java:443) at org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.getPermission(RawLocalFileSyst em.java:418) Here is the detail output C:\spark-1.1.0\bin> spark-shell 15/01/29 10:13:13 INFO SecurityManager: Changing view acls to: ningjun.wang, 15/01/29 10:13:13 INFO SecurityManager: Changing modify acls to: ningjun.wang, 15/01/29 10:13:13 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(ningjun.wang, ); users with modify permissions: Set(ningjun.wang, ) 15/01/29 10:13:13 INFO HttpServer: Starting HTTP Server 15/01/29 10:13:14 INFO Server: jetty-8.y.z-SNAPSHOT 15/01/29 10:13:14 INFO AbstractConnector: Started SocketConnector@0.0.0.0:53692<http://SocketConnector@0.0.0.0:53692> 15/01/29 10:13:14 INFO Utils: Successfully started service 'HTTP class server' on port 53692. Failed to created SparkJLineReader: java.lang.NoClassDefFoundError: Could not initialize class org.f usesource.jansi.internal.Kernel32 Falling back to SimpleReader. Welcome to __ / __/__ ___ _/ /__ _\ \/ _ \/ _ `/ __/ '_/ /___/ .__/\_,_/_/ /_/\_\ version 1.1.0 /_/ Using Scala version 2.10.4 (Java HotSpot(TM) 64-Bit Server VM, Java 1.7.0_71) Type in expressions to have them evaluated. Type :help for more information. 15/01/29 10:13:18 INFO SecurityManager: Changing view acls to: ningjun.wang, 15/01/29 10:13:18 INFO SecurityManager: Changing modify acls to: ningjun.wang, 15/01/29 10:13:18 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(ningjun.wang, ); users with modify permissions: Set(ningjun.wang, ) 15/01/29 10:13:18 INFO Slf4jLogger: Slf4jLogger started 15/01/29 10:13:18 INFO Remoting: Starting remoting 15/01/29 10:13:19 INFO Remoting: Remoting started; listening on addresses :[akka.tcp://sparkDriver@L AB4-WIN01.pcc.lexisnexis.com:53705<http://AB4-WIN01.pcc.lexisnexis.com:53705>] 15/01/29 10:13:19 INFO Remoting: Remoting now listens on addresses: [akka.tcp://sparkDriver@LAB4-WIN 01.pcc.lexisnexis.com:53705<http://01.pcc.lexisnexis.com:53705>] 15/01/29 10:13:19 INFO Utils: Successfully started service 'sparkDriver' on port 53705. 15/01/29 10:13:19 INFO SparkEnv: Registering MapOutputTracker 15/01/29 10:13:19 INFO SparkEnv: Registering BlockManagerMaster 15/01/29 10:13:19 INFO DiskBlockManager: Created local directory at C:\Users\NINGJU~1.WAN\AppData\Lo cal\Temp\3\spark-local-20150129101319-f9da 15/01/29 10:13:19 INFO Utils: Successfully started service 'Connection manager for block manager' on port 53708. 15/01/29 10:13:19 INFO ConnectionManager: Bound socket to port 53708 with id = ConnectionManagerId(L AB4-WIN01.pcc.lexisnexis.com<http://AB4-WIN01.pcc.lexisnexis.com>,53708) 15/01/29 10:13:19 INFO MemoryStore: MemoryStore started with capacity 265.4 MB 15/01/29 10:13:19 INFO BlockManagerMaster: Trying to register BlockManager 15/01/29 10:13:19 INFO BlockManagerMasterActor: Registering block manager
Re: Fail to launch spark-shell on windows 2008 R2
Hi, I tried to use spark under windows once. However the only solution that I found is to install virtualbox Hope this can help you. Best Gen On Thu, Jan 29, 2015 at 4:18 PM, Wang, Ningjun (LNG-NPV) < ningjun.w...@lexisnexis.com> wrote: > I deployed spark-1.1.0 on Windows 7 and was albe to launch the > spark-shell. I then deploy it to windows 2008 R2 and launch the > spark-shell, I got the error > > > > java.lang.RuntimeException: Error while running command to get file > permissions : java.io.IOExceptio > > n: Cannot run program "ls": CreateProcess error=2, The system cannot find > the file specified > > at java.lang.ProcessBuilder.start(Unknown Source) > > at org.apache.hadoop.util.Shell.runCommand(Shell.java:200) > > at org.apache.hadoop.util.Shell.run(Shell.java:182) > > at > org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:375) > > at org.apache.hadoop.util.Shell.execCommand(Shell.java:461) > > at org.apache.hadoop.util.Shell.execCommand(Shell.java:444) > > at org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:710) > > at > org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.loadPermissionInfo(RawLocalFil > > eSystem.java:443) > > at > org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.getPermission(RawLocalFileSyst > > em.java:418) > > > > > > > > Here is the detail output > > > > > > C:\spark-1.1.0\bin> spark-shell > > 15/01/29 10:13:13 INFO SecurityManager: Changing view acls to: > ningjun.wang, > > 15/01/29 10:13:13 INFO SecurityManager: Changing modify acls to: > ningjun.wang, > > 15/01/29 10:13:13 INFO SecurityManager: SecurityManager: authentication > disabled; ui acls disabled; > > users with view permissions: Set(ningjun.wang, ); users with modify > permissions: Set(ningjun.wang, ) > > > > 15/01/29 10:13:13 INFO HttpServer: Starting HTTP Server > > 15/01/29 10:13:14 INFO Server: jetty-8.y.z-SNAPSHOT > > 15/01/29 10:13:14 INFO AbstractConnector: Started > SocketConnector@0.0.0.0:53692 > > 15/01/29 10:13:14 INFO Utils: Successfully started service 'HTTP class > server' on port 53692. > > Failed to created SparkJLineReader: java.lang.NoClassDefFoundError: Could > not initialize class org.f > > usesource.jansi.internal.Kernel32 > > Falling back to SimpleReader. > > Welcome to > > __ > > / __/__ ___ _/ /__ > > _\ \/ _ \/ _ `/ __/ '_/ > >/___/ .__/\_,_/_/ /_/\_\ version 1.1.0 > > /_/ > > > > Using Scala version 2.10.4 (Java HotSpot(TM) 64-Bit Server VM, Java > 1.7.0_71) > > Type in expressions to have them evaluated. > > Type :help for more information. > > 15/01/29 10:13:18 INFO SecurityManager: Changing view acls to: > ningjun.wang, > > 15/01/29 10:13:18 INFO SecurityManager: Changing modify acls to: > ningjun.wang, > > 15/01/29 10:13:18 INFO SecurityManager: SecurityManager: authentication > disabled; ui acls disabled; > > users with view permissions: Set(ningjun.wang, ); users with modify > permissions: Set(ningjun.wang, ) > > > > 15/01/29 10:13:18 INFO Slf4jLogger: Slf4jLogger started > > 15/01/29 10:13:18 INFO Remoting: Starting remoting > > 15/01/29 10:13:19 INFO Remoting: Remoting started; listening on addresses > :[akka.tcp://sparkDriver@L > > AB4-WIN01.pcc.lexisnexis.com:53705] > > 15/01/29 10:13:19 INFO Remoting: Remoting now listens on addresses: > [akka.tcp://sparkDriver@LAB4-WIN > > 01.pcc.lexisnexis.com:53705] > > 15/01/29 10:13:19 INFO Utils: Successfully started service 'sparkDriver' > on port 53705. > > 15/01/29 10:13:19 INFO SparkEnv: Registering MapOutputTracker > > 15/01/29 10:13:19 INFO SparkEnv: Registering BlockManagerMaster > > 15/01/29 10:13:19 INFO DiskBlockManager: Created local directory at > C:\Users\NINGJU~1.WAN\AppData\Lo > > cal\Temp\3\spark-local-20150129101319-f9da > > 15/01/29 10:13:19 INFO Utils: Successfully started service 'Connection > manager for block manager' on > > port 53708. > > 15/01/29 10:13:19 INFO ConnectionManager: Bound socket to port 53708 with > id = ConnectionManagerId(L > > AB4-WIN01.pcc.lexisnexis.com,53708) > > 15/01/29 10:13:19 INFO MemoryStore: MemoryStore started with capacity > 265.4 MB > > 15/01/29 10:13:19 INFO BlockManagerMaster: Trying to register BlockManager > > 15/01/29 10:13:19 INFO BlockManagerMasterActor: Registering block manager > LAB4-WIN01.pcc.lexisnexis. > > com:53708 with 265.4 MB RAM > > 15/01/29 10:13:19 INFO BlockManagerMaster: Registered BlockManager > > 15/01/29 10:13:19 INFO HttpFileServer: HTTP File server directory is > C:\Users\NINGJU~1.WAN\AppData\L > > ocal\Temp\3\spark-2f65b1c3-00e2-489b-967c-4e1f41520583 > > 15/01/29 10:13:19 INFO HttpServer: Starting HTTP Server > > 15/01/29 10:13:19 INFO Server: jetty-8.y.z-SNAPSHOT > > 15/01/29 10:13:19 INFO AbstractConnector: Started > SocketConnector@0.0.0.0:53709 > > 15/01/29 10:13:19 INFO Utils: Successfully started service 'HTTP file > server' on port 53709. > > 15/01/29 10:13:20 INFO Server: jet