Re: Fail to launch spark-shell on windows 2008 R2
Hi Ningjun, I have been working with Spark 1.2 on Windows 7 and Windows 2008 R2 (purely for development purposes). I had most recently installed them utilizing Java 1.8, Scala 2.10.4, and Spark 1.2 Precompiled for Hadoop 2.4+. A handy thread concerning the null\bin\winutils issue is addressed in an earlier thread at: http://apache-spark-user-list.1001560.n3.nabble.com/Run-spark-unit-test-on-Windows-7-td8656.html Hope this helps a little bit! Denny On Tue Feb 03 2015 at 8:24:24 AM Wang, Ningjun (LNG-NPV) ningjun.w...@lexisnexis.com wrote: Hi Gen Thanks for your feedback. We do have a business reason to run spark on windows. We have an existing application that is built on C# .NET running on windows. We are considering adding spark to the application for parallel processing of large data. We want spark to run on windows so it integrate with our existing app easily. Has anybody use spark on windows for production system? Is spark reliable on windows? Ningjun *From:* gen tang [mailto:gen.tan...@gmail.com] *Sent:* Thursday, January 29, 2015 12:53 PM *To:* Wang, Ningjun (LNG-NPV) *Cc:* user@spark.apache.org *Subject:* Re: Fail to launch spark-shell on windows 2008 R2 Hi, Using spark under windows is a really bad idea, because even you solve the problems about hadoop, you probably will meet the problem of java.net.SocketException. connection reset by peer. It is caused by the fact we ask socket port too frequently under windows. In my knowledge, it is really difficult to solve. And you will find something really funny: the same code sometimes works and sometimes not, even in the shell mode. And I am sorry but I don't see the interest to run spark under windows and moreover using local file system in a business environment. Do you have a cluster in windows? FYI, I have used spark prebuilt on hadoop 1 under windows 7 and there is no problem to launch, but have problem of java.net.SocketException. If you are using spark prebuilt on hadoop 2, you should consider follow the solution provided by https://issues.apache.org/jira/browse/SPARK-2356 Cheers Gen On Thu, Jan 29, 2015 at 5:54 PM, Wang, Ningjun (LNG-NPV) ningjun.w...@lexisnexis.com wrote: Install virtual box which run Linux? That does not help us. We have business reason to run it on Windows operating system, e.g. Windows 2008 R2. If anybody have done that, please give some advise on what version of spark, which version of Hadoop do you built spark against, etc…. Note that we only use local file system and do not have any hdfs file system at all. I don’t understand why spark generate so many error on Hadoop while we don’t even need hdfs. Ningjun *From:* gen tang [mailto:gen.tan...@gmail.com] *Sent:* Thursday, January 29, 2015 10:45 AM *To:* Wang, Ningjun (LNG-NPV) *Cc:* user@spark.apache.org *Subject:* Re: Fail to launch spark-shell on windows 2008 R2 Hi, I tried to use spark under windows once. However the only solution that I found is to install virtualbox Hope this can help you. Best Gen On Thu, Jan 29, 2015 at 4:18 PM, Wang, Ningjun (LNG-NPV) ningjun.w...@lexisnexis.com wrote: I deployed spark-1.1.0 on Windows 7 and was albe to launch the spark-shell. I then deploy it to windows 2008 R2 and launch the spark-shell, I got the error java.lang.RuntimeException: Error while running command to get file permissions : java.io.IOExceptio n: Cannot run program ls: CreateProcess error=2, The system cannot find the file specified at java.lang.ProcessBuilder.start(Unknown Source) at org.apache.hadoop.util.Shell.runCommand(Shell.java:200) at org.apache.hadoop.util.Shell.run(Shell.java:182) at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:375) at org.apache.hadoop.util.Shell.execCommand(Shell.java:461) at org.apache.hadoop.util.Shell.execCommand(Shell.java:444) at org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:710) at org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.loadPermissionInfo(RawLocalFil eSystem.java:443) at org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.getPermission(RawLocalFileSyst em.java:418) Here is the detail output C:\spark-1.1.0\bin spark-shell 15/01/29 10:13:13 INFO SecurityManager: Changing view acls to: ningjun.wang, 15/01/29 10:13:13 INFO SecurityManager: Changing modify acls to: ningjun.wang, 15/01/29 10:13:13 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(ningjun.wang, ); users with modify permissions: Set(ningjun.wang, ) 15/01/29 10:13:13 INFO HttpServer: Starting HTTP Server 15/01/29 10:13:14 INFO Server: jetty-8.y.z-SNAPSHOT 15/01/29 10:13:14 INFO AbstractConnector: Started SocketConnector@0.0.0.0:53692 15/01/29 10:13:14 INFO Utils: Successfully
RE: Fail to launch spark-shell on windows 2008 R2
Hi Gen Thanks for your feedback. We do have a business reason to run spark on windows. We have an existing application that is built on C# .NET running on windows. We are considering adding spark to the application for parallel processing of large data. We want spark to run on windows so it integrate with our existing app easily. Has anybody use spark on windows for production system? Is spark reliable on windows? Ningjun From: gen tang [mailto:gen.tan...@gmail.com] Sent: Thursday, January 29, 2015 12:53 PM To: Wang, Ningjun (LNG-NPV) Cc: user@spark.apache.org Subject: Re: Fail to launch spark-shell on windows 2008 R2 Hi, Using spark under windows is a really bad idea, because even you solve the problems about hadoop, you probably will meet the problem of java.net.SocketException. connection reset by peer. It is caused by the fact we ask socket port too frequently under windows. In my knowledge, it is really difficult to solve. And you will find something really funny: the same code sometimes works and sometimes not, even in the shell mode. And I am sorry but I don't see the interest to run spark under windows and moreover using local file system in a business environment. Do you have a cluster in windows? FYI, I have used spark prebuilt on hadoop 1 under windows 7 and there is no problem to launch, but have problem of java.net.SocketException. If you are using spark prebuilt on hadoop 2, you should consider follow the solution provided by https://issues.apache.org/jira/browse/SPARK-2356 Cheers Gen On Thu, Jan 29, 2015 at 5:54 PM, Wang, Ningjun (LNG-NPV) ningjun.w...@lexisnexis.commailto:ningjun.w...@lexisnexis.com wrote: Install virtual box which run Linux? That does not help us. We have business reason to run it on Windows operating system, e.g. Windows 2008 R2. If anybody have done that, please give some advise on what version of spark, which version of Hadoop do you built spark against, etc…. Note that we only use local file system and do not have any hdfs file system at all. I don’t understand why spark generate so many error on Hadoop while we don’t even need hdfs. Ningjun From: gen tang [mailto:gen.tan...@gmail.commailto:gen.tan...@gmail.com] Sent: Thursday, January 29, 2015 10:45 AM To: Wang, Ningjun (LNG-NPV) Cc: user@spark.apache.orgmailto:user@spark.apache.org Subject: Re: Fail to launch spark-shell on windows 2008 R2 Hi, I tried to use spark under windows once. However the only solution that I found is to install virtualbox Hope this can help you. Best Gen On Thu, Jan 29, 2015 at 4:18 PM, Wang, Ningjun (LNG-NPV) ningjun.w...@lexisnexis.commailto:ningjun.w...@lexisnexis.com wrote: I deployed spark-1.1.0 on Windows 7 and was albe to launch the spark-shell. I then deploy it to windows 2008 R2 and launch the spark-shell, I got the error java.lang.RuntimeException: Error while running command to get file permissions : java.io.IOExceptio n: Cannot run program ls: CreateProcess error=2, The system cannot find the file specified at java.lang.ProcessBuilder.start(Unknown Source) at org.apache.hadoop.util.Shell.runCommand(Shell.java:200) at org.apache.hadoop.util.Shell.run(Shell.java:182) at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:375) at org.apache.hadoop.util.Shell.execCommand(Shell.java:461) at org.apache.hadoop.util.Shell.execCommand(Shell.java:444) at org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:710) at org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.loadPermissionInfo(RawLocalFil eSystem.java:443) at org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.getPermission(RawLocalFileSyst em.java:418) Here is the detail output C:\spark-1.1.0\bin spark-shell 15/01/29 10:13:13 INFO SecurityManager: Changing view acls to: ningjun.wang, 15/01/29 10:13:13 INFO SecurityManager: Changing modify acls to: ningjun.wang, 15/01/29 10:13:13 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(ningjun.wang, ); users with modify permissions: Set(ningjun.wang, ) 15/01/29 10:13:13 INFO HttpServer: Starting HTTP Server 15/01/29 10:13:14 INFO Server: jetty-8.y.z-SNAPSHOT 15/01/29 10:13:14 INFO AbstractConnector: Started SocketConnector@0.0.0.0:53692http://SocketConnector@0.0.0.0:53692 15/01/29 10:13:14 INFO Utils: Successfully started service 'HTTP class server' on port 53692. Failed to created SparkJLineReader: java.lang.NoClassDefFoundError: Could not initialize class org.f usesource.jansi.internal.Kernel32 Falling back to SimpleReader. Welcome to __ / __/__ ___ _/ /__ _\ \/ _ \/ _ `/ __/ '_/ /___/ .__/\_,_/_/ /_/\_\ version 1.1.0 /_/ Using Scala version 2.10.4 (Java HotSpot(TM) 64-Bit Server VM, Java 1.7.0_71) Type in expressions to have them evaluated. Type :help for more information. 15/01/29 10:13:18 INFO
RE: Fail to launch spark-shell on windows 2008 R2
Install virtual box which run Linux? That does not help us. We have business reason to run it on Windows operating system, e.g. Windows 2008 R2. If anybody have done that, please give some advise on what version of spark, which version of Hadoop do you built spark against, etc…. Note that we only use local file system and do not have any hdfs file system at all. I don’t understand why spark generate so many error on Hadoop while we don’t even need hdfs. Ningjun From: gen tang [mailto:gen.tan...@gmail.com] Sent: Thursday, January 29, 2015 10:45 AM To: Wang, Ningjun (LNG-NPV) Cc: user@spark.apache.org Subject: Re: Fail to launch spark-shell on windows 2008 R2 Hi, I tried to use spark under windows once. However the only solution that I found is to install virtualbox Hope this can help you. Best Gen On Thu, Jan 29, 2015 at 4:18 PM, Wang, Ningjun (LNG-NPV) ningjun.w...@lexisnexis.commailto:ningjun.w...@lexisnexis.com wrote: I deployed spark-1.1.0 on Windows 7 and was albe to launch the spark-shell. I then deploy it to windows 2008 R2 and launch the spark-shell, I got the error java.lang.RuntimeException: Error while running command to get file permissions : java.io.IOExceptio n: Cannot run program ls: CreateProcess error=2, The system cannot find the file specified at java.lang.ProcessBuilder.start(Unknown Source) at org.apache.hadoop.util.Shell.runCommand(Shell.java:200) at org.apache.hadoop.util.Shell.run(Shell.java:182) at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:375) at org.apache.hadoop.util.Shell.execCommand(Shell.java:461) at org.apache.hadoop.util.Shell.execCommand(Shell.java:444) at org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:710) at org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.loadPermissionInfo(RawLocalFil eSystem.java:443) at org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.getPermission(RawLocalFileSyst em.java:418) Here is the detail output C:\spark-1.1.0\bin spark-shell 15/01/29 10:13:13 INFO SecurityManager: Changing view acls to: ningjun.wang, 15/01/29 10:13:13 INFO SecurityManager: Changing modify acls to: ningjun.wang, 15/01/29 10:13:13 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(ningjun.wang, ); users with modify permissions: Set(ningjun.wang, ) 15/01/29 10:13:13 INFO HttpServer: Starting HTTP Server 15/01/29 10:13:14 INFO Server: jetty-8.y.z-SNAPSHOT 15/01/29 10:13:14 INFO AbstractConnector: Started SocketConnector@0.0.0.0:53692http://SocketConnector@0.0.0.0:53692 15/01/29 10:13:14 INFO Utils: Successfully started service 'HTTP class server' on port 53692. Failed to created SparkJLineReader: java.lang.NoClassDefFoundError: Could not initialize class org.f usesource.jansi.internal.Kernel32 Falling back to SimpleReader. Welcome to __ / __/__ ___ _/ /__ _\ \/ _ \/ _ `/ __/ '_/ /___/ .__/\_,_/_/ /_/\_\ version 1.1.0 /_/ Using Scala version 2.10.4 (Java HotSpot(TM) 64-Bit Server VM, Java 1.7.0_71) Type in expressions to have them evaluated. Type :help for more information. 15/01/29 10:13:18 INFO SecurityManager: Changing view acls to: ningjun.wang, 15/01/29 10:13:18 INFO SecurityManager: Changing modify acls to: ningjun.wang, 15/01/29 10:13:18 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(ningjun.wang, ); users with modify permissions: Set(ningjun.wang, ) 15/01/29 10:13:18 INFO Slf4jLogger: Slf4jLogger started 15/01/29 10:13:18 INFO Remoting: Starting remoting 15/01/29 10:13:19 INFO Remoting: Remoting started; listening on addresses :[akka.tcp://sparkDriver@L AB4-WIN01.pcc.lexisnexis.com:53705http://AB4-WIN01.pcc.lexisnexis.com:53705] 15/01/29 10:13:19 INFO Remoting: Remoting now listens on addresses: [akka.tcp://sparkDriver@LAB4-WIN 01.pcc.lexisnexis.com:53705http://01.pcc.lexisnexis.com:53705] 15/01/29 10:13:19 INFO Utils: Successfully started service 'sparkDriver' on port 53705. 15/01/29 10:13:19 INFO SparkEnv: Registering MapOutputTracker 15/01/29 10:13:19 INFO SparkEnv: Registering BlockManagerMaster 15/01/29 10:13:19 INFO DiskBlockManager: Created local directory at C:\Users\NINGJU~1.WAN\AppData\Lo cal\Temp\3\spark-local-20150129101319-f9da 15/01/29 10:13:19 INFO Utils: Successfully started service 'Connection manager for block manager' on port 53708. 15/01/29 10:13:19 INFO ConnectionManager: Bound socket to port 53708 with id = ConnectionManagerId(L AB4-WIN01.pcc.lexisnexis.comhttp://AB4-WIN01.pcc.lexisnexis.com,53708) 15/01/29 10:13:19 INFO MemoryStore: MemoryStore started with capacity 265.4 MB 15/01/29 10:13:19 INFO BlockManagerMaster: Trying to register BlockManager 15/01/29 10:13:19 INFO BlockManagerMasterActor: Registering block manager LAB4-WIN01.pcc.lexisnexis. com:53708 with 265.4 MB RAM 15/01/29 10:13:19
Re: Fail to launch spark-shell on windows 2008 R2
Hi, Using spark under windows is a really bad idea, because even you solve the problems about hadoop, you probably will meet the problem of java.net.SocketException. connection reset by peer. It is caused by the fact we ask socket port too frequently under windows. In my knowledge, it is really difficult to solve. And you will find something really funny: the same code sometimes works and sometimes not, even in the shell mode. And I am sorry but I don't see the interest to run spark under windows and moreover using local file system in a business environment. Do you have a cluster in windows? FYI, I have used spark prebuilt on hadoop 1 under windows 7 and there is no problem to launch, but have problem of java.net.SocketException. If you are using spark prebuilt on hadoop 2, you should consider follow the solution provided by https://issues.apache.org/jira/browse/SPARK-2356 Cheers Gen On Thu, Jan 29, 2015 at 5:54 PM, Wang, Ningjun (LNG-NPV) ningjun.w...@lexisnexis.com wrote: Install virtual box which run Linux? That does not help us. We have business reason to run it on Windows operating system, e.g. Windows 2008 R2. If anybody have done that, please give some advise on what version of spark, which version of Hadoop do you built spark against, etc…. Note that we only use local file system and do not have any hdfs file system at all. I don’t understand why spark generate so many error on Hadoop while we don’t even need hdfs. Ningjun *From:* gen tang [mailto:gen.tan...@gmail.com] *Sent:* Thursday, January 29, 2015 10:45 AM *To:* Wang, Ningjun (LNG-NPV) *Cc:* user@spark.apache.org *Subject:* Re: Fail to launch spark-shell on windows 2008 R2 Hi, I tried to use spark under windows once. However the only solution that I found is to install virtualbox Hope this can help you. Best Gen On Thu, Jan 29, 2015 at 4:18 PM, Wang, Ningjun (LNG-NPV) ningjun.w...@lexisnexis.com wrote: I deployed spark-1.1.0 on Windows 7 and was albe to launch the spark-shell. I then deploy it to windows 2008 R2 and launch the spark-shell, I got the error java.lang.RuntimeException: Error while running command to get file permissions : java.io.IOExceptio n: Cannot run program ls: CreateProcess error=2, The system cannot find the file specified at java.lang.ProcessBuilder.start(Unknown Source) at org.apache.hadoop.util.Shell.runCommand(Shell.java:200) at org.apache.hadoop.util.Shell.run(Shell.java:182) at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:375) at org.apache.hadoop.util.Shell.execCommand(Shell.java:461) at org.apache.hadoop.util.Shell.execCommand(Shell.java:444) at org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:710) at org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.loadPermissionInfo(RawLocalFil eSystem.java:443) at org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.getPermission(RawLocalFileSyst em.java:418) Here is the detail output C:\spark-1.1.0\bin spark-shell 15/01/29 10:13:13 INFO SecurityManager: Changing view acls to: ningjun.wang, 15/01/29 10:13:13 INFO SecurityManager: Changing modify acls to: ningjun.wang, 15/01/29 10:13:13 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(ningjun.wang, ); users with modify permissions: Set(ningjun.wang, ) 15/01/29 10:13:13 INFO HttpServer: Starting HTTP Server 15/01/29 10:13:14 INFO Server: jetty-8.y.z-SNAPSHOT 15/01/29 10:13:14 INFO AbstractConnector: Started SocketConnector@0.0.0.0:53692 15/01/29 10:13:14 INFO Utils: Successfully started service 'HTTP class server' on port 53692. Failed to created SparkJLineReader: java.lang.NoClassDefFoundError: Could not initialize class org.f usesource.jansi.internal.Kernel32 Falling back to SimpleReader. Welcome to __ / __/__ ___ _/ /__ _\ \/ _ \/ _ `/ __/ '_/ /___/ .__/\_,_/_/ /_/\_\ version 1.1.0 /_/ Using Scala version 2.10.4 (Java HotSpot(TM) 64-Bit Server VM, Java 1.7.0_71) Type in expressions to have them evaluated. Type :help for more information. 15/01/29 10:13:18 INFO SecurityManager: Changing view acls to: ningjun.wang, 15/01/29 10:13:18 INFO SecurityManager: Changing modify acls to: ningjun.wang, 15/01/29 10:13:18 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(ningjun.wang, ); users with modify permissions: Set(ningjun.wang, ) 15/01/29 10:13:18 INFO Slf4jLogger: Slf4jLogger started 15/01/29 10:13:18 INFO Remoting: Starting remoting 15/01/29 10:13:19 INFO Remoting: Remoting started; listening on addresses :[akka.tcp://sparkDriver@L AB4-WIN01.pcc.lexisnexis.com:53705] 15/01/29 10:13:19 INFO Remoting: Remoting now listens on addresses: [akka.tcp
Re: Fail to launch spark-shell on windows 2008 R2
Hi, I tried to use spark under windows once. However the only solution that I found is to install virtualbox Hope this can help you. Best Gen On Thu, Jan 29, 2015 at 4:18 PM, Wang, Ningjun (LNG-NPV) ningjun.w...@lexisnexis.com wrote: I deployed spark-1.1.0 on Windows 7 and was albe to launch the spark-shell. I then deploy it to windows 2008 R2 and launch the spark-shell, I got the error java.lang.RuntimeException: Error while running command to get file permissions : java.io.IOExceptio n: Cannot run program ls: CreateProcess error=2, The system cannot find the file specified at java.lang.ProcessBuilder.start(Unknown Source) at org.apache.hadoop.util.Shell.runCommand(Shell.java:200) at org.apache.hadoop.util.Shell.run(Shell.java:182) at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:375) at org.apache.hadoop.util.Shell.execCommand(Shell.java:461) at org.apache.hadoop.util.Shell.execCommand(Shell.java:444) at org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:710) at org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.loadPermissionInfo(RawLocalFil eSystem.java:443) at org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.getPermission(RawLocalFileSyst em.java:418) Here is the detail output C:\spark-1.1.0\bin spark-shell 15/01/29 10:13:13 INFO SecurityManager: Changing view acls to: ningjun.wang, 15/01/29 10:13:13 INFO SecurityManager: Changing modify acls to: ningjun.wang, 15/01/29 10:13:13 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(ningjun.wang, ); users with modify permissions: Set(ningjun.wang, ) 15/01/29 10:13:13 INFO HttpServer: Starting HTTP Server 15/01/29 10:13:14 INFO Server: jetty-8.y.z-SNAPSHOT 15/01/29 10:13:14 INFO AbstractConnector: Started SocketConnector@0.0.0.0:53692 15/01/29 10:13:14 INFO Utils: Successfully started service 'HTTP class server' on port 53692. Failed to created SparkJLineReader: java.lang.NoClassDefFoundError: Could not initialize class org.f usesource.jansi.internal.Kernel32 Falling back to SimpleReader. Welcome to __ / __/__ ___ _/ /__ _\ \/ _ \/ _ `/ __/ '_/ /___/ .__/\_,_/_/ /_/\_\ version 1.1.0 /_/ Using Scala version 2.10.4 (Java HotSpot(TM) 64-Bit Server VM, Java 1.7.0_71) Type in expressions to have them evaluated. Type :help for more information. 15/01/29 10:13:18 INFO SecurityManager: Changing view acls to: ningjun.wang, 15/01/29 10:13:18 INFO SecurityManager: Changing modify acls to: ningjun.wang, 15/01/29 10:13:18 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(ningjun.wang, ); users with modify permissions: Set(ningjun.wang, ) 15/01/29 10:13:18 INFO Slf4jLogger: Slf4jLogger started 15/01/29 10:13:18 INFO Remoting: Starting remoting 15/01/29 10:13:19 INFO Remoting: Remoting started; listening on addresses :[akka.tcp://sparkDriver@L AB4-WIN01.pcc.lexisnexis.com:53705] 15/01/29 10:13:19 INFO Remoting: Remoting now listens on addresses: [akka.tcp://sparkDriver@LAB4-WIN 01.pcc.lexisnexis.com:53705] 15/01/29 10:13:19 INFO Utils: Successfully started service 'sparkDriver' on port 53705. 15/01/29 10:13:19 INFO SparkEnv: Registering MapOutputTracker 15/01/29 10:13:19 INFO SparkEnv: Registering BlockManagerMaster 15/01/29 10:13:19 INFO DiskBlockManager: Created local directory at C:\Users\NINGJU~1.WAN\AppData\Lo cal\Temp\3\spark-local-20150129101319-f9da 15/01/29 10:13:19 INFO Utils: Successfully started service 'Connection manager for block manager' on port 53708. 15/01/29 10:13:19 INFO ConnectionManager: Bound socket to port 53708 with id = ConnectionManagerId(L AB4-WIN01.pcc.lexisnexis.com,53708) 15/01/29 10:13:19 INFO MemoryStore: MemoryStore started with capacity 265.4 MB 15/01/29 10:13:19 INFO BlockManagerMaster: Trying to register BlockManager 15/01/29 10:13:19 INFO BlockManagerMasterActor: Registering block manager LAB4-WIN01.pcc.lexisnexis. com:53708 with 265.4 MB RAM 15/01/29 10:13:19 INFO BlockManagerMaster: Registered BlockManager 15/01/29 10:13:19 INFO HttpFileServer: HTTP File server directory is C:\Users\NINGJU~1.WAN\AppData\L ocal\Temp\3\spark-2f65b1c3-00e2-489b-967c-4e1f41520583 15/01/29 10:13:19 INFO HttpServer: Starting HTTP Server 15/01/29 10:13:19 INFO Server: jetty-8.y.z-SNAPSHOT 15/01/29 10:13:19 INFO AbstractConnector: Started SocketConnector@0.0.0.0:53709 15/01/29 10:13:19 INFO Utils: Successfully started service 'HTTP file server' on port 53709. 15/01/29 10:13:20 INFO Server: jetty-8.y.z-SNAPSHOT 15/01/29 10:13:20 INFO AbstractConnector: Started SelectChannelConnector@0.0.0.0:4040 15/01/29 10:13:20 INFO Utils: Successfully started service 'SparkUI' on