Hi, I tried to use spark under windows once. However the only solution that I found is to install virtualbox....
Hope this can help you. Best Gen On Thu, Jan 29, 2015 at 4:18 PM, Wang, Ningjun (LNG-NPV) < ningjun.w...@lexisnexis.com> wrote: > I deployed spark-1.1.0 on Windows 7 and was albe to launch the > spark-shell. I then deploy it to windows 2008 R2 and launch the > spark-shell, I got the error > > > > java.lang.RuntimeException: Error while running command to get file > permissions : java.io.IOExceptio > > n: Cannot run program "ls": CreateProcess error=2, The system cannot find > the file specified > > at java.lang.ProcessBuilder.start(Unknown Source) > > at org.apache.hadoop.util.Shell.runCommand(Shell.java:200) > > at org.apache.hadoop.util.Shell.run(Shell.java:182) > > at > org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:375) > > at org.apache.hadoop.util.Shell.execCommand(Shell.java:461) > > at org.apache.hadoop.util.Shell.execCommand(Shell.java:444) > > at org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:710) > > at > org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.loadPermissionInfo(RawLocalFil > > eSystem.java:443) > > at > org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.getPermission(RawLocalFileSyst > > em.java:418) > > > > > > > > Here is the detail output > > > > > > C:\spark-1.1.0\bin> spark-shell > > 15/01/29 10:13:13 INFO SecurityManager: Changing view acls to: > ningjun.wang, > > 15/01/29 10:13:13 INFO SecurityManager: Changing modify acls to: > ningjun.wang, > > 15/01/29 10:13:13 INFO SecurityManager: SecurityManager: authentication > disabled; ui acls disabled; > > users with view permissions: Set(ningjun.wang, ); users with modify > permissions: Set(ningjun.wang, ) > > > > 15/01/29 10:13:13 INFO HttpServer: Starting HTTP Server > > 15/01/29 10:13:14 INFO Server: jetty-8.y.z-SNAPSHOT > > 15/01/29 10:13:14 INFO AbstractConnector: Started > SocketConnector@0.0.0.0:53692 > > 15/01/29 10:13:14 INFO Utils: Successfully started service 'HTTP class > server' on port 53692. > > Failed to created SparkJLineReader: java.lang.NoClassDefFoundError: Could > not initialize class org.f > > usesource.jansi.internal.Kernel32 > > Falling back to SimpleReader. > > Welcome to > > ____ __ > > / __/__ ___ _____/ /__ > > _\ \/ _ \/ _ `/ __/ '_/ > > /___/ .__/\_,_/_/ /_/\_\ version 1.1.0 > > /_/ > > > > Using Scala version 2.10.4 (Java HotSpot(TM) 64-Bit Server VM, Java > 1.7.0_71) > > Type in expressions to have them evaluated. > > Type :help for more information. > > 15/01/29 10:13:18 INFO SecurityManager: Changing view acls to: > ningjun.wang, > > 15/01/29 10:13:18 INFO SecurityManager: Changing modify acls to: > ningjun.wang, > > 15/01/29 10:13:18 INFO SecurityManager: SecurityManager: authentication > disabled; ui acls disabled; > > users with view permissions: Set(ningjun.wang, ); users with modify > permissions: Set(ningjun.wang, ) > > > > 15/01/29 10:13:18 INFO Slf4jLogger: Slf4jLogger started > > 15/01/29 10:13:18 INFO Remoting: Starting remoting > > 15/01/29 10:13:19 INFO Remoting: Remoting started; listening on addresses > :[akka.tcp://sparkDriver@L > > AB4-WIN01.pcc.lexisnexis.com:53705] > > 15/01/29 10:13:19 INFO Remoting: Remoting now listens on addresses: > [akka.tcp://sparkDriver@LAB4-WIN > > 01.pcc.lexisnexis.com:53705] > > 15/01/29 10:13:19 INFO Utils: Successfully started service 'sparkDriver' > on port 53705. > > 15/01/29 10:13:19 INFO SparkEnv: Registering MapOutputTracker > > 15/01/29 10:13:19 INFO SparkEnv: Registering BlockManagerMaster > > 15/01/29 10:13:19 INFO DiskBlockManager: Created local directory at > C:\Users\NINGJU~1.WAN\AppData\Lo > > cal\Temp\3\spark-local-20150129101319-f9da > > 15/01/29 10:13:19 INFO Utils: Successfully started service 'Connection > manager for block manager' on > > port 53708. > > 15/01/29 10:13:19 INFO ConnectionManager: Bound socket to port 53708 with > id = ConnectionManagerId(L > > AB4-WIN01.pcc.lexisnexis.com,53708) > > 15/01/29 10:13:19 INFO MemoryStore: MemoryStore started with capacity > 265.4 MB > > 15/01/29 10:13:19 INFO BlockManagerMaster: Trying to register BlockManager > > 15/01/29 10:13:19 INFO BlockManagerMasterActor: Registering block manager > LAB4-WIN01.pcc.lexisnexis. > > com:53708 with 265.4 MB RAM > > 15/01/29 10:13:19 INFO BlockManagerMaster: Registered BlockManager > > 15/01/29 10:13:19 INFO HttpFileServer: HTTP File server directory is > C:\Users\NINGJU~1.WAN\AppData\L > > ocal\Temp\3\spark-2f65b1c3-00e2-489b-967c-4e1f41520583 > > 15/01/29 10:13:19 INFO HttpServer: Starting HTTP Server > > 15/01/29 10:13:19 INFO Server: jetty-8.y.z-SNAPSHOT > > 15/01/29 10:13:19 INFO AbstractConnector: Started > SocketConnector@0.0.0.0:53709 > > 15/01/29 10:13:19 INFO Utils: Successfully started service 'HTTP file > server' on port 53709. > > 15/01/29 10:13:20 INFO Server: jetty-8.y.z-SNAPSHOT > > 15/01/29 10:13:20 INFO AbstractConnector: Started > SelectChannelConnector@0.0.0.0:4040 > > 15/01/29 10:13:20 INFO Utils: Successfully started service 'SparkUI' on > port 4040. > > 15/01/29 10:13:20 INFO SparkUI: Started SparkUI at > http://LAB4-WIN01.pcc.lexisnexis.com:4040 > > java.lang.RuntimeException: Error while running command to get file > permissions : java.io.IOExceptio > > n: Cannot run program "ls": CreateProcess error=2, The system cannot find > the file specified > > at java.lang.ProcessBuilder.start(Unknown Source) > > at org.apache.hadoop.util.Shell.runCommand(Shell.java:200) > > at org.apache.hadoop.util.Shell.run(Shell.java:182) > > at > org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:375) > > at org.apache.hadoop.util.Shell.execCommand(Shell.java:461) > > at org.apache.hadoop.util.Shell.execCommand(Shell.java:444) > > at org.apache.hadoop.fs.FileUtil.execCommand(FileUtil.java:710) > > at > org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.loadPermissionInfo(RawLocalFil > > eSystem.java:443) > > at > org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.getPermission(RawLocalFileSyst > > em.java:418) > > at > org.apache.spark.util.FileLogger.createLogDir(FileLogger.scala:113) > > at org.apache.spark.util.FileLogger.start(FileLogger.scala:91) > > at > org.apache.spark.scheduler.EventLoggingListener.start(EventLoggingListener.scala:79) > > at org.apache.spark.SparkContext.<init>(SparkContext.scala:252) > > at > org.apache.spark.repl.SparkILoop.createSparkContext(SparkILoop.scala:972) > > at $iwC$$iwC.<init>(<console>:8) > > at $iwC.<init>(<console>:14) > > at <init>(<console>:16) > > at .<init>(<console>:20) > > at .<clinit>(<console>) > > at .<init>(<console>:7) > > at .<clinit>(<console>) > > at $print(<console>) > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source) > > at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) > > at java.lang.reflect.Method.invoke(Unknown Source) > > at > org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:789) > > at > org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkIMain.scala:1062) > > at > org.apache.spark.repl.SparkIMain.loadAndRunReq$1(SparkIMain.scala:615) > > at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:646) > > at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:610) > > at > org.apache.spark.repl.SparkILoop.reallyInterpret$1(SparkILoop.scala:814) > > at > org.apache.spark.repl.SparkILoop.interpretStartingWith(SparkILoop.scala:859) > > at org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:771) > > at > org.apache.spark.repl.SparkILoopInit$$anonfun$initializeSpark$1.apply(SparkILoopInit.scal > > a:121) > > at > org.apache.spark.repl.SparkILoopInit$$anonfun$initializeSpark$1.apply(SparkILoopInit.scal > > a:120) > > at > org.apache.spark.repl.SparkIMain.beQuietDuring(SparkIMain.scala:264) > > at > org.apache.spark.repl.SparkILoopInit$class.initializeSpark(SparkILoopInit.scala:120) > > at > org.apache.spark.repl.SparkILoop.initializeSpark(SparkILoop.scala:56) > > at > org.apache.spark.repl.SparkILoop$$anonfun$process$1$$anonfun$apply$mcZ$sp$5.apply$mcV$sp( > > SparkILoop.scala:931) > > at > org.apache.spark.repl.SparkILoopInit$class.runThunks(SparkILoopInit.scala:142) > > at org.apache.spark.repl.SparkILoop.runThunks(SparkILoop.scala:56) > > at > org.apache.spark.repl.SparkILoopInit$class.postInitialization(SparkILoopInit.scala:104) > > at > org.apache.spark.repl.SparkILoop.postInitialization(SparkILoop.scala:56) > > at > org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply$mcZ$sp(SparkILoop.scala:948) > > at > org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply(SparkILoop.scala:902) > > at > org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply(SparkILoop.scala:902) > > at > scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:135) > > at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:902) > > at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:997) > > at org.apache.spark.repl.Main$.main(Main.scala:31) > > at org.apache.spark.repl.Main.main(Main.scala) > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source) > > at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) > > at java.lang.reflect.Method.invoke(Unknown Source) > > at > org.apache.spark.deploy.SparkSubmit$.launch(SparkSubmit.scala:328) > > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:75) > > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > > Caused by: java.io.IOException: CreateProcess error=2, The system cannot > find the file specified > > at java.lang.ProcessImpl.create(Native Method) > > at java.lang.ProcessImpl.<init>(Unknown Source) > > at java.lang.ProcessImpl.start(Unknown Source) > > ... 59 more > > > > at > org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.loadPermissionInfo(RawLocalFil > > eSystem.java:468) > > at > org.apache.hadoop.fs.RawLocalFileSystem$RawLocalFileStatus.getPermission(RawLocalFileSyst > > em.java:418) > > at > org.apache.spark.util.FileLogger.createLogDir(FileLogger.scala:113) > > at org.apache.spark.util.FileLogger.start(FileLogger.scala:91) > > at > org.apache.spark.scheduler.EventLoggingListener.start(EventLoggingListener.scala:79) > > at org.apache.spark.SparkContext.<init>(SparkContext.scala:252) > > at > org.apache.spark.repl.SparkILoop.createSparkContext(SparkILoop.scala:972) > > at $iwC$$iwC.<init>(<console>:8) > > at $iwC.<init>(<console>:14) > > at <init>(<console>:16) > > at .<init>(<console>:20) > > at .<clinit>(<console>) > > at .<init>(<console>:7) > > at .<clinit>(<console>) > > at $print(<console>) > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source) > > at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) > > at java.lang.reflect.Method.invoke(Unknown Source) > > at > org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:789) > > at > org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkIMain.scala:1062) > > at > org.apache.spark.repl.SparkIMain.loadAndRunReq$1(SparkIMain.scala:615) > > at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:646) > > at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:610) > > at > org.apache.spark.repl.SparkILoop.reallyInterpret$1(SparkILoop.scala:814) > > at > org.apache.spark.repl.SparkILoop.interpretStartingWith(SparkILoop.scala:859) > > at org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:771) > > at > org.apache.spark.repl.SparkILoopInit$$anonfun$initializeSpark$1.apply(SparkILoopInit.scal > > a:121) > > at > org.apache.spark.repl.SparkILoopInit$$anonfun$initializeSpark$1.apply(SparkILoopInit.scal > > a:120) > > at > org.apache.spark.repl.SparkIMain.beQuietDuring(SparkIMain.scala:264) > > at > org.apache.spark.repl.SparkILoopInit$class.initializeSpark(SparkILoopInit.scala:120) > > at > org.apache.spark.repl.SparkILoop.initializeSpark(SparkILoop.scala:56) > > at > org.apache.spark.repl.SparkILoop$$anonfun$process$1$$anonfun$apply$mcZ$sp$5.apply$mcV$sp( > > SparkILoop.scala:931) > > at > org.apache.spark.repl.SparkILoopInit$class.runThunks(SparkILoopInit.scala:142) > > at org.apache.spark.repl.SparkILoop.runThunks(SparkILoop.scala:56) > > at > org.apache.spark.repl.SparkILoopInit$class.postInitialization(SparkILoopInit.scala:104) > > at > org.apache.spark.repl.SparkILoop.postInitialization(SparkILoop.scala:56) > > at > org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply$mcZ$sp(SparkILoop.scala:948) > > at > org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply(SparkILoop.scala:902) > > at > org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply(SparkILoop.scala:902) > > at > scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:135) > > at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:902) > > at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:997) > > at org.apache.spark.repl.Main$.main(Main.scala:31) > > at org.apache.spark.repl.Main.main(Main.scala) > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source) > > at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) > > at java.lang.reflect.Method.invoke(Unknown Source) > > at > org.apache.spark.deploy.SparkSubmit$.launch(SparkSubmit.scala:328) > > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:75) > > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > > > > Spark context available as sc. > > > > > > Please advise > > > > > > Ningjun > > >