You can also run Java & Python API.
though sbt should run.
Regards
Mayur

Mayur Rustagi
Ph: +919632149971
h <https://twitter.com/mayur_rustagi>ttp://www.sigmoidanalytics.com
https://twitter.com/mayur_rustagi



On Thu, Feb 20, 2014 at 3:54 AM, lihu <[email protected]> wrote:

> When I use the* sbt/sbt run* command, then occurred the following
> exception:
>
> *java.io.IOException: No locks available*
> * at sun.nio.ch.FileChannelImpl.lock0(Native Method)*
> * at sun.nio.ch.FileChannelImpl.tryLock(FileChannelImpl.java:871)*
> * at java.nio.channels.FileChannel.tryLock(FileChannel.java:962)*
> * at xsbt.boot.Locks$GlobalLock.withChannel$1(Locks.scala:65)*
> * at xsbt.boot.Locks$GlobalLock.withChannelRetries$1(Locks.scala:58)*
> * at
> xsbt.boot.Locks$GlobalLock$$anonfun$withFileLock$1.apply(Locks.scala:79)*
> * at xsbt.boot.Using$.withResource(Using.scala:11)*
> * at xsbt.boot.Using$.apply(Using.scala:10)*
> * at xsbt.boot.Locks$GlobalLock.liftedTree1$1(Locks.scala:51)*
> * at xsbt.boot.Locks$GlobalLock.withLock(Locks.scala:51)*
> * at xsbt.boot.Locks$.apply0(Locks.scala:30)*
> * at xsbt.boot.Locks$.apply(Locks.scala:27)*
> * at xsbt.boot.Update.apply(Checks.java:100)*
> * at xsbt.boot.Launch.update(Launch.scala:274)*
> * at xsbt.boot.Launch$$anonfun$jnaLoader$1.apply(Launch.scala:119)*
> * at scala.Option.getOrElse(Option.scala:108)*
> * at xsbt.boot.Launch.jnaLoader$2f324eef(Launch.scala:114)*
> * at xsbt.boot.Launch.<init>(Launch.scala:94)*
> * at xsbt.boot.Launcher$.apply(Launch.scala:289)*
> * at xsbt.boot.Launch$.apply(Launch.scala:16)*
> * at xsbt.boot.Boot$.runImpl(Boot.scala:31)*
> * at xsbt.boot.Boot$.main(Boot.scala:20)*
> * at xsbt.boot.Boot.main(Boot.scala)*
> *Error during sbt execution: java.io.IOException: No locks available*
>
>
> On Thu, Feb 20, 2014 at 7:51 PM, lihu <[email protected]> wrote:
>
>> Hi,
>>
>>     I run the spark on the cluster, but our cluster can not use the *sbt
>> run* command, so I can not write my own app, I can only use the
>> spark-shell and run-example now, this is not very  convenient,  can anyone
>> know how to solve this,  this question may not very relate with spark, but
>> I can not find any suggestion through the search. Thanks!
>>
>>
>>
>>
>>
>>
>
>
> --
> *Best Wishes!*
>
> *Li Hu(李浒) | Graduate Student*
>
> *Institute for Interdisciplinary Information Sciences(IIIS
> <http://iiis.tsinghua.edu.cn/>)*
> *Tsinghua University, China*
>
> *Email: [email protected] <[email protected]>*
> *Tel  : +86 15120081920 <%2B86%2015120081920>*
> *Homepage: http://iiis.tsinghua.edu.cn/zh/lihu/
> <http://iiis.tsinghua.edu.cn/zh/lihu/>*
>
>
>

Reply via email to