Try passing jar using --jars option
On Wed, Mar 2, 2016 at 10:17 AM Ashish Soni <asoni.le...@gmail.com> wrote:

> I made some progress but now i am stuck at this point , Please help as
> looks like i am close to get it working
>
> I have everything running in docker container including mesos slave and
> master
>
> When i try to submit the pi example i get below error
> *Error: Cannot load main class from JAR file:/opt/spark/Example*
>
> Below is the command i use to submit as a docker container
>
> docker run -it --rm -e SPARK_MASTER="mesos://10.0.2.15:7077"  -e
> SPARK_IMAGE="spark_driver:latest" spark_driver:latest ./bin/spark-submit
> --deploy-mode cluster --name "PI Example" --class
> org.apache.spark.examples.SparkPi --driver-memory 512m --executor-memory
> 512m --executor-cores 1
> http://10.0.2.15/spark-examples-1.6.0-hadoop2.6.0.jar
>
>
> On Tue, Mar 1, 2016 at 2:59 PM, Timothy Chen <t...@mesosphere.io> wrote:
>
>> Can you go through the Mesos UI and look at the driver/executor log from
>> steer file and see what the problem is?
>>
>> Tim
>>
>> On Mar 1, 2016, at 8:05 AM, Ashish Soni <asoni.le...@gmail.com> wrote:
>>
>> Not sure what is the issue but i am getting below error  when i try to
>> run spark PI example
>>
>> Blacklisting Mesos slave value: "5345asdasdasdkas234234asdasdasdasd"
>>    due to too many failures; is Spark installed on it?
>>     WARN TaskSchedulerImpl: Initial job has not accepted any resources; 
>> check your cluster UI to ensure that workers are registered and have 
>> sufficient resources
>>
>>
>> On Mon, Feb 29, 2016 at 1:39 PM, Sathish Kumaran Vairavelu <
>> vsathishkuma...@gmail.com> wrote:
>>
>>> May be the Mesos executor couldn't find spark image or the constraints
>>> are not satisfied. Check your Mesos UI if you see Spark application in the
>>> Frameworks tab
>>>
>>> On Mon, Feb 29, 2016 at 12:23 PM Ashish Soni <asoni.le...@gmail.com>
>>> wrote:
>>>
>>>> What is the Best practice , I have everything running as docker
>>>> container in single host ( mesos and marathon also as docker container )
>>>>  and everything comes up fine but when i try to launch the spark shell i
>>>> get below error
>>>>
>>>>
>>>> SQL context available as sqlContext.
>>>>
>>>> scala> val data = sc.parallelize(1 to 100)
>>>> data: org.apache.spark.rdd.RDD[Int] = ParallelCollectionRDD[0] at
>>>> parallelize at <console>:27
>>>>
>>>> scala> data.count
>>>> [Stage 0:>                                                          (0
>>>> + 0) / 2]16/02/29 18:21:12 WARN TaskSchedulerImpl: Initial job has not
>>>> accepted any resources; check your cluster UI to ensure that workers are
>>>> registered and have sufficient resources
>>>> 16/02/29 18:21:27 WARN TaskSchedulerImpl: Initial job has not accepted
>>>> any resources; check your cluster UI to ensure that workers are registered
>>>> and have sufficient resources
>>>>
>>>>
>>>>
>>>> On Mon, Feb 29, 2016 at 12:04 PM, Tim Chen <t...@mesosphere.io> wrote:
>>>>
>>>>> No you don't have to run Mesos in docker containers to run Spark in
>>>>> docker containers.
>>>>>
>>>>> Once you have Mesos cluster running you can then specfiy the Spark
>>>>> configurations in your Spark job (i.e: 
>>>>> spark.mesos.executor.docker.image=mesosphere/spark:1.6)
>>>>> and Mesos will automatically launch docker containers for you.
>>>>>
>>>>> Tim
>>>>>
>>>>> On Mon, Feb 29, 2016 at 7:36 AM, Ashish Soni <asoni.le...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> Yes i read that and not much details here.
>>>>>>
>>>>>> Is it true that we need to have spark installed on each mesos docker
>>>>>> container ( master and slave ) ...
>>>>>>
>>>>>> Ashish
>>>>>>
>>>>>> On Fri, Feb 26, 2016 at 2:14 PM, Tim Chen <t...@mesosphere.io> wrote:
>>>>>>
>>>>>>> https://spark.apache.org/docs/latest/running-on-mesos.html should
>>>>>>> be the best source, what problems were you running into?
>>>>>>>
>>>>>>> Tim
>>>>>>>
>>>>>>> On Fri, Feb 26, 2016 at 11:06 AM, Yin Yang <yy201...@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Have you read this ?
>>>>>>>> https://spark.apache.org/docs/latest/running-on-mesos.html
>>>>>>>>
>>>>>>>> On Fri, Feb 26, 2016 at 11:03 AM, Ashish Soni <
>>>>>>>> asoni.le...@gmail.com> wrote:
>>>>>>>>
>>>>>>>>> Hi All ,
>>>>>>>>>
>>>>>>>>> Is there any proper documentation as how to run spark on mesos , I
>>>>>>>>> am trying from the last few days and not able to make it work.
>>>>>>>>>
>>>>>>>>> Please help
>>>>>>>>>
>>>>>>>>> Ashish
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>
>

Reply via email to