Hi,

Thanks for that input, I tried doing that but apparently thats not working
as well. I thought i am having problems with my spark installation so I ran
simple word count and that works, so I am not really sure what the problem
is now.

Is my translation of the scala code correct? I don't understand the syntax
of scala very well thus wrote my own implementation of streaming kmeans in
java, so i am hoping thats correct.

Thanks & Regards
Biplob Biswas

On Sun, Jun 19, 2016 at 3:23 AM, Akhil Das <ak...@hacked.work> wrote:

> SparkStreaming does not pick up old files by default, so you need to start
> your job with master=local[2] (It needs 2 or more working threads, 1 to
> read the files and the other to do your computation) and once the job start
> to run, place your input files in the input directories and you can see
> them being picked up by sparkstreaming.
>
> On Sun, Jun 19, 2016 at 12:37 AM, Biplob Biswas <revolutioni...@gmail.com>
> wrote:
>
>> Hi,
>>
>> I tried local[*] and local[2] and the result is the same. I don't really
>> understand the problem here.
>> How can I confirm that the files are read properly?
>>
>> Thanks & Regards
>> Biplob Biswas
>>
>> On Sat, Jun 18, 2016 at 5:59 PM, Akhil Das <ak...@hacked.work> wrote:
>>
>>> Looks like you need to set your master to local[2] or local[*]
>>>
>>> On Sat, Jun 18, 2016 at 4:54 PM, Biplob Biswas <revolutioni...@gmail.com
>>> > wrote:
>>>
>>>> Hi,
>>>>
>>>> I implemented the streamingKmeans example provided in the spark website
>>>> but
>>>> in Java.
>>>> The full implementation is here,
>>>>
>>>> http://pastebin.com/CJQfWNvk
>>>>
>>>> But i am not getting anything in the output except occasional timestamps
>>>> like one below:
>>>>
>>>> -------------------------------------------
>>>> Time: 1466176935000 ms
>>>> -------------------------------------------
>>>>
>>>> Also, i have 2 directories:
>>>> "D:\spark\streaming example\Data Sets\training"
>>>> "D:\spark\streaming example\Data Sets\test"
>>>>
>>>> and inside these directories i have 1 file each
>>>> "samplegpsdata_train.txt"
>>>> and "samplegpsdata_test.txt" with training data having 500 datapoints
>>>> and
>>>> test data with 60 datapoints.
>>>>
>>>> I am very new to the spark systems and any help is highly appreciated.
>>>>
>>>> Thank you so much
>>>> Biplob Biswas
>>>>
>>>>
>>>>
>>>> --
>>>> View this message in context:
>>>> http://apache-spark-user-list.1001560.n3.nabble.com/Running-JavaBased-Implementationof-StreamingKmeans-tp27192.html
>>>> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>>>>
>>>> ---------------------------------------------------------------------
>>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>>>> For additional commands, e-mail: user-h...@spark.apache.org
>>>>
>>>>
>>>
>>>
>>> --
>>> Cheers!
>>>
>>>
>>
>
>
> --
> Cheers!
>
>

Reply via email to