Hi,

What's the result type of sliding(2,1)?

Pozdrawiam,
Jacek Laskowski
----
https://medium.com/@jaceklaskowski/
Mastering Apache Spark 2.0 http://bit.ly/mastering-apache-spark
Follow me at https://twitter.com/jaceklaskowski


On Sun, Jul 31, 2016 at 9:23 AM, sri hari kali charan Tummala
<kali.tumm...@gmail.com> wrote:
> tried this no luck, wht is non-empty iterator here ?
>
> OP:-
> (-987,non-empty iterator)
> (-987,non-empty iterator)
> (-987,non-empty iterator)
> (-987,non-empty iterator)
> (-987,non-empty iterator)
>
>
> sc.textFile(file).keyBy(x => x.split("\\~") (0))
>   .map(x => x._2.split("\\~"))
>   .map(x => (x(0),x(2)))
>     .map { case (key,value) => (key,value.toArray.toSeq.sliding(2,1).map(x
> => x.sum/x.size))}.foreach(println)
>
>
> On Sun, Jul 31, 2016 at 12:03 AM, sri hari kali charan Tummala
> <kali.tumm...@gmail.com> wrote:
>>
>> Hi All,
>>
>> I managed to write using sliding function but can it get key as well in my
>> output ?
>>
>> sc.textFile(file).keyBy(x => x.split("\\~") (0))
>>       .map(x => x._2.split("\\~"))
>>       .map(x => (x(2).toDouble)).toArray().sliding(2,1).map(x =>
>> (x,x.size)).foreach(println)
>>
>>
>> at the moment my output:-
>>
>> 75.0
>> -25.0
>> 50.0
>> -50.0
>> -100.0
>>
>> I want with key how to get moving average output based on key ?
>>
>>
>> 987,75.0
>> 987,-25
>> 987,50.0
>>
>> Thanks
>> Sri
>>
>>
>>
>>
>>
>>
>> On Sat, Jul 30, 2016 at 11:40 AM, sri hari kali charan Tummala
>> <kali.tumm...@gmail.com> wrote:
>>>
>>> for knowledge just wondering how to write it up in scala or spark RDD.
>>>
>>> Thanks
>>> Sri
>>>
>>> On Sat, Jul 30, 2016 at 11:24 AM, Jacek Laskowski <ja...@japila.pl>
>>> wrote:
>>>>
>>>> Why?
>>>>
>>>> Pozdrawiam,
>>>> Jacek Laskowski
>>>> ----
>>>> https://medium.com/@jaceklaskowski/
>>>> Mastering Apache Spark 2.0 http://bit.ly/mastering-apache-spark
>>>> Follow me at https://twitter.com/jaceklaskowski
>>>>
>>>>
>>>> On Sat, Jul 30, 2016 at 4:42 AM, kali.tumm...@gmail.com
>>>> <kali.tumm...@gmail.com> wrote:
>>>> > Hi All,
>>>> >
>>>> > I managed to write business requirement in spark-sql and hive I am
>>>> > still
>>>> > learning scala how this below sql be written using spark RDD not spark
>>>> > data
>>>> > frames.
>>>> >
>>>> > SELECT DATE,balance,
>>>> > SUM(balance) OVER (ORDER BY DATE ROWS BETWEEN UNBOUNDED PRECEDING AND
>>>> > CURRENT ROW) daily_balance
>>>> > FROM  table
>>>> >
>>>> >
>>>> >
>>>> >
>>>> >
>>>> > --
>>>> > View this message in context:
>>>> > http://apache-spark-user-list.1001560.n3.nabble.com/sql-to-spark-scala-rdd-tp27433.html
>>>> > Sent from the Apache Spark User List mailing list archive at
>>>> > Nabble.com.
>>>> >
>>>> > ---------------------------------------------------------------------
>>>> > To unsubscribe e-mail: user-unsubscr...@spark.apache.org
>>>> >
>>>
>>>
>>>
>>>
>>> --
>>> Thanks & Regards
>>> Sri Tummala
>>>
>>
>>
>>
>> --
>> Thanks & Regards
>> Sri Tummala
>>
>
>
>
> --
> Thanks & Regards
> Sri Tummala
>

---------------------------------------------------------------------
To unsubscribe e-mail: user-unsubscr...@spark.apache.org

Reply via email to