Hi,

Impossible - see
http://www.scala-lang.org/api/current/index.html#scala.collection.Seq@sliding(size:Int,step:Int):Iterator[Repr].

I tried to show you why you ended up with "non-empty iterator" after
println. You should really start with
http://www.scala-lang.org/documentation/

Pozdrawiam,
Jacek Laskowski
----
https://medium.com/@jaceklaskowski/
Mastering Apache Spark 2.0 http://bit.ly/mastering-apache-spark
Follow me at https://twitter.com/jaceklaskowski


On Sun, Jul 31, 2016 at 8:49 PM, sri hari kali charan Tummala
<kali.tumm...@gmail.com> wrote:
> Tuple
>
> [Lscala.Tuple2;@65e4cb84
>
> On Sun, Jul 31, 2016 at 1:00 AM, Jacek Laskowski <ja...@japila.pl> wrote:
>>
>> Hi,
>>
>> What's the result type of sliding(2,1)?
>>
>> Pozdrawiam,
>> Jacek Laskowski
>> ----
>> https://medium.com/@jaceklaskowski/
>> Mastering Apache Spark 2.0 http://bit.ly/mastering-apache-spark
>> Follow me at https://twitter.com/jaceklaskowski
>>
>>
>> On Sun, Jul 31, 2016 at 9:23 AM, sri hari kali charan Tummala
>> <kali.tumm...@gmail.com> wrote:
>> > tried this no luck, wht is non-empty iterator here ?
>> >
>> > OP:-
>> > (-987,non-empty iterator)
>> > (-987,non-empty iterator)
>> > (-987,non-empty iterator)
>> > (-987,non-empty iterator)
>> > (-987,non-empty iterator)
>> >
>> >
>> > sc.textFile(file).keyBy(x => x.split("\\~") (0))
>> >   .map(x => x._2.split("\\~"))
>> >   .map(x => (x(0),x(2)))
>> >     .map { case (key,value) =>
>> > (key,value.toArray.toSeq.sliding(2,1).map(x
>> > => x.sum/x.size))}.foreach(println)
>> >
>> >
>> > On Sun, Jul 31, 2016 at 12:03 AM, sri hari kali charan Tummala
>> > <kali.tumm...@gmail.com> wrote:
>> >>
>> >> Hi All,
>> >>
>> >> I managed to write using sliding function but can it get key as well in
>> >> my
>> >> output ?
>> >>
>> >> sc.textFile(file).keyBy(x => x.split("\\~") (0))
>> >>       .map(x => x._2.split("\\~"))
>> >>       .map(x => (x(2).toDouble)).toArray().sliding(2,1).map(x =>
>> >> (x,x.size)).foreach(println)
>> >>
>> >>
>> >> at the moment my output:-
>> >>
>> >> 75.0
>> >> -25.0
>> >> 50.0
>> >> -50.0
>> >> -100.0
>> >>
>> >> I want with key how to get moving average output based on key ?
>> >>
>> >>
>> >> 987,75.0
>> >> 987,-25
>> >> 987,50.0
>> >>
>> >> Thanks
>> >> Sri
>> >>
>> >>
>> >>
>> >>
>> >>
>> >>
>> >> On Sat, Jul 30, 2016 at 11:40 AM, sri hari kali charan Tummala
>> >> <kali.tumm...@gmail.com> wrote:
>> >>>
>> >>> for knowledge just wondering how to write it up in scala or spark RDD.
>> >>>
>> >>> Thanks
>> >>> Sri
>> >>>
>> >>> On Sat, Jul 30, 2016 at 11:24 AM, Jacek Laskowski <ja...@japila.pl>
>> >>> wrote:
>> >>>>
>> >>>> Why?
>> >>>>
>> >>>> Pozdrawiam,
>> >>>> Jacek Laskowski
>> >>>> ----
>> >>>> https://medium.com/@jaceklaskowski/
>> >>>> Mastering Apache Spark 2.0 http://bit.ly/mastering-apache-spark
>> >>>> Follow me at https://twitter.com/jaceklaskowski
>> >>>>
>> >>>>
>> >>>> On Sat, Jul 30, 2016 at 4:42 AM, kali.tumm...@gmail.com
>> >>>> <kali.tumm...@gmail.com> wrote:
>> >>>> > Hi All,
>> >>>> >
>> >>>> > I managed to write business requirement in spark-sql and hive I am
>> >>>> > still
>> >>>> > learning scala how this below sql be written using spark RDD not
>> >>>> > spark
>> >>>> > data
>> >>>> > frames.
>> >>>> >
>> >>>> > SELECT DATE,balance,
>> >>>> > SUM(balance) OVER (ORDER BY DATE ROWS BETWEEN UNBOUNDED PRECEDING
>> >>>> > AND
>> >>>> > CURRENT ROW) daily_balance
>> >>>> > FROM  table
>> >>>> >
>> >>>> >
>> >>>> >
>> >>>> >
>> >>>> >
>> >>>> > --
>> >>>> > View this message in context:
>> >>>> >
>> >>>> > http://apache-spark-user-list.1001560.n3.nabble.com/sql-to-spark-scala-rdd-tp27433.html
>> >>>> > Sent from the Apache Spark User List mailing list archive at
>> >>>> > Nabble.com.
>> >>>> >
>> >>>> >
>> >>>> > ---------------------------------------------------------------------
>> >>>> > To unsubscribe e-mail: user-unsubscr...@spark.apache.org
>> >>>> >
>> >>>
>> >>>
>> >>>
>> >>>
>> >>> --
>> >>> Thanks & Regards
>> >>> Sri Tummala
>> >>>
>> >>
>> >>
>> >>
>> >> --
>> >> Thanks & Regards
>> >> Sri Tummala
>> >>
>> >
>> >
>> >
>> > --
>> > Thanks & Regards
>> > Sri Tummala
>> >
>
>
>
>
> --
> Thanks & Regards
> Sri Tummala
>

---------------------------------------------------------------------
To unsubscribe e-mail: user-unsubscr...@spark.apache.org

Reply via email to