Hi, Impossible - see http://www.scala-lang.org/api/current/index.html#scala.collection.Seq@sliding(size:Int,step:Int):Iterator[Repr].
I tried to show you why you ended up with "non-empty iterator" after println. You should really start with http://www.scala-lang.org/documentation/ Pozdrawiam, Jacek Laskowski ---- https://medium.com/@jaceklaskowski/ Mastering Apache Spark 2.0 http://bit.ly/mastering-apache-spark Follow me at https://twitter.com/jaceklaskowski On Sun, Jul 31, 2016 at 8:49 PM, sri hari kali charan Tummala <kali.tumm...@gmail.com> wrote: > Tuple > > [Lscala.Tuple2;@65e4cb84 > > On Sun, Jul 31, 2016 at 1:00 AM, Jacek Laskowski <ja...@japila.pl> wrote: >> >> Hi, >> >> What's the result type of sliding(2,1)? >> >> Pozdrawiam, >> Jacek Laskowski >> ---- >> https://medium.com/@jaceklaskowski/ >> Mastering Apache Spark 2.0 http://bit.ly/mastering-apache-spark >> Follow me at https://twitter.com/jaceklaskowski >> >> >> On Sun, Jul 31, 2016 at 9:23 AM, sri hari kali charan Tummala >> <kali.tumm...@gmail.com> wrote: >> > tried this no luck, wht is non-empty iterator here ? >> > >> > OP:- >> > (-987,non-empty iterator) >> > (-987,non-empty iterator) >> > (-987,non-empty iterator) >> > (-987,non-empty iterator) >> > (-987,non-empty iterator) >> > >> > >> > sc.textFile(file).keyBy(x => x.split("\\~") (0)) >> > .map(x => x._2.split("\\~")) >> > .map(x => (x(0),x(2))) >> > .map { case (key,value) => >> > (key,value.toArray.toSeq.sliding(2,1).map(x >> > => x.sum/x.size))}.foreach(println) >> > >> > >> > On Sun, Jul 31, 2016 at 12:03 AM, sri hari kali charan Tummala >> > <kali.tumm...@gmail.com> wrote: >> >> >> >> Hi All, >> >> >> >> I managed to write using sliding function but can it get key as well in >> >> my >> >> output ? >> >> >> >> sc.textFile(file).keyBy(x => x.split("\\~") (0)) >> >> .map(x => x._2.split("\\~")) >> >> .map(x => (x(2).toDouble)).toArray().sliding(2,1).map(x => >> >> (x,x.size)).foreach(println) >> >> >> >> >> >> at the moment my output:- >> >> >> >> 75.0 >> >> -25.0 >> >> 50.0 >> >> -50.0 >> >> -100.0 >> >> >> >> I want with key how to get moving average output based on key ? >> >> >> >> >> >> 987,75.0 >> >> 987,-25 >> >> 987,50.0 >> >> >> >> Thanks >> >> Sri >> >> >> >> >> >> >> >> >> >> >> >> >> >> On Sat, Jul 30, 2016 at 11:40 AM, sri hari kali charan Tummala >> >> <kali.tumm...@gmail.com> wrote: >> >>> >> >>> for knowledge just wondering how to write it up in scala or spark RDD. >> >>> >> >>> Thanks >> >>> Sri >> >>> >> >>> On Sat, Jul 30, 2016 at 11:24 AM, Jacek Laskowski <ja...@japila.pl> >> >>> wrote: >> >>>> >> >>>> Why? >> >>>> >> >>>> Pozdrawiam, >> >>>> Jacek Laskowski >> >>>> ---- >> >>>> https://medium.com/@jaceklaskowski/ >> >>>> Mastering Apache Spark 2.0 http://bit.ly/mastering-apache-spark >> >>>> Follow me at https://twitter.com/jaceklaskowski >> >>>> >> >>>> >> >>>> On Sat, Jul 30, 2016 at 4:42 AM, kali.tumm...@gmail.com >> >>>> <kali.tumm...@gmail.com> wrote: >> >>>> > Hi All, >> >>>> > >> >>>> > I managed to write business requirement in spark-sql and hive I am >> >>>> > still >> >>>> > learning scala how this below sql be written using spark RDD not >> >>>> > spark >> >>>> > data >> >>>> > frames. >> >>>> > >> >>>> > SELECT DATE,balance, >> >>>> > SUM(balance) OVER (ORDER BY DATE ROWS BETWEEN UNBOUNDED PRECEDING >> >>>> > AND >> >>>> > CURRENT ROW) daily_balance >> >>>> > FROM table >> >>>> > >> >>>> > >> >>>> > >> >>>> > >> >>>> > >> >>>> > -- >> >>>> > View this message in context: >> >>>> > >> >>>> > http://apache-spark-user-list.1001560.n3.nabble.com/sql-to-spark-scala-rdd-tp27433.html >> >>>> > Sent from the Apache Spark User List mailing list archive at >> >>>> > Nabble.com. >> >>>> > >> >>>> > >> >>>> > --------------------------------------------------------------------- >> >>>> > To unsubscribe e-mail: user-unsubscr...@spark.apache.org >> >>>> > >> >>> >> >>> >> >>> >> >>> >> >>> -- >> >>> Thanks & Regards >> >>> Sri Tummala >> >>> >> >> >> >> >> >> >> >> -- >> >> Thanks & Regards >> >> Sri Tummala >> >> >> > >> > >> > >> > -- >> > Thanks & Regards >> > Sri Tummala >> > > > > > > -- > Thanks & Regards > Sri Tummala > --------------------------------------------------------------------- To unsubscribe e-mail: user-unsubscr...@spark.apache.org