Tuple [Lscala.Tuple2;@65e4cb84
On Sun, Jul 31, 2016 at 1:00 AM, Jacek Laskowski <ja...@japila.pl> wrote: > Hi, > > What's the result type of sliding(2,1)? > > Pozdrawiam, > Jacek Laskowski > ---- > https://medium.com/@jaceklaskowski/ > Mastering Apache Spark 2.0 http://bit.ly/mastering-apache-spark > Follow me at https://twitter.com/jaceklaskowski > > > On Sun, Jul 31, 2016 at 9:23 AM, sri hari kali charan Tummala > <kali.tumm...@gmail.com> wrote: > > tried this no luck, wht is non-empty iterator here ? > > > > OP:- > > (-987,non-empty iterator) > > (-987,non-empty iterator) > > (-987,non-empty iterator) > > (-987,non-empty iterator) > > (-987,non-empty iterator) > > > > > > sc.textFile(file).keyBy(x => x.split("\\~") (0)) > > .map(x => x._2.split("\\~")) > > .map(x => (x(0),x(2))) > > .map { case (key,value) => > (key,value.toArray.toSeq.sliding(2,1).map(x > > => x.sum/x.size))}.foreach(println) > > > > > > On Sun, Jul 31, 2016 at 12:03 AM, sri hari kali charan Tummala > > <kali.tumm...@gmail.com> wrote: > >> > >> Hi All, > >> > >> I managed to write using sliding function but can it get key as well in > my > >> output ? > >> > >> sc.textFile(file).keyBy(x => x.split("\\~") (0)) > >> .map(x => x._2.split("\\~")) > >> .map(x => (x(2).toDouble)).toArray().sliding(2,1).map(x => > >> (x,x.size)).foreach(println) > >> > >> > >> at the moment my output:- > >> > >> 75.0 > >> -25.0 > >> 50.0 > >> -50.0 > >> -100.0 > >> > >> I want with key how to get moving average output based on key ? > >> > >> > >> 987,75.0 > >> 987,-25 > >> 987,50.0 > >> > >> Thanks > >> Sri > >> > >> > >> > >> > >> > >> > >> On Sat, Jul 30, 2016 at 11:40 AM, sri hari kali charan Tummala > >> <kali.tumm...@gmail.com> wrote: > >>> > >>> for knowledge just wondering how to write it up in scala or spark RDD. > >>> > >>> Thanks > >>> Sri > >>> > >>> On Sat, Jul 30, 2016 at 11:24 AM, Jacek Laskowski <ja...@japila.pl> > >>> wrote: > >>>> > >>>> Why? > >>>> > >>>> Pozdrawiam, > >>>> Jacek Laskowski > >>>> ---- > >>>> https://medium.com/@jaceklaskowski/ > >>>> Mastering Apache Spark 2.0 http://bit.ly/mastering-apache-spark > >>>> Follow me at https://twitter.com/jaceklaskowski > >>>> > >>>> > >>>> On Sat, Jul 30, 2016 at 4:42 AM, kali.tumm...@gmail.com > >>>> <kali.tumm...@gmail.com> wrote: > >>>> > Hi All, > >>>> > > >>>> > I managed to write business requirement in spark-sql and hive I am > >>>> > still > >>>> > learning scala how this below sql be written using spark RDD not > spark > >>>> > data > >>>> > frames. > >>>> > > >>>> > SELECT DATE,balance, > >>>> > SUM(balance) OVER (ORDER BY DATE ROWS BETWEEN UNBOUNDED PRECEDING > AND > >>>> > CURRENT ROW) daily_balance > >>>> > FROM table > >>>> > > >>>> > > >>>> > > >>>> > > >>>> > > >>>> > -- > >>>> > View this message in context: > >>>> > > http://apache-spark-user-list.1001560.n3.nabble.com/sql-to-spark-scala-rdd-tp27433.html > >>>> > Sent from the Apache Spark User List mailing list archive at > >>>> > Nabble.com. > >>>> > > >>>> > > --------------------------------------------------------------------- > >>>> > To unsubscribe e-mail: user-unsubscr...@spark.apache.org > >>>> > > >>> > >>> > >>> > >>> > >>> -- > >>> Thanks & Regards > >>> Sri Tummala > >>> > >> > >> > >> > >> -- > >> Thanks & Regards > >> Sri Tummala > >> > > > > > > > > -- > > Thanks & Regards > > Sri Tummala > > > -- Thanks & Regards Sri Tummala