Hi, What's the result type of sliding(2,1)?
Pozdrawiam, Jacek Laskowski ---- https://medium.com/@jaceklaskowski/ Mastering Apache Spark 2.0 http://bit.ly/mastering-apache-spark Follow me at https://twitter.com/jaceklaskowski On Sun, Jul 31, 2016 at 9:23 AM, sri hari kali charan Tummala <kali.tumm...@gmail.com> wrote: > tried this no luck, wht is non-empty iterator here ? > > OP:- > (-987,non-empty iterator) > (-987,non-empty iterator) > (-987,non-empty iterator) > (-987,non-empty iterator) > (-987,non-empty iterator) > > > sc.textFile(file).keyBy(x => x.split("\\~") (0)) > .map(x => x._2.split("\\~")) > .map(x => (x(0),x(2))) > .map { case (key,value) => (key,value.toArray.toSeq.sliding(2,1).map(x > => x.sum/x.size))}.foreach(println) > > > On Sun, Jul 31, 2016 at 12:03 AM, sri hari kali charan Tummala > <kali.tumm...@gmail.com> wrote: >> >> Hi All, >> >> I managed to write using sliding function but can it get key as well in my >> output ? >> >> sc.textFile(file).keyBy(x => x.split("\\~") (0)) >> .map(x => x._2.split("\\~")) >> .map(x => (x(2).toDouble)).toArray().sliding(2,1).map(x => >> (x,x.size)).foreach(println) >> >> >> at the moment my output:- >> >> 75.0 >> -25.0 >> 50.0 >> -50.0 >> -100.0 >> >> I want with key how to get moving average output based on key ? >> >> >> 987,75.0 >> 987,-25 >> 987,50.0 >> >> Thanks >> Sri >> >> >> >> >> >> >> On Sat, Jul 30, 2016 at 11:40 AM, sri hari kali charan Tummala >> <kali.tumm...@gmail.com> wrote: >>> >>> for knowledge just wondering how to write it up in scala or spark RDD. >>> >>> Thanks >>> Sri >>> >>> On Sat, Jul 30, 2016 at 11:24 AM, Jacek Laskowski <ja...@japila.pl> >>> wrote: >>>> >>>> Why? >>>> >>>> Pozdrawiam, >>>> Jacek Laskowski >>>> ---- >>>> https://medium.com/@jaceklaskowski/ >>>> Mastering Apache Spark 2.0 http://bit.ly/mastering-apache-spark >>>> Follow me at https://twitter.com/jaceklaskowski >>>> >>>> >>>> On Sat, Jul 30, 2016 at 4:42 AM, kali.tumm...@gmail.com >>>> <kali.tumm...@gmail.com> wrote: >>>> > Hi All, >>>> > >>>> > I managed to write business requirement in spark-sql and hive I am >>>> > still >>>> > learning scala how this below sql be written using spark RDD not spark >>>> > data >>>> > frames. >>>> > >>>> > SELECT DATE,balance, >>>> > SUM(balance) OVER (ORDER BY DATE ROWS BETWEEN UNBOUNDED PRECEDING AND >>>> > CURRENT ROW) daily_balance >>>> > FROM table >>>> > >>>> > >>>> > >>>> > >>>> > >>>> > -- >>>> > View this message in context: >>>> > http://apache-spark-user-list.1001560.n3.nabble.com/sql-to-spark-scala-rdd-tp27433.html >>>> > Sent from the Apache Spark User List mailing list archive at >>>> > Nabble.com. >>>> > >>>> > --------------------------------------------------------------------- >>>> > To unsubscribe e-mail: user-unsubscr...@spark.apache.org >>>> > >>> >>> >>> >>> >>> -- >>> Thanks & Regards >>> Sri Tummala >>> >> >> >> >> -- >> Thanks & Regards >> Sri Tummala >> > > > > -- > Thanks & Regards > Sri Tummala > --------------------------------------------------------------------- To unsubscribe e-mail: user-unsubscr...@spark.apache.org