What I would like to do it's to count the number of elements and if
it's greater than a number, I have to iterate all them and store them
in mysql or another system. So, I need to count them and preserve the
values because saving in other system.

I know about this map(line => ("key", line)), it was just a test, I
want to change "key" for a value which comes from a RE.

2014-12-17 17:28 GMT+01:00 Gerard Maas <gerard.m...@gmail.com>:
>
> You can create a DStream that contains the count, transforming the grouped
> windowed RDD, like this:
> val errorCount = grouping.map{case (k,v) => v.size }
>
> If you need to preserve the key:
> val errorCount = grouping.map{case (k,v) => (k,v.size) }
>
> or you if you don't care about the content of the values, you could count
> directly, instead of grouping first:
>
> val errorCount = mapErrorLines.countByWindow(Seconds(8), Seconds(4))
>
> Not sure why you're using map(line => ("key", line)) as there only seem to
> be one key. If that's not required, we can simplify one more step:
>
> val errorCount = errorLines.countByWindow(Seconds(8), Seconds(4))
>
>
> The question is: what do you want to do with that count afterwards?
>
> -kr, Gerard.
>
>
> On Wed, Dec 17, 2014 at 5:11 PM, Guillermo Ortiz <konstt2...@gmail.com>
> wrote:
>>
>> I'm a newbie with Spark,,, a simple question
>>
>> val errorLines = lines.filter(_.contains("h"))
>> val mapErrorLines = errorLines.map(line => ("key", line))
>> val grouping = errorLinesValue.groupByKeyAndWindow(Seconds(8), Seconds(4))
>>
>> I get something like:
>>
>> 604: -------------------------------------------
>> 605: Time: 1418832180000 ms
>> 606: -------------------------------------------
>> 607: (key,ArrayBuffer(h2, h3, h4))
>>
>> Now, I would like to get that ArrayBuffer and count the number of
>> elements,,
>> How could I get that arrayBuffer??? something like:
>> val values = grouping.getValue()... How could I do this in Spark with
>> Scala?
>>
>> ---------------------------------------------------------------------
>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>> For additional commands, e-mail: user-h...@spark.apache.org
>>
>

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to