+1

On Wed, Jun 22, 2016 at 1:07 PM, Kousuke Saruta <saru...@oss.nttdata.co.jp>
wrote:

> +1 (non-binding)
>
> On 2016/06/23 4:53, Reynold Xin wrote:
>
> +1 myself
>
>
> On Wed, Jun 22, 2016 at 12:19 PM, Sean McNamara <
> sean.mcnam...@webtrends.com> wrote:
>
>> +1
>>
>> On Jun 22, 2016, at 1:14 PM, Michael Armbrust <mich...@databricks.com>
>> wrote:
>>
>> +1
>>
>> On Wed, Jun 22, 2016 at 11:33 AM, Jonathan Kelly <jonathaka...@gmail.com>
>> wrote:
>>
>>> +1
>>>
>>> On Wed, Jun 22, 2016 at 10:41 AM Tim Hunter < <timhun...@databricks.com>
>>> timhun...@databricks.com> wrote:
>>>
>>>> +1 This release passes all tests on the graphframes and tensorframes
>>>> packages.
>>>>
>>>> On Wed, Jun 22, 2016 at 7:19 AM, Cody Koeninger < <c...@koeninger.org>
>>>> c...@koeninger.org> wrote:
>>>>
>>>>> If we're considering backporting changes for the 0.8 kafka
>>>>> integration, I am sure there are people who would like to get
>>>>>
>>>>> https://issues.apache.org/jira/browse/SPARK-10963
>>>>>
>>>>> into 1.6.x as well
>>>>>
>>>>> On Wed, Jun 22, 2016 at 7:41 AM, Sean Owen < <so...@cloudera.com>
>>>>> so...@cloudera.com> wrote:
>>>>> > Good call, probably worth back-porting, I'll try to do that. I don't
>>>>> > think it blocks a release, but would be good to get into a next RC if
>>>>> > any.
>>>>> >
>>>>> > On Wed, Jun 22, 2016 at 11:38 AM, Pete Robbins <
>>>>> <robbin...@gmail.com>robbin...@gmail.com> wrote:
>>>>> >> This has failed on our 1.6 stream builds regularly.
>>>>> >> ( <https://issues.apache.org/jira/browse/SPARK-6005>
>>>>> https://issues.apache.org/jira/browse/SPARK-6005) looks fixed in 2.0?
>>>>> >>
>>>>> >> On Wed, 22 Jun 2016 at 11:15 Sean Owen < <so...@cloudera.com>
>>>>> so...@cloudera.com> wrote:
>>>>> >>>
>>>>> >>> Oops, one more in the "does anybody else see this" department:
>>>>> >>>
>>>>> >>> - offset recovery *** FAILED ***
>>>>> >>>   recoveredOffsetRanges.forall(((or:
>>>>> (org.apache.spark.streaming.Time,
>>>>> >>> Array[org.apache.spark.streaming.kafka.OffsetRange])) =>
>>>>> >>>
>>>>> >>>
>>>>> earlierOffsetRangesAsSets.contains(scala.Tuple2.apply[org.apache.spark.streaming.Time,
>>>>> >>>
>>>>> >>>
>>>>> scala.collection.immutable.Set[org.apache.spark.streaming.kafka.OffsetRange]](or._1,
>>>>> >>>
>>>>> >>>
>>>>> scala.this.Predef.refArrayOps[org.apache.spark.streaming.kafka.OffsetRange](or._2).toSet[org.apache.spark.streaming.kafka.OffsetRange]))))
>>>>> >>> was false Recovered ranges are not the same as the ones generated
>>>>> >>> (DirectKafkaStreamSuite.scala:301)
>>>>> >>>
>>>>> >>> This actually fails consistently for me too in the Kafka
>>>>> integration
>>>>> >>> code. Not timezone related, I think.
>>>>> >
>>>>> > ---------------------------------------------------------------------
>>>>> > To unsubscribe, e-mail: <dev-unsubscr...@spark.apache.org>
>>>>> dev-unsubscr...@spark.apache.org
>>>>> > For additional commands, e-mail: <dev-h...@spark.apache.org>
>>>>> dev-h...@spark.apache.org
>>>>> >
>>>>>
>>>>> ---------------------------------------------------------------------
>>>>> To unsubscribe, e-mail: <dev-unsubscr...@spark.apache.org>
>>>>> dev-unsubscr...@spark.apache.org
>>>>> For additional commands, e-mail: <dev-h...@spark.apache.org>
>>>>> dev-h...@spark.apache.org
>>>>>
>>>>>
>>>>
>>
>>
>
>


-- 
Sameer Agarwal
Software Engineer | Databricks Inc.
http://cs.berkeley.edu/~sameerag

Reply via email to