I believe TD just forgot to set the fix version on the JIRA. There is
a fix for this in 1.3:

https://github.com/apache/spark/commit/03e263f5b527cf574f4ffcd5cd886f7723e3756e

- Patrick

On Mon, Apr 6, 2015 at 2:31 PM, Mark Hamstra <m...@clearstorydata.com> wrote:
> Is that correct, or is the JIRA just out of sync, since TD's PR was merged?
> https://github.com/apache/spark/pull/5008
>
> On Mon, Apr 6, 2015 at 11:10 AM, Hari Shreedharan
> <hshreedha...@cloudera.com> wrote:
>>
>> It does not look like https://issues.apache.org/jira/browse/SPARK-6222
>> made it. It was targeted towards this release.
>>
>>
>>
>>
>> Thanks, Hari
>>
>> On Mon, Apr 6, 2015 at 11:04 AM, York, Brennon
>> <brennon.y...@capitalone.com> wrote:
>>
>> > +1 (non-binding)
>> > Tested GraphX, build infrastructure, & core test suite on OSX 10.9 w/
>> > Java
>> > 1.7/1.8
>> > On 4/6/15, 5:21 AM, "Sean Owen" <so...@cloudera.com> wrote:
>> >>SPARK-6673 is not, in the end, relevant for 1.3.x I believe; we just
>> >>resolved it for 1.4 anyway. False alarm there.
>> >>
>> >>I back-ported SPARK-6205 into the 1.3 branch for next time. We'll pick
>> >>it up if there's another RC, but by itself is not something that needs
>> >>a new RC. (I will give the same treatment to branch 1.2 if needed in
>> >>light of the 1.2.2 release.)
>> >>
>> >>I applied the simple change in SPARK-6205 in order to continue
>> >>executing tests and all was well. I still see a few failures in Hive
>> >>tests:
>> >>
>> >>- show_create_table_serde *** FAILED ***
>> >>- show_tblproperties *** FAILED ***
>> >>- udf_std *** FAILED ***
>> >>- udf_stddev *** FAILED ***
>> >>
>> >>with ...
>> >>
>> >>mvn -Phadoop-2.4 -Pyarn -Phive -Phive-0.13.1 -Dhadoop.version=2.6.0
>> >>-DskipTests clean package; mvn -Phadoop-2.4 -Pyarn -Phive
>> >>-Phive-0.13.1 -Dhadoop.version=2.6.0 test
>> >>
>> >>... but these are not regressions from 1.3.0.
>> >>
>> >>+1 from me at this point on the current artifacts.
>> >>
>> >>On Sun, Apr 5, 2015 at 9:24 AM, Sean Owen <so...@cloudera.com> wrote:
>> >>> Signatures and hashes are good.
>> >>> LICENSE, NOTICE still check out.
>> >>> Compiles for a Hadoop 2.6 + YARN + Hive profile.
>> >>>
>> >>> I still see the UISeleniumSuite test failure observed in 1.3.0, which
>> >>> is minor and already fixed. I don't know why I didn't back-port it:
>> >>> https://issues.apache.org/jira/browse/SPARK-6205
>> >>>
>> >>> If we roll another, let's get this easy fix in, but it is only an
>> >>> issue with tests.
>> >>>
>> >>>
>> >>> On JIRA, I checked open issues with Fix Version = 1.3.0 or 1.3.1 and
>> >>> all look legitimate (e.g. reopened or in progress)
>> >>>
>> >>>
>> >>> There is 1 open Blocker for 1.3.1 per Andrew:
>> >>> https://issues.apache.org/jira/browse/SPARK-6673 spark-shell.cmd can't
>> >>> start even when spark was built in Windows
>> >>>
>> >>> I believe this can be resolved quickly but as a matter of hygiene
>> >>> should be fixed or demoted before release.
>> >>>
>> >>>
>> >>> FYI there are 16 Critical issues marked for 1.3.0 / 1.3.1; worth
>> >>> examining before release to see how critical they are:
>> >>>
>> >>> SPARK-6701,Flaky test: o.a.s.deploy.yarn.YarnClusterSuite Python
>> >>> application,,Open,4/3/15
>> >>> SPARK-6484,"Ganglia metrics xml reporter doesn't escape
>> >>> correctly",Josh Rosen,Open,3/24/15
>> >>> SPARK-6270,Standalone Master hangs when streaming job
>> >>>completes,,Open,3/11/15
>> >>> SPARK-6209,ExecutorClassLoader can leak connections after failing to
>> >>> load classes from the REPL class server,Josh Rosen,In Progress,4/2/15
>> >>> SPARK-5113,Audit and document use of hostnames and IP addresses in
>> >>> Spark,,Open,3/24/15
>> >>> SPARK-5098,Number of running tasks become negative after tasks
>> >>> lost,,Open,1/14/15
>> >>> SPARK-4925,Publish Spark SQL hive-thriftserver maven artifact,Patrick
>> >>> Wendell,Reopened,3/23/15
>> >>> SPARK-4922,Support dynamic allocation for coarse-grained
>> >>>Mesos,,Open,3/31/15
>> >>> SPARK-4888,"Spark EC2 doesn't mount local disks for i2.8xlarge
>> >>> instances",,Open,1/27/15
>> >>> SPARK-4879,Missing output partitions after job completes with
>> >>> speculative execution,Josh Rosen,Open,3/5/15
>> >>> SPARK-4751,Support dynamic allocation for standalone mode,Andrew
>> >>> Or,Open,12/22/14
>> >>> SPARK-4454,Race condition in DAGScheduler,Josh Rosen,Reopened,2/18/15
>> >>> SPARK-4452,Shuffle data structures can starve others on the same
>> >>> thread for memory,Tianshuo Deng,Open,1/24/15
>> >>> SPARK-4352,Incorporate locality preferences in dynamic allocation
>> >>> requests,,Open,1/26/15
>> >>> SPARK-4227,Document external shuffle service,,Open,3/23/15
>> >>> SPARK-3650,Triangle Count handles reverse edges
>> >>>incorrectly,,Open,2/23/15
>> >>>
>> >>> On Sun, Apr 5, 2015 at 1:09 AM, Patrick Wendell <pwend...@gmail.com>
>> >>>wrote:
>> >>>> Please vote on releasing the following candidate as Apache Spark
>> >>>>version 1.3.1!
>> >>>>
>> >>>> The tag to be voted on is v1.3.1-rc1 (commit 0dcb5d9f):
>> >>>>
>>
>> >>>> >>>>https://git-wip-us.apache.org/repos/asf?p=spark.git;a=commit;h=0dcb5d9f3
>> >>>>1b713ed90bcec63ebc4e530cbb69851
>> >>>>
>> >>>> The list of fixes present in this release can be found at:
>> >>>> http://bit.ly/1C2nVPY
>> >>>>
>> >>>> The release files, including signatures, digests, etc. can be found
>> >>>> at:
>> >>>> http://people.apache.org/~pwendell/spark-1.3.1-rc1/
>> >>>>
>> >>>> Release artifacts are signed with the following key:
>> >>>> https://people.apache.org/keys/committer/pwendell.asc
>> >>>>
>> >>>> The staging repository for this release can be found at:
>> >>>>
>> >>>> https://repository.apache.org/content/repositories/orgapachespark-1080
>> >>>>
>> >>>> The documentation corresponding to this release can be found at:
>> >>>> http://people.apache.org/~pwendell/spark-1.3.1-rc1-docs/
>> >>>>
>> >>>> Please vote on releasing this package as Apache Spark 1.3.1!
>> >>>>
>> >>>> The vote is open until Wednesday, April 08, at 01:10 UTC and passes
>> >>>> if a majority of at least 3 +1 PMC votes are cast.
>> >>>>
>> >>>> [ ] +1 Release this package as Apache Spark 1.3.1
>> >>>> [ ] -1 Do not release this package because ...
>> >>>>
>> >>>> To learn more about Apache Spark, please see
>> >>>> http://spark.apache.org/
>> >>>>
>> >>>> - Patrick
>> >>>>
>> >>>> ---------------------------------------------------------------------
>> >>>> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
>> >>>> For additional commands, e-mail: dev-h...@spark.apache.org
>> >>>>
>> >>
>> >>---------------------------------------------------------------------
>> >>To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
>> >>For additional commands, e-mail: dev-h...@spark.apache.org
>> >>
>> > ________________________________________________________
>> > The information contained in this e-mail is confidential and/or
>> > proprietary to Capital One and/or its affiliates. The information
>> > transmitted herewith is intended only for use by the individual or entity 
>> > to
>> > which it is addressed.  If the reader of this message is not the intended
>> > recipient, you are hereby notified that any review, retransmission,
>> > dissemination, distribution, copying or other use of, or taking of any
>> > action in reliance upon this information is strictly prohibited. If you 
>> > have
>> > received this communication in error, please contact the sender and delete
>> > the material from your computer.
>> > ---------------------------------------------------------------------
>> > To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
>> > For additional commands, e-mail: dev-h...@spark.apache.org
>
>

---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
For additional commands, e-mail: dev-h...@spark.apache.org

Reply via email to