-1 : SPARK-16228 [SQL]  - "Percentile" needs explicit cast to double,
otherwise it throws an error. I can not move my existing 100500 quires to
2.0 transparently.

2016-06-24 11:52 GMT-07:00 Matt Cheah <mch...@palantir.com>:

> -1 because of SPARK-16181 which is a correctness regression from 1.6.
> Looks like the patch is ready though:
> https://github.com/apache/spark/pull/13884 – it would be ideal for this
> patch to make it into the release.
>
> -Matt Cheah
>
> From: Nick Pentreath <nick.pentre...@gmail.com>
> Date: Friday, June 24, 2016 at 4:37 AM
> To: "dev@spark.apache.org" <dev@spark.apache.org>
> Subject: Re: [VOTE] Release Apache Spark 2.0.0 (RC1)
>
> I'm getting the following when trying to run ./dev/run-tests (not
> happening on master) from the extracted source tar. Anyone else seeing
> this?
>
> error: Could not access 'fc0a1475ef'
> **********************************************************************
> File "./dev/run-tests.py", line 69, in
> __main__.identify_changed_files_from_git_commits
> Failed example:
>     [x.name
> <https://urldefense.proofpoint.com/v2/url?u=http-3A__x.name&d=DQMFaQ&c=izlc9mHr637UR4lpLEZLFFS3Vn2UXBrZ4tFb6oOnmz8&r=hzwIMNQ9E99EMYGuqHI0kXhVbvX3nU3OSDadUnJxjAs&m=Y3d-oJvw2gK_2KXYjXY8_yzfAosPOqqaV4wtMg6ZPwM&s=wx5Qjw-efxMVvKXnjUsSkkQcEF6zQHQLQaGtAK9pxIw&e=>
> for x in determine_modules_for_files(
> identify_changed_files_from_git_commits("fc0a1475ef",
> target_ref="5da21f07"))]
> Exception raised:
>     Traceback (most recent call last):
>       File "/Users/nick/miniconda2/lib/python2.7/doctest.py", line 1315,
> in __run
>         compileflags, 1) in test.globs
>       File "<doctest
> __main__.identify_changed_files_from_git_commits[0]>", line 1, in <module>
>         [x.name
> <https://urldefense.proofpoint.com/v2/url?u=http-3A__x.name&d=DQMFaQ&c=izlc9mHr637UR4lpLEZLFFS3Vn2UXBrZ4tFb6oOnmz8&r=hzwIMNQ9E99EMYGuqHI0kXhVbvX3nU3OSDadUnJxjAs&m=Y3d-oJvw2gK_2KXYjXY8_yzfAosPOqqaV4wtMg6ZPwM&s=wx5Qjw-efxMVvKXnjUsSkkQcEF6zQHQLQaGtAK9pxIw&e=>
> for x in determine_modules_for_files(
> identify_changed_files_from_git_commits("fc0a1475ef",
> target_ref="5da21f07"))]
>       File "./dev/run-tests.py", line 86, in
> identify_changed_files_from_git_commits
>         universal_newlines=True)
>       File "/Users/nick/miniconda2/lib/python2.7/subprocess.py", line 573,
> in check_output
>         raise CalledProcessError(retcode, cmd, output=output)
>     CalledProcessError: Command '['git', 'diff', '--name-only',
> 'fc0a1475ef', '5da21f07']' returned non-zero exit status 1
> error: Could not access '50a0496a43'
> **********************************************************************
> File "./dev/run-tests.py", line 71, in
> __main__.identify_changed_files_from_git_commits
> Failed example:
>     'root' in [x.name
> <https://urldefense.proofpoint.com/v2/url?u=http-3A__x.name&d=DQMFaQ&c=izlc9mHr637UR4lpLEZLFFS3Vn2UXBrZ4tFb6oOnmz8&r=hzwIMNQ9E99EMYGuqHI0kXhVbvX3nU3OSDadUnJxjAs&m=Y3d-oJvw2gK_2KXYjXY8_yzfAosPOqqaV4wtMg6ZPwM&s=wx5Qjw-efxMVvKXnjUsSkkQcEF6zQHQLQaGtAK9pxIw&e=>
> for x in determine_modules_for_files(
>  identify_changed_files_from_git_commits("50a0496a43",
> target_ref="6765ef9"))]
> Exception raised:
>     Traceback (most recent call last):
>       File "/Users/nick/miniconda2/lib/python2.7/doctest.py", line 1315,
> in __run
>         compileflags, 1) in test.globs
>       File "<doctest
> __main__.identify_changed_files_from_git_commits[1]>", line 1, in <module>
>         'root' in [x.name
> <https://urldefense.proofpoint.com/v2/url?u=http-3A__x.name&d=DQMFaQ&c=izlc9mHr637UR4lpLEZLFFS3Vn2UXBrZ4tFb6oOnmz8&r=hzwIMNQ9E99EMYGuqHI0kXhVbvX3nU3OSDadUnJxjAs&m=Y3d-oJvw2gK_2KXYjXY8_yzfAosPOqqaV4wtMg6ZPwM&s=wx5Qjw-efxMVvKXnjUsSkkQcEF6zQHQLQaGtAK9pxIw&e=>
> for x in determine_modules_for_files(
>  identify_changed_files_from_git_commits("50a0496a43",
> target_ref="6765ef9"))]
>       File "./dev/run-tests.py", line 86, in
> identify_changed_files_from_git_commits
>         universal_newlines=True)
>       File "/Users/nick/miniconda2/lib/python2.7/subprocess.py", line 573,
> in check_output
>         raise CalledProcessError(retcode, cmd, output=output)
>     CalledProcessError: Command '['git', 'diff', '--name-only',
> '50a0496a43', '6765ef9']' returned non-zero exit status 1
> **********************************************************************
> 1 items had failures:
>    2 of   2 in __main__.identify_changed_files_from_git_commits
> ***Test Failed*** 2 failures.
>
>
>
> On Fri, 24 Jun 2016 at 06:59 Yin Huai <yh...@databricks.com> wrote:
>
>> -1 because of https://issues.apache.org/jira/browse/SPARK-16121
>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__issues.apache.org_jira_browse_SPARK-2D16121&d=DQMFaQ&c=izlc9mHr637UR4lpLEZLFFS3Vn2UXBrZ4tFb6oOnmz8&r=hzwIMNQ9E99EMYGuqHI0kXhVbvX3nU3OSDadUnJxjAs&m=Y3d-oJvw2gK_2KXYjXY8_yzfAosPOqqaV4wtMg6ZPwM&s=9200NP4SpeJSUNrSrlWWEC7vFvjWSyCHnx5LD7Sj9u4&e=>.
>>
>>
>> This jira was resolved after 2.0.0-RC1 was cut. Without the fix, Spark
>> SQL effectively only uses the driver to list files when loading datasets
>> and the driver-side file listing is very slow for datasets having many
>> files and partitions. Since this bug causes a serious performance
>> regression, I am giving -1.
>>
>> On Thu, Jun 23, 2016 at 1:25 AM, Pete Robbins <robbin...@gmail.com>
>> wrote:
>>
>>> I'm also seeing some of these same failures:
>>>
>>> - spilling with compression *** FAILED ***
>>> I have seen this occassionaly
>>>
>>> - to UTC timestamp *** FAILED ***
>>> This was fixed yesterday in branch-2.0 (
>>> https://issues.apache.org/jira/browse/SPARK-16078
>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__issues.apache.org_jira_browse_SPARK-2D16078&d=DQMFaQ&c=izlc9mHr637UR4lpLEZLFFS3Vn2UXBrZ4tFb6oOnmz8&r=hzwIMNQ9E99EMYGuqHI0kXhVbvX3nU3OSDadUnJxjAs&m=Y3d-oJvw2gK_2KXYjXY8_yzfAosPOqqaV4wtMg6ZPwM&s=SuVdXUNGdAhYgtA2fMLe5vZ2PFrPOaeO3i3cbhYU4tc&e=>)
>>>
>>>
>>> - offset recovery *** FAILED ***
>>> Haven't seen this for a while and thought the flaky test was fixed but
>>> it popped up again in one of our builds.
>>>
>>> StateStoreSuite:
>>> - maintenance *** FAILED ***
>>> Just seen this has been failing for last 2 days on one build machine
>>> (linux amd64)
>>>
>>>
>>> On 23 June 2016 at 08:51, Sean Owen <so...@cloudera.com> wrote:
>>>
>>>> First pass of feedback on the RC: all the sigs, hashes, etc are fine.
>>>> Licensing is up to date to the best of my knowledge.
>>>>
>>>> I'm hitting test failures, some of which may be spurious. Just putting
>>>> them out there to see if they ring bells. This is Java 8 on Ubuntu 16.
>>>>
>>>>
>>>> - spilling with compression *** FAILED ***
>>>>   java.lang.Exception: Test failed with compression using codec
>>>> org.apache.spark.io.SnappyCompressionCodec:
>>>> assertion failed: expected cogroup to spill, but did not
>>>>   at scala.Predef$.assert(Predef.scala:170)
>>>>   at org.apache.spark.TestUtils$.assertSpilled(TestUtils.scala:170)
>>>>   at org.apache.spark.util.collection.ExternalAppendOnlyMapSuite.org
>>>> <https://urldefense.proofpoint.com/v2/url?u=http-3A__org.apache.spark.util.collection.ExternalAppendOnlyMapSuite.org&d=DQMFaQ&c=izlc9mHr637UR4lpLEZLFFS3Vn2UXBrZ4tFb6oOnmz8&r=hzwIMNQ9E99EMYGuqHI0kXhVbvX3nU3OSDadUnJxjAs&m=Y3d-oJvw2gK_2KXYjXY8_yzfAosPOqqaV4wtMg6ZPwM&s=goarAptcJYfLg44f7BAwhbipqJlRFKz9Y6Z36HItiKg&e=>
>>>> $apache$spark$util$collection$ExternalAppendOnlyMapSuite$$testSimpleSpilling(ExternalAppendOnlyMapSuite.scala:263)
>>>> ...
>>>>
>>>> I feel like I've seen this before, and see some possibly relevant
>>>> fixes, but they're in 2.0.0 already:
>>>> https://github.com/apache/spark/pull/10990
>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__github.com_apache_spark_pull_10990&d=DQMFaQ&c=izlc9mHr637UR4lpLEZLFFS3Vn2UXBrZ4tFb6oOnmz8&r=hzwIMNQ9E99EMYGuqHI0kXhVbvX3nU3OSDadUnJxjAs&m=Y3d-oJvw2gK_2KXYjXY8_yzfAosPOqqaV4wtMg6ZPwM&s=dFymYD9NRVHIJ5MKpmzPcH_NYwLjOWcZd7FUuQBpTUU&e=>
>>>> Is this something where a native library needs to be installed or
>>>> something?
>>>>
>>>>
>>>> - to UTC timestamp *** FAILED ***
>>>>   "2016-03-13 [02]:00:00.0" did not equal "2016-03-13 [10]:00:00.0"
>>>> (DateTimeUtilsSuite.scala:506)
>>>>
>>>> I know, we talked about this for the 1.6.2 RC, but I reproduced this
>>>> locally too. I will investigate, could still be spurious.
>>>>
>>>>
>>>> StateStoreSuite:
>>>> - maintenance *** FAILED ***
>>>>   The code passed to eventually never returned normally. Attempted 627
>>>> times over 10.000180116 seconds. Last failure message:
>>>> StateStoreSuite.this.fileExists(provider, 1L, false) was true earliest
>>>> file not deleted. (StateStoreSuite.scala:395)
>>>>
>>>> No idea.
>>>>
>>>>
>>>> - offset recovery *** FAILED ***
>>>>   The code passed to eventually never returned normally. Attempted 197
>>>> times over 10.040864806 seconds. Last failure message:
>>>> strings.forall({
>>>>     ((x$1: Any) => DirectKafkaStreamSuite.collectedData.contains(x$1))
>>>>   }) was false. (DirectKafkaStreamSuite.scala:250)
>>>>
>>>> Also something that was possibly fixed already for 2.0.0 and that I
>>>> just back-ported into 1.6. Could be just a very similar failure.
>>>>
>>>> On Wed, Jun 22, 2016 at 2:26 AM, Reynold Xin <r...@databricks.com>
>>>> wrote:
>>>> > Please vote on releasing the following candidate as Apache Spark
>>>> version
>>>> > 2.0.0. The vote is open until Friday, June 24, 2016 at 19:00 PDT and
>>>> passes
>>>> > if a majority of at least 3+1 PMC votes are cast.
>>>> >
>>>> > [ ] +1 Release this package as Apache Spark 2.0.0
>>>> > [ ] -1 Do not release this package because ...
>>>> >
>>>> >
>>>> > The tag to be voted on is v2.0.0-rc1
>>>> > (0c66ca41afade6db73c9aeddd5aed6e5dcea90df).
>>>> >
>>>> > This release candidate resolves ~2400 issues:
>>>> > https://s.apache.org/spark-2.0.0-rc1-jira
>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__s.apache.org_spark-2D2.0.0-2Drc1-2Djira&d=DQMFaQ&c=izlc9mHr637UR4lpLEZLFFS3Vn2UXBrZ4tFb6oOnmz8&r=hzwIMNQ9E99EMYGuqHI0kXhVbvX3nU3OSDadUnJxjAs&m=Y3d-oJvw2gK_2KXYjXY8_yzfAosPOqqaV4wtMg6ZPwM&s=ZD_PezvsJ1GyDhv7MhaeUrVba_uhED5mPkqKpfenKEE&e=>
>>>> >
>>>> > The release files, including signatures, digests, etc. can be found
>>>> at:
>>>> >
>>>> http://people.apache.org/~pwendell/spark-releases/spark-2.0.0-rc1-bin/
>>>> <https://urldefense.proofpoint.com/v2/url?u=http-3A__people.apache.org_-7Epwendell_spark-2Dreleases_spark-2D2.0.0-2Drc1-2Dbin_&d=DQMFaQ&c=izlc9mHr637UR4lpLEZLFFS3Vn2UXBrZ4tFb6oOnmz8&r=hzwIMNQ9E99EMYGuqHI0kXhVbvX3nU3OSDadUnJxjAs&m=Y3d-oJvw2gK_2KXYjXY8_yzfAosPOqqaV4wtMg6ZPwM&s=wSbzZ2LyuDcNKaCijEPdt9rokQ0R9w66tn2jMfjKN2I&e=>
>>>> >
>>>> > Release artifacts are signed with the following key:
>>>> > https://people.apache.org/keys/committer/pwendell.asc
>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__people.apache.org_keys_committer_pwendell.asc&d=DQMFaQ&c=izlc9mHr637UR4lpLEZLFFS3Vn2UXBrZ4tFb6oOnmz8&r=hzwIMNQ9E99EMYGuqHI0kXhVbvX3nU3OSDadUnJxjAs&m=Y3d-oJvw2gK_2KXYjXY8_yzfAosPOqqaV4wtMg6ZPwM&s=i1Uxw1NyUf2iuA3CXbyiEODD1RR24rAXUvkc42ut8Ao&e=>
>>>> >
>>>> > The staging repository for this release can be found at:
>>>> >
>>>> https://repository.apache.org/content/repositories/orgapachespark-1187/
>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__repository.apache.org_content_repositories_orgapachespark-2D1187_&d=DQMFaQ&c=izlc9mHr637UR4lpLEZLFFS3Vn2UXBrZ4tFb6oOnmz8&r=hzwIMNQ9E99EMYGuqHI0kXhVbvX3nU3OSDadUnJxjAs&m=Y3d-oJvw2gK_2KXYjXY8_yzfAosPOqqaV4wtMg6ZPwM&s=QjsvnxXe6JBQqXwKw6r-fIIHI9E0ugeeICAqjRXRNwc&e=>
>>>> >
>>>> > The documentation corresponding to this release can be found at:
>>>> >
>>>> http://people.apache.org/~pwendell/spark-releases/spark-2.0.0-rc1-docs/
>>>> <https://urldefense.proofpoint.com/v2/url?u=http-3A__people.apache.org_-7Epwendell_spark-2Dreleases_spark-2D2.0.0-2Drc1-2Ddocs_&d=DQMFaQ&c=izlc9mHr637UR4lpLEZLFFS3Vn2UXBrZ4tFb6oOnmz8&r=hzwIMNQ9E99EMYGuqHI0kXhVbvX3nU3OSDadUnJxjAs&m=Y3d-oJvw2gK_2KXYjXY8_yzfAosPOqqaV4wtMg6ZPwM&s=_6IZExLgc8WoxW0kft_weR7AvELgbFXnHZdezQ_IYGk&e=>
>>>> >
>>>> >
>>>> > =======================================
>>>> > == How can I help test this release? ==
>>>> > =======================================
>>>> > If you are a Spark user, you can help us test this release by taking
>>>> an
>>>> > existing Spark workload and running on this release candidate, then
>>>> > reporting any regressions from 1.x.
>>>> >
>>>> > ================================================
>>>> > == What justifies a -1 vote for this release? ==
>>>> > ================================================
>>>> > Critical bugs impacting major functionalities.
>>>> >
>>>> > Bugs already present in 1.x, missing features, or bugs related to new
>>>> > features will not necessarily block this release. Note that
>>>> historically
>>>> > Spark documentation has been published on the website separately from
>>>> the
>>>> > main release so we do not need to block the release due to
>>>> documentation
>>>> > errors either.
>>>> >
>>>> >
>>>>
>>>> ---------------------------------------------------------------------
>>>> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
>>>> For additional commands, e-mail: dev-h...@spark.apache.org
>>>>
>>>>
>>>
>>


-- 


*Sincerely yoursEgor Pakhomov*

Reply via email to