Hi Kevin,

Can you try https://issues.apache.org/jira/browse/SPARK-1898 to see if it
fixes your issue?

Running in YARN cluster mode, I had a similar issue where Spark was able to
create a Driver and an Executor via YARN, but then it stopped making any
progress.

Note: I was using a pre-release version of CDH5.1.0, not 2.3 like you were
using.

best,
Colin


On Wed, May 21, 2014 at 3:34 PM, Kevin Markey <kevin.mar...@oracle.com>wrote:

> 0
>
> Abstaining because I'm not sure if my failures are due to Spark,
> configuration, or other factors...
>
> Compiled and deployed RC10 for YARN, Hadoop 2.3 per Spark 1.0.0 Yarn
> documentation.  No problems.
> Rebuilt applications against RC10 and Hadoop 2.3.0 (plain vanilla Apache
> release).
> Updated scripts for various applications.
> Application had successfully compiled and run against Spark 0.9.1 and
> Hadoop 2.3.0.
> Ran in "yarn-cluster" mode.
> Application ran to conclusion except that it ultimately failed because of
> an exception when Spark tried to clean up the staging directory.  Also,
> where before Yarn would report the running program as "RUNNING", it only
> reported this application as "ACCEPTED".  It appeared to run two containers
> when the first instance never reported that it was RUNNING.
>
> I will post a separate note to the USER list about the specifics.
>
> Thanks
> Kevin Markey
>
>
>
> On 05/21/2014 10:58 AM, Mark Hamstra wrote:
>
>> +1
>>
>>
>> On Tue, May 20, 2014 at 11:09 PM, Henry Saputra <henry.sapu...@gmail.com>
>> wrote:
>>
>>  Signature and hash for source looks good
>>> No external executable package with source - good
>>> Compiled with git and maven - good
>>> Ran examples and sample programs locally and standalone -good
>>>
>>> +1
>>>
>>> - Henry
>>>
>>>
>>>
>>> On Tue, May 20, 2014 at 1:13 PM, Tathagata Das
>>> <tathagata.das1...@gmail.com> wrote:
>>>
>>>> Please vote on releasing the following candidate as Apache Spark version
>>>>
>>> 1.0.0!
>>>
>>>> This has a few bug fixes on top of rc9:
>>>> SPARK-1875: https://github.com/apache/spark/pull/824
>>>> SPARK-1876: https://github.com/apache/spark/pull/819
>>>> SPARK-1878: https://github.com/apache/spark/pull/822
>>>> SPARK-1879: https://github.com/apache/spark/pull/823
>>>>
>>>> The tag to be voted on is v1.0.0-rc10 (commit d8070234):
>>>>
>>>>  https://git-wip-us.apache.org/repos/asf?p=spark.git;a=commit;h=
>>> d807023479ce10aec28ef3c1ab646ddefc2e663c
>>>
>>>> The release files, including signatures, digests, etc. can be found at:
>>>> http://people.apache.org/~tdas/spark-1.0.0-rc10/
>>>>
>>>> The release artifacts are signed with the following key:
>>>> https://people.apache.org/keys/committer/tdas.asc
>>>>
>>>> The staging repository for this release can be found at:
>>>> https://repository.apache.org/content/repositories/orgapachespark-1018/
>>>>
>>>> The documentation corresponding to this release can be found at:
>>>> http://people.apache.org/~tdas/spark-1.0.0-rc10-docs/
>>>>
>>>> The full list of changes in this release can be found at:
>>>>
>>>>  https://git-wip-us.apache.org/repos/asf?p=spark.git;a=blob;
>>> f=CHANGES.txt;h=d21f0ace6326e099360975002797eb7cba9d5273;hb=
>>> d807023479ce10aec28ef3c1ab646ddefc2e663c
>>>
>>>> Please vote on releasing this package as Apache Spark 1.0.0!
>>>>
>>>> The vote is open until Friday, May 23, at 20:00 UTC and passes if
>>>> amajority of at least 3 +1 PMC votes are cast.
>>>>
>>>> [ ] +1 Release this package as Apache Spark 1.0.0
>>>> [ ] -1 Do not release this package because ...
>>>>
>>>> To learn more about Apache Spark, please see
>>>> http://spark.apache.org/
>>>>
>>>> ====== API Changes ======
>>>> We welcome users to compile Spark applications against 1.0. There are
>>>> a few API changes in this release. Here are links to the associated
>>>> upgrade guides - user facing changes have been kept as small as
>>>> possible.
>>>>
>>>> Changes to ML vector specification:
>>>>
>>>>  http://people.apache.org/~tdas/spark-1.0.0-rc10-docs/
>>> mllib-guide.html#from-09-to-10
>>>
>>>> Changes to the Java API:
>>>>
>>>>  http://people.apache.org/~tdas/spark-1.0.0-rc10-docs/
>>> java-programming-guide.html#upgrading-from-pre-10-versions-of-spark
>>>
>>>> Changes to the streaming API:
>>>>
>>>>  http://people.apache.org/~tdas/spark-1.0.0-rc10-docs/
>>> streaming-programming-guide.html#migration-guide-from-091-or-below-to-1x
>>>
>>>> Changes to the GraphX API:
>>>>
>>>>  http://people.apache.org/~tdas/spark-1.0.0-rc10-docs/
>>> graphx-programming-guide.html#upgrade-guide-from-spark-091
>>>
>>>> Other changes:
>>>> coGroup and related functions now return Iterable[T] instead of Seq[T]
>>>> ==> Call toSeq on the result to restore the old behavior
>>>>
>>>> SparkContext.jarOfClass returns Option[String] instead of Seq[String]
>>>> ==> Call toSeq on the result to restore old behavior
>>>>
>>>
>

Reply via email to