Hi Josh,

Thanks for the informative answer. Sounds like one should await your changes in 
1.3. As information, I found the following set of options for doing the visual 
in a notebook.

http://nbviewer.ipython.org/github/ipython/ipython/blob/3607712653c66d63e0d7f13f073bde8c0f209ba8/docs/examples/notebooks/Animations_and_Progress.ipynb


> On Dec 27, 2014, at 4:07 PM, Josh Rosen <rosenvi...@gmail.com> wrote:
> 
> The console progress bars are implemented on top of a new stable "status API" 
> that was added in Spark 1.2.  It's possible to query job progress using this 
> interface (in older versions of Spark, you could implement a custom 
> SparkListener and maintain the counts of completed / running / failed tasks / 
> stages yourself). 
> 
> There are actually several subtleties involved in implementing "job-level" 
> progress bars which behave in an intuitive way; there's a pretty extensive 
> discussion of the challenges at https://github.com/apache/spark/pull/3009.  
> Also, check out the pull request for the console progress bars for an 
> interesting design discussion around how they handle parallel stages: 
> https://github.com/apache/spark/pull/3029.
> 
> I'm not sure about the plumbing that would be necessary to display live 
> progress updates in the IPython notebook UI, though.  The general pattern 
> would probably involve a mapping to relate notebook cells to Spark jobs (you 
> can do this with job groups, I think), plus some periodic timer that polls 
> the driver for the status of the current job in order to update the progress 
> bar.
> 
> For Spark 1.3, I'm working on designing a REST interface to accesses this 
> type of job / stage / task progress information, as well as expanding the 
> types of information exposed through the stable status API interface.
> 
> - Josh
> 
>> On Thu, Dec 25, 2014 at 10:01 AM, Eric Friedman <eric.d.fried...@gmail.com> 
>> wrote:
>> Spark 1.2.0 is SO much more usable than previous releases -- many thanks to 
>> the team for this release.
>> 
>> A question about progress of actions.  I can see how things are progressing 
>> using the Spark UI.  I can also see the nice ASCII art animation on the 
>> spark driver console.
>> 
>> Has anyone come up with a way to accomplish something similar in an iPython 
>> notebook using pyspark?
>> 
>> Thanks
>> Eric
> 

Reply via email to