[ https://issues.apache.org/jira/browse/SPARK-6928?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Sean Owen reopened SPARK-6928: ------------------------------ OK, given there's a PR to consider, I'll reopen it, but I think the PR has some issues. I still personally suspect this isn't important to support. > spark-shell stops working after the replay command > -------------------------------------------------- > > Key: SPARK-6928 > URL: https://issues.apache.org/jira/browse/SPARK-6928 > Project: Spark > Issue Type: Bug > Components: Spark Shell > Affects Versions: 1.3.0 > Environment: Scala Version :Scala-2.10 > Reporter: Tijo Thomas > > Step to reproduce this issues. > Step 1 : > scala> sc.parallelize(1 to 10).map(_+"2").count(); > res0: Long = 10 > Step 2 : > scala> :replay > Replaying: sc.parallelize(1 to 10).map(_+"2").count(); > <console>:8: error: not found: value sc > sc.parallelize(1 to 10).map(_+"2").count(); > ^ > // Note : After Replay command , Non of the spark api's are working as the > SparkContext has gone out of scope. > eg: getting this exception as given below > scala> exit > error: > while compiling: <console> > during phase: jvm > library version: version 2.10.4 > compiler version: version 2.10.4 > reconstructed args: > last tree to typer: Apply(constructor $read) > symbol: constructor $read in class $read (flags: <method> > <triedcooking>) > symbol definition: def <init>(): $line20.$read > tpe: $line20.$read > symbol owners: constructor $read -> class $read -> package $line20 > context owners: class iwC -> package $line20 > ............ -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org