I believe this problem led to opening SPARK-34389 where the problem is
discussed further.
[1] https://issues.apache.org/jira/browse/SPARK-34389
--
Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/
-
To
A much better one-liner (easier to understand the UI because it will be 1
simple job with 2 stages):
```
spark.read.text("README.md").repartition(2).take(1)
```
Attila Zsolt Piros wrote
> No, it won't be reused.
> You should reuse the dateframe for reusing the shuffle blocks (and cached
>