juliuszsompolski opened a new pull request, #42331:
URL: https://github.com/apache/spark/pull/42331

   ### What changes were proposed in this pull request?
   
   This makes sure that all iterators used in Spark Connect scala client are 
`CloseableIterator`.
   
   1. Makes `CustomSparkConnectBlockingStub.executePlan` return 
`CloseableIterator` and make all wrappers respect that.
   2. Makes `ExecutePlanResponseReattachableIterator` a `CloseableIterator`, 
with an implementation that will inform the server that query result can be 
released with ReleaseExecute.
   3. Makes `SparkResult.iterator` explicitly a `CloseableIterator`, and also 
register the `SparkResult.responses` iterator as with the 
`SparkResultCloseable` cleaner, which will make it close upon GC, if not closed 
explicitly sooner.
   4. Because `Dataset.toLocalIterator` requires a Java iterator, implement a 
conversion to `java.util.Iterator with AutoCloseable` to be returned there
   5. Using `CloseableIterator` consistently everywhere else removes the need 
to convert between iterator types.
   
   ### Why are the changes needed?
   
   Properly closeable iterators are needed for resource management, and with 
reattachable execution to inform server that processing finished.
   
   ### Does this PR introduce _any_ user-facing change?
   
   No.
   
   ### How was this patch tested?
   
   Exercise current E2E tests.
   
   Co-authored-by: Alice Sayutina <alice.sayut...@databricks.com>
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to