[ https://issues.apache.org/jira/browse/SPARK-27124?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16789562#comment-16789562 ]
Hyukjin Kwon commented on SPARK-27124: -------------------------------------- I think we can mention that generally somewhere, say, Py4J approach can be used to access to internal codes in Spark. To be honest, this approach requires to understand Py4J. So, we shouldn't focus on how to use it tho. > Expose org.apache.spark.sql.avro.SchemaConverters as developer API > ------------------------------------------------------------------ > > Key: SPARK-27124 > URL: https://issues.apache.org/jira/browse/SPARK-27124 > Project: Spark > Issue Type: Improvement > Components: PySpark, SQL > Affects Versions: 3.0.0 > Reporter: Gabor Somogyi > Priority: Minor > > org.apache.spark.sql.avro.SchemaConverters provides extremely useful APIs to > convert schema between Spark SQL and avro. This is reachable from scala side > but not from pyspark. I suggest to add this as a developer API to ease > development for pyspark users. -- This message was sent by Atlassian JIRA (v7.6.3#76005) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org