[ https://issues.apache.org/jira/browse/SPARK-13842?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15231209#comment-15231209 ]
holdenk commented on SPARK-13842: --------------------------------- So the testing frameworks is a mixture of doctests along with the standard unittest2 stuff (in the tests.py file in each sub directory). Let me know if you have any questions while your doing this that I can help with :) > Consider __iter__ and __getitem__ methods for pyspark.sql.types.StructType > -------------------------------------------------------------------------- > > Key: SPARK-13842 > URL: https://issues.apache.org/jira/browse/SPARK-13842 > Project: Spark > Issue Type: Improvement > Components: PySpark > Affects Versions: 1.6.1 > Reporter: Shea Parkes > Priority: Minor > Original Estimate: 24h > Remaining Estimate: 24h > > It would be nice to consider adding \_\_iter\_\_ and \_\_getitem\_\_ to > {{pyspark.sql.types.StructType}}. Here are some simplistic suggestions: > {code} > def __iter__(self): > """Iterate the fields upon request.""" > return iter(self.fields) > def __getitem__(self, key): > """Return the corresponding StructField""" > _fields_dict = dict(zip(self.names, self.fields)) > try: > return _fields_dict[key] > except KeyError: > raise KeyError('No field named {}'.format(key)) > {code} > I realize the latter might be a touch more controversial since there could be > name collisions. Still, I doubt there are that many in practice and it would > be quite nice to work with. > Privately, I have more extensive metadata extraction methods overlaid on this > class, but I imagine the rest of what I have done might go too far for the > common user. If this request gains traction though, I'll share those other > layers. -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org