Fernando Pereira created SPARK-22250:
----------------------------------------

             Summary: Be less restrictive on type checking
                 Key: SPARK-22250
                 URL: https://issues.apache.org/jira/browse/SPARK-22250
             Project: Spark
          Issue Type: Bug
          Components: PySpark
    Affects Versions: 2.0.0
            Reporter: Fernando Pereira
            Priority: Minor


I find types.py._verify_type() often too restrictive. E.g. 
{code}
TypeError: FloatType can not accept object 0 in type <type 'int'>
{code}
I believe it would be globally acceptable to fill a float field with an int, 
especially since in some formats (json) you don't have a way of inferring the 
type correctly.

Another situation relates to other equivalent numerical types, like array.array 
or numpy. A numpy scalar int is not accepted as an int, and these arrays have 
always to be converted down to plain lists, which can be prohibitively large 
and computationally expensive.

Any thoughts?



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to