Artem Kalchenko created SPARK-27591:
---------------------------------------

             Summary: A bug in UnivocityParser prevents using UDT
                 Key: SPARK-27591
                 URL: https://issues.apache.org/jira/browse/SPARK-27591
             Project: Spark
          Issue Type: Bug
          Components: SQL
    Affects Versions: 2.4.2
            Reporter: Artem Kalchenko


I am trying to define a UserDefinedType based on String but different from 
StringType in Spark 2.4.1 but it looks like there is a bug in Spark or I am 
doing smth incorrectly.

I define my type as follows:
{code:java}
class MyType extends UserDefinedType[MyValue] {
  override def sqlType: DataType = StringType
  ...
}

@SQLUserDefinedType(udt = classOf[MyType])
case class MyValue
{code}
I expect it to be read and stored as String with just a custom SQL type. In 
fact Spark can't read the string at all:
{code:java}
java.lang.ClassCastException: 
org.apache.spark.sql.execution.datasources.csv.UnivocityParser$$anonfun$makeConverter$11
 cannot be cast to org.apache.spark.unsafe.types.UTF8String
    at 
org.apache.spark.sql.catalyst.expressions.BaseGenericInternalRow$class.getUTF8String(rows.scala:46)
    at 
org.apache.spark.sql.catalyst.expressions.GenericInternalRow.getUTF8String(rows.scala:195)
    at 
org.apache.spark.sql.catalyst.expressions.JoinedRow.getUTF8String(JoinedRow.scala:102)
{code}
the problem is with UnivocityParser.makeConverter that doesn't return (String 
=> Any) function but (String => (String => Any)) in the case of UDT, see 
UnivocityParser:184
{code:java}
case udt: UserDefinedType[_] => (datum: String) =>
  makeConverter(name, udt.sqlType, nullable, options)

{code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to