Hi,
I am trying to use Spark to read from Oracle (12.1) table using Spark 2.0.
My table has JSON data.  I am getting below exception in my code. Any clue?

>>>>>
java.sql.SQLException: Unsupported type -101

at
org.apache.spark.sql.execution.datasources.jdbc.JdbcUtils$.org$apache$spark$sql$execution$datasources$jdbc$JdbcUtils$$getCatalystType(JdbcUtils.scala:233)
at
org.apache.spark.sql.execution.datasources.jdbc.JdbcUtils$$anonfun$8.apply(JdbcUtils.scala:290)
at
org.apache.spark.sql.execution.datasources.jdbc.JdbcUtils$$anonfun$8.apply(JdbcUtils.scala:290)
at scala.Option.getOrElse(Option.scala:121)
at

==========
My code is very simple.

SparkSession spark = SparkSession
        .builder()
        .appName("Oracle Example")
        .master("local[4]")
        .getOrCreate();

final Properties connectionProperties = new Properties();
connectionProperties.put("user", *"some_user"*));
connectionProperties.put("password", "some_pwd"));

final String dbTable =
        "(select *  from  MySampleTable)";

Dataset<Row> jdbcDF = spark.read().jdbc(*URL*, dbTable, connectionProperties);

Reply via email to