deng created SPARK-15427: ---------------------------- Summary: Spark SQL doesn't support field case sensitive when load data use Phoenix Key: SPARK-15427 URL: https://issues.apache.org/jira/browse/SPARK-15427 Project: Spark Issue Type: Bug Components: Spark Core, SQL Affects Versions: 1.5.0 Reporter: deng
i am use sparkSql load data from Apache Phoenix. SQLContext sqlContext = new SQLContext(sc); Map<String, String> options = new HashMap(); options.put("driver", driver); options.put("url", PhoenixUtil.p.getProperty("phoenixURL")); options.put("dbtable", "(select "value","name" from "user")"); DataFrame jdbcDF = sqlContext.load("jdbc", options); It will always throws exception, like "can't find field VALUE". I track the code and find spark will use: val rs = conn.prepareStatement(s"SELECT * FROM $table WHERE 1=0").executeQuery() to get the field.But the field already be uppercase like "value" to VALUE. So it will always throws "can't find field VALUE"; It didn't think of the the case when data loaded from source in which filed is case sensitive. -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org