You may access it via something like |SELECT filterIp.element FROM tb|, just like Hive. Or if you’re using Spark SQL DSL, you can use |tb.select("filterIp.element".attr)|.

On 12/8/14 1:08 PM, Xuelin Cao wrote:


Hi,

    I'm generating a Spark SQL table from an offline Json file.

The difficulty is, in the original json file, there is a hierarchical structure. And, as a result, this is what I get:

scala> tb.printSchema
root
 |-- budget: double (nullable = true)
* |-- filterIp: array (nullable = true)*
* |    |-- element: string (containsNull = false)*
 |-- status: integer (nullable = true)
 |-- third_party: integer (nullable = true)
 |-- userId: integer (nullable = true)

As you may have noticed, the table schema is with a hierarchical structure ("element" field is a sub-field under the "filterIp" field). Then, my question is, how do I access the "element" field with SQL?


Reply via email to