Drill Members, I have an intriguing problem where I have hundreds of thousands or even millions of records stored in jsonb format in uneven JSon objects in a PostgreSQL database. I would like to be able to implicitly grab column names and data types using your tool since neither PostreSQL or Pentaho have this function. My current tool for this task is Scala but I am not dedicated to writing and keeping up with the JSon format specs; etc. in real time.
Is it possible by conversion or otherwise to read the jsonb or a text string from PostgreSQL as if it were being queried from a "json" type storage instead of a "jdbc" type storage? If so, could I pull in different columns from PostgreSQL without some sort of key (with the original query)? Is there the ability to do some thing like SELECT pkey, split_to_columns(convert_to(field,'JSON')) FROM postgres.mytable? For context, I have posted an example record below: pkey , data 1423234, "{"loadCaseResult": {"case": {"type": "Adoption (Family)", "judge": null, "style": "Confidential", "caseId": "12", "events": {"event": [{"date": "08/06/2014", "type": "Request: Action", "comment": "900125-Request for Action \n Filed/Issued By"}}} Thank you for your time, Andrew Evans Java/Scala/Python Programmer at Hygenics Data,LLC