I have an RDD of logs that look like this: /no_cache/bi_event?Log=0&pg_inst=517638988975678942&pg=fow_mwe&ver=c.2.1.8&site=xyz.com&pid=156431807121222351&rid=156431666543211500&srch_id=156431666581865115&row=6&seq=1&tot=1&tsp=1&cmp=thmb_12&co_txt_url=Viewing&et=click&thmb_type=p&ct=u&c=579855&lnx=SPGOOGBRANDCAMP&ref_url=http%3A%2F%2Fwww.abcd.com
The pairs are separated by "&", and the keys/values of each pair are separated by "=". Hive has a str_to_map function <https://cwiki.apache.org/confluence/display/Hive/LanguageManual+UDF#LanguageManualUDF-StringFunctions> that will convert this String to a map that will make the following work: mappedString["site"] will return "xyz.com" What's the most efficient way to do this in Scala + Spark? -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Converting-Apache-log-string-into-map-using-delimiter-tp18641.html Sent from the Apache Spark User List mailing list archive at Nabble.com. --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org