1 ) using mysql DB
2 ) will be inserting/update/overwrite to the same table
3 ) i want to update a specific column in a record, the data is read via
Spark SQL,

on the below table which is read via sparkSQL, i would like to update the
NumOfSamples column .

consider DF as the dataFrame which holds the records,  registered as
temporary table MS .

spark.sqlContext.write.format("jdbc").option("url", url ).option("dbtable",
"update ms  set NumOfSamples = 20 where 'TimeSeriesID = '1000'" As MS ).save

I believe updating a record via sparkSQL is not supported,  the only
workaround is to open up a jdbc connection without using spark API's and do
a direct update ?..

Sample Ex : -

mysql> show columns from ms;
+--------------+-------------+------+-----+---------+-------+
| Field        | Type        | Null | Key | Default | Extra |
+--------------+-------------+------+-----+---------+-------+
| TimeSeriesID | varchar(20) | YES  |     | NULL    |       |
| NumOfSamples | int(11)     | YES  |     | NULL    |       |
+--------------+-------------+------+-----+---------+-------+


Thanks,
Sujeet



On Tue, Aug 9, 2016 at 6:31 PM, Mich Talebzadeh <mich.talebza...@gmail.com>
wrote:

> Hi,
>
>
>    1. what is the underlying DB, say Hive etc
>    2. Is table transactional or you are going to do insert/overwrite to
>    the same table
>    3. can you do all this in the database itself assuming it is an RDBMS
>    4. Can you provide the sql or pseudo code for such an update
>
>
> HTH
>
> Dr Mich Talebzadeh
>
>
>
> LinkedIn * 
> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*
>
>
>
> http://talebzadehmich.wordpress.com
>
>
> *Disclaimer:* Use it at your own risk. Any and all responsibility for any
> loss, damage or destruction of data or any other property which may arise
> from relying on this email's technical content is explicitly disclaimed.
> The author will in no case be liable for any monetary damages arising from
> such loss, damage or destruction.
>
>
>
> On 9 August 2016 at 13:39, sujeet jog <sujeet....@gmail.com> wrote:
>
>> Hi,
>>
>> Is it possible to update certain columnr records  in DB  from spark,
>>
>> for example i have 10 rows with 3 columns  which are read from Spark SQL,
>>
>> i want to update specific column entries  and write back to DB, but since
>> RDD"s are immutable i believe this would be difficult, is there a
>> workaround.
>>
>>
>> Thanks,
>> Sujeet
>>
>
>

Reply via email to