Hi , I am writing spark rdd to Ignie Cache . If I do ,
*val *df = ic.fromCache("EntitySettingsCache").sql("select * from entity_settingsā).withColumn("partitionkey",*col*("partitionkey")/2) ic.fromCache("EntitySettingsCache").saveValues(df.*rdd*) *Not able to reflect changes in cache* , but i do , val rddPair: RDD[EntitySettingsKey , EntitySettings] = converRddToPair(rdd) ic.fromCache("EntitySettingsCache").savePairs(df.*rdd , true*) above code is relfecting changes in Cache . If I see implementation of *saveValues* and *savePairs* , there is one difference , *streamer.allowOverwrite(overwrite)* above line is missing in saveValues , can anyone please help me , how do I solve the problem , because I dont' want to convert *rdd* to *rddPair* manually everytime .