Hi What are the design approaches I can follow to ensure that data is consistent from an application perspective (not from individual tables perspective). I am thinking of issues which arise due to unavailability of rollback or executing atomic transactions in Cassandra. Is Cassandra not suitable for my project?
Cassandra recommends creating a new table for each query. This results in data duplication (which doesn’t bother me). Take the following scenario. An application which allows users to create, share and manage food recipes. Each of the function below adds records in a separate database for {savedRecipe <- saveInRecipeRepository(...) recipeTagRepository <- saveRecipeTag(...) partitionInfoOfRecipes <- savePartitionOfTheTag(...) updatedUserProfile <- updateInUserProfile(...) recipesByUser <- saveRecipesCreatedByUser(...) supportedRecipes <- updateSupportedRecipesInformation(tag)} If say updateInUserProfile fails, then I'll have to manage rollback in the application itself as Cassandra doesn’t do it. My concerns is that the rollback process could itself fail due to network issues say. Is there a recommended way or a design principle I can follow to keep data consistent? Thanks Manu Sent from Mail<https://go.microsoft.com/fwlink/?LinkId=550986> for Windows 10