Le sam. 11 juil. 2015 à 14:53, Roman Sokolov <ole...@gmail.com> a écrit :
> Hello. Had the same question. What if I need to store 4-6 Tb and do > queries? Can't find any clue in documentation. > Am 11.07.2015 03:28 schrieb "Mohammed Guller" <moham...@glassbeam.com>: > >> Hi Ravi, >> >> First, Neither Spark nor Spark SQL is a database. Both are compute >> engines, which need to be paired with a storage system. Seconds, they are >> designed for processing large distributed datasets. If you have only >> 100,000 records or even a million records, you don’t need Spark. A RDBMS >> will perform much better for that volume of data. >> >> >> >> Mohammed >> >> >> >> *From:* Ravisankar Mani [mailto:rrav...@gmail.com] >> *Sent:* Friday, July 10, 2015 3:50 AM >> *To:* user@spark.apache.org >> *Subject:* Spark performance >> >> >> >> Hi everyone, >> >> I have planned to move mssql server to spark?. I have using around >> 50,000 to 1l records. >> >> The spark performance is slow when compared to mssql server. >> >> >> >> What is the best data base(Spark or sql) to store or retrieve data around >> 50,000 to 1l records ? >> >> regards, >> >> Ravi >> >> >> >