(Starting over...) The best place to look for the requirements would be at the individual pages of each technology.
As for absolute minimum requirements, I would suggest 50GB of disk space and at least 8GB of memory. This is the absolute minimum. "Architecting" a solution like you are looking for is very complex. If you are just looking for a proof-of-concept consider a Docker image or going to Cloudera/Hortonworks/MapR and look for their "express VMs" which can usually run on Oracle Virtualbox or VMware. Marco On Thu, Jan 22, 2015 at 7:36 AM, Sudipta Banerjee < asudipta.baner...@gmail.com> wrote: > > > Hi Apache-Spark team , > > What are the system requirements installing Hadoop and Apache Spark? > I have attached the screen shot of Gparted. > > > Thanks and regards, > Sudipta > > > > > -- > Sudipta Banerjee > Consultant, Business Analytics and Cloud Based Architecture > Call me +919019578099 > > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > For additional commands, e-mail: user-h...@spark.apache.org >