(Starting over...)

The best place to look for the requirements would be at the individual
pages of each technology.

As for absolute minimum requirements, I would suggest 50GB of disk space
and at least 8GB of memory.  This is the absolute minimum.

"Architecting" a solution like you are looking for is very complex.  If you
are just looking for a proof-of-concept consider a Docker image or going to
Cloudera/Hortonworks/MapR and look for their "express VMs" which can
usually run on Oracle Virtualbox or VMware.

Marco


On Thu, Jan 22, 2015 at 7:36 AM, Sudipta Banerjee <
asudipta.baner...@gmail.com> wrote:

>
>
> Hi Apache-Spark team ,
>
> What are the system requirements installing Hadoop and Apache Spark?
> I have attached the screen shot of Gparted.
>
>
> Thanks and regards,
> Sudipta
>
>
>
>
> --
> Sudipta Banerjee
> Consultant, Business Analytics and Cloud Based Architecture
> Call me +919019578099
>
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
>

Reply via email to