Hi, Yay, I followed the steps you described and got the following error. Any idea?
vagrant up creating provisioner directive for running tests Bringing machine 'bigtop1' up with 'virtualbox' provider... ==> bigtop1: Box 'puppetlab-centos-64-nocm' could not be found. Attempting to find and install... bigtop1: Box Provider: virtualbox bigtop1: Box Version: >= 0 ==> bigtop1: Adding box 'puppetlab-centos-64-nocm' (v0) for provider: virtualbox bigtop1: Downloading: http://puppet-vagrant-boxes.puppetlabs.com/centos-64-x64-vbox4210-nocm.box ==> bigtop1: Successfully added box 'puppetlab-centos-64-nocm' (v0) for 'virtualbox'! There are errors in the configuration of this machine. Please fix the following errors and try again: vm: * The 'hostmanager' provisioner could not be found. Thanks Jim On Nov 4, 2014, at 6:36 PM, jay vyas <jayunit100.apa...@gmail.com> wrote: > Hi daemon: Actually, for most folks who would want to actually use a hadoop > cluster, i would think setting up bigtop is super easy ! If you have issues > with it ping me and I can help you get started. > Also, we have docker containers - so you dont even *need* a VM to run a 4 or > 5 node hadoop cluster. > > install vagrant > install VirtualBox > git clone https://github.com/apache/bigtop > cd bigtop/bigtop-deploy/vm/vagrant-puppet > vagrant up > Then vagrant destroy when your done. > > This to me is easier than manually downloading an appliance, picking memory > starting the virtualbox gui, loading the appliance , etc... and also its > easy to turn the simple single node bigtop VM into a multinode one, > by just modifying the vagrantile. > > > On Tue, Nov 4, 2014 at 5:32 PM, daemeon reiydelle <daeme...@gmail.com> wrote: > What you want as a sandbox depends on what you are trying to learn. > > If you are trying to learn to code in e.g PigLatin, Sqooz, or similar, all of > the suggestions (perhaps excluding BigTop due to its setup complexities) are > great. Laptop? perhaps but laptop's are really kind of infuriatingly slow > (because of the hardware - you pay a price for a 30-45watt average heating > bill). A laptop is an OK place to start if it is e.g. an i5 or i7 with lots > of memory. What do you think of the thought that you will pretty quickly > graduate to wanting a small'ish desktop for your sandbox? > > A simple, single node, Hadoop instance will let you learn many things. The > next level of complexity comes when you are attempting to deal with data > whose processing needs to be split up, so you can learn about how to split > data in Mapping, reduce the splits via reduce jobs, etc. For that, you could > get a windows desktop box or e.g. RedHat/CentOS and use virtualization. > Something like a 4 core i5 with 32gb of memory, running 3 or for some things > 4, vm's. You could load e.g. hortonworks into each of the vm's and practice > setting up a 3/4 way cluster. Throw in 2-3 1tb drives off of eBay and you can > have a lot of learning. > > > > > > ....... > “The race is not to the swift, > nor the battle to the strong, > but to those who can see it coming and jump aside.” - Hunter Thompson > Daemeon > > On Tue, Nov 4, 2014 at 1:24 PM, oscar sumano <osum...@gmail.com> wrote: > you can try the pivotal vm as well. > > http://pivotalhd.docs.pivotal.io/tutorial/getting-started/pivotalhd-vm.html > > On Tue, Nov 4, 2014 at 3:13 PM, Leonid Fedotov <lfedo...@hortonworks.com> > wrote: > Tim, > download Sandbox from http://hortonworks/com > You will have everything needed in a small VM instance which will run on your > home desktop. > > > Thank you! > > > > Sincerely, > > Leonid Fedotov > > Systems Architect - Professional Services > > lfedo...@hortonworks.com > > office: +1 855 846 7866 ext 292 > > mobile: +1 650 430 1673 > > > On Tue, Nov 4, 2014 at 11:28 AM, Tim Dunphy <bluethu...@gmail.com> wrote: > Hey all, > > I want to setup an environment where I can teach myself hadoop. Usually the > way I'll handle this is to grab a machine off the Amazon free tier and setup > whatever software I want. > > However I realize that Hadoop is a memory intensive, big data solution. So > what I'm wondering is, would a t2.micro instance be sufficient for setting up > a cluster of hadoop nodes with the intention of learning it? To keep things > running longer in the free tier I would either setup however many nodes as I > want and keep them stopped when I'm not actively using them. Or just setup a > few nodes with a few different accounts (with a different gmail address for > each one.. easy enough to do). > > Failing that, what are some other free/cheap solutions for setting up a > hadoop learning environment? > > Thanks, > Tim > > -- > GPG me!! > > gpg --keyserver pool.sks-keyservers.net --recv-keys F186197B > > > > CONFIDENTIALITY NOTICE > NOTICE: This message is intended for the use of the individual or entity to > which it is addressed and may contain information that is confidential, > privileged and exempt from disclosure under applicable law. If the reader of > this message is not the intended recipient, you are hereby notified that any > printing, copying, dissemination, distribution, disclosure or forwarding of > this communication is strictly prohibited. If you have received this > communication in error, please contact the sender immediately and delete it > from your system. Thank You. > > > > > > -- > jay vyas