Yes, I'd prefer Terraform too.

On 2018-01-31 06:32:21, Michael Mior <mm...@uwaterloo.ca> wrote:
> While whatever format this comes out in would be helpful, you might want to
> consider Terraform. 1Password recently published a blog post on their
> experience with Terraform vs. CloudFormation.
> 
> https://blog.agilebits.com/2018/01/25/terraforming-1password/
> 
> --
> Michael Mior
> mm...@apache.org
> 
> 2018-01-31 2:34 GMT-05:00 Kenneth Brotman <kenbrot...@yahoo.com.invalid>:
> 
> > Hi Yuri,
> >
> > If possible I will do everything with AWS Cloudformation.  I'm working on
> > it now.  Nothing published yet.
> >
> > Kenneth Brotman
> >
> > -----Original Message-----
> > From: Yuri Subach [mailto:ysub...@gmail.com]
> > Sent: Tuesday, January 30, 2018 7:02 PM
> > To: user@cassandra.apache.org
> > Subject: RE: Slender Cassandra Cluster Project
> >
> > Hi Kenneth,
> >
> > I like this project idea!
> >
> > A couple of questions:
> > - What tools are you going to use for AWS cluster setup?
> > - Do you have anything published already (github)?
> >
> > On 2018-01-22 22:42:11, Kenneth Brotman <kenbrot...@yahoo.com.INVALID>
> > wrote:
> > > Thanks Anthony!  I’ve made a note to include that information in the
> > documentation. You’re right.  It won’t work as intended unless that is
> > configured properly.
> > >
> > >
> > >
> > > I’m also favoring a couple other guidelines for Slender Cassandra:
> > >
> > > 1.       SSD’s only, no spinning disks
> > >
> > > 2.       At least two cores per node
> > >
> > >
> > >
> > > For AWS, I’m favoring the c3.large on Linux.  It’s available in these
> > regions: US-East, US-West and US-West2.  The specifications are listed as:
> > >
> > > ·         Two (2) vCPU’s
> > >
> > > ·         3.7 Gib Memory
> > >
> > > ·         Two (2) 16 GB SSD’s
> > >
> > > ·         Moderate I/O
> > >
> > >
> > >
> > > It’s going to be hard to beat the inexpensive cost of operating a
> > Slender Cluster on demand in the cloud – and it fits a lot of the use cases
> > well:
> > >
> > >
> > >
> > > ·         For under a $100 a month, in current pricing for EC2
> > instances, you can operate an eighteen (18) node Slender Cluster for five
> > (5) hours a day, ten (10) days a month.  That’s fine for demonstrations,
> > teaching or experiments that last half a day or less.
> > >
> > > ·         For under $20, you can have that Slender Cluster up all day
> > long, up to ten (10) hours, for whatever demonstrations or experiments you
> > want it for.
> > >
> > >
> > >
> > > As always, feedback is encouraged.
> > >
> > >
> > >
> > > Thanks,
> > >
> > >
> > >
> > > Kenneth Brotman
> > >
> > >
> > >
> > > From: Anthony Grasso [mailto:anthony.gra...@gmail.com]
> > > Sent: Sunday, January 21, 2018 3:57 PM
> > > To: user
> > > Subject: Re: Slender Cassandra Cluster Project
> > >
> > >
> > >
> > > Hi Kenneth,
> > >
> > >
> > >
> > > Fantastic idea!
> > >
> > >
> > >
> > > One thing that came to mind from my reading of the proposed setup was
> > rack awareness of each node. Given that the proposed setup contains three
> > DCs, I assume that each node will be made rack aware? If not, consider
> > defining three racks for each DC and placing two nodes in each rack. This
> > will ensure that all the nodes in a single rack contain at most one replica
> > of the data.
> > >
> > >
> > >
> > > Regards,
> > >
> > > Anthony
> > >
> > >
> > >
> > > On 17 January 2018 at 11:24, Kenneth Brotman
> > <kenbrot...@yahoo.com.invalid> wrote:
> > >
> > > Sure.  That takes the project from awesome to 10X awesome.  I absolutely
> > would be willing to do that.  Thanks Kurt!
> > >
> > >
> > >
> > > Regarding your comment on the keyspaces, I agree.  There should be a few
> > simple examples one way or the other that can be duplicated and observed,
> > and then an example to duplicate and play with that has a nice real world
> > mix, with some keyspaces that replicate over only a subset of DC’s and some
> > that replicate to all DC’s.
> > >
> > >
> > >
> > > Kenneth Brotman
> > >
> > >
> > >
> > > From: kurt greaves [mailto:k...@instaclustr.com]
> > > Sent: Tuesday, January 16, 2018 1:31 PM
> > > To: User
> > > Subject: Re: Slender Cassandra Cluster Project
> > >
> > >
> > >
> > > Sounds like a great idea. Probably would be valuable to add to the
> > official docs as an example set up if you're willing.
> > >
> > >
> > >
> > > Only thing I'd add is that you should have keyspaces that replicate over
> > only a subset of DC's, plus one/some replicated to all DC's
> > >
> > >
> > >
> > > On 17 Jan. 2018 03:26, "Kenneth Brotman" <kenbrot...@yahoo.com.invalid>
> > wrote:
> > >
> > > I’ve begun working on a reference project intended to provide guidance
> > on configuring and operating a modest Cassandra cluster of about 18 nodes
> > suitable for the economic study, demonstration, experimentation and testing
> > of a Cassandra cluster.
> > >
> > >
> > >
> > > The slender cluster would be designed to be as inexpensive as possible
> > while still using real world hardware in order to lower the cost to those
> > with limited initial resources. Sorry no Raspberry Pi’s for this project.
> > >
> > >
> > >
> > > There would be an on-premises version and a cloud version.  Guidance
> > would be provided on configuring the cluster, on demonstrating key
> > Cassandra behaviors, on files sizes, capacity to use with the Slender
> > Cassandra Cluster, and so on.
> > >
> > >
> > >
> > > Why about eighteen nodes? I tried to figure out what the minimum number
> > of nodes needed for Cassandra to be Cassandra is?  Here were my
> > considerations:
> > >
> > >
> > >
> > > •             A user wouldn’t run Cassandra in just one data center; so
> > at least two datacenters.
> > >
> > > •             A user probably would want a third data center available
> > for analytics.
> > >
> > > •             There needs to be enough nodes for enough parallelism to
> > observe Cassandra’s distributed nature.
> > >
> > > •             The cluster should have enough nodes that one gets a sense
> > of the need for cluster wide management tools to do things like repairs,
> > snapshots and cluster monitoring.
> > >
> > > •             The cluster should be able to demonstrate a RF=3 with
> > local quorum.  If replicated in all three data centers, one write would
> > impact half the 18 nodes, 3 datacenters X 3 nodes per data center = 9 nodes
> > of 18 nodes  If replicated in two of the data centers, one write would
> > still impact one third of the 18 nodes, 2 DC’s X 3 nodes per DC = 6 of the
> > 18 nodes.
> > >
> > >
> > >
> > > So eighteen seems like the minimum number of nodes needed.  That’s six
> > nodes in each of three data centers.
> > >
> > >
> > >
> > > Before I get too carried away with this project, I’m looking for some
> > feedback on whether this project would indeed be helpful to others? Also,
> > should the project be changed in any way?
> > >
> > >
> > >
> > > It’s always a pleasure to connect with the Cassandra users’ community.
> > Thanks for all the hard work, the expertise, the civil dialog.
> > >
> > >
> > >
> > > Kenneth Brotman
> > >
> > >
> > >
> >
> > ---------------------------------------------------------------------
> > To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org
> > For additional commands, e-mail: user-h...@cassandra.apache.org
> >
> >
> > ---------------------------------------------------------------------
> > To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org
> > For additional commands, e-mail: user-h...@cassandra.apache.org
> >
> >

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org
For additional commands, e-mail: user-h...@cassandra.apache.org

Reply via email to