While whatever format this comes out in would be helpful, you might want to
consider Terraform. 1Password recently published a blog post on their
experience with Terraform vs. CloudFormation.

https://blog.agilebits.com/2018/01/25/terraforming-1password/

--
Michael Mior
mm...@apache.org

2018-01-31 2:34 GMT-05:00 Kenneth Brotman <kenbrot...@yahoo.com.invalid>:

> Hi Yuri,
>
> If possible I will do everything with AWS Cloudformation.  I'm working on
> it now.  Nothing published yet.
>
> Kenneth Brotman
>
> -----Original Message-----
> From: Yuri Subach [mailto:ysub...@gmail.com]
> Sent: Tuesday, January 30, 2018 7:02 PM
> To: user@cassandra.apache.org
> Subject: RE: Slender Cassandra Cluster Project
>
> Hi Kenneth,
>
> I like this project idea!
>
> A couple of questions:
> - What tools are you going to use for AWS cluster setup?
> - Do you have anything published already (github)?
>
> On 2018-01-22 22:42:11, Kenneth Brotman <kenbrot...@yahoo.com.INVALID>
> wrote:
> > Thanks Anthony!  I’ve made a note to include that information in the
> documentation. You’re right.  It won’t work as intended unless that is
> configured properly.
> >
> >
> >
> > I’m also favoring a couple other guidelines for Slender Cassandra:
> >
> > 1.       SSD’s only, no spinning disks
> >
> > 2.       At least two cores per node
> >
> >
> >
> > For AWS, I’m favoring the c3.large on Linux.  It’s available in these
> regions: US-East, US-West and US-West2.  The specifications are listed as:
> >
> > ·         Two (2) vCPU’s
> >
> > ·         3.7 Gib Memory
> >
> > ·         Two (2) 16 GB SSD’s
> >
> > ·         Moderate I/O
> >
> >
> >
> > It’s going to be hard to beat the inexpensive cost of operating a
> Slender Cluster on demand in the cloud – and it fits a lot of the use cases
> well:
> >
> >
> >
> > ·         For under a $100 a month, in current pricing for EC2
> instances, you can operate an eighteen (18) node Slender Cluster for five
> (5) hours a day, ten (10) days a month.  That’s fine for demonstrations,
> teaching or experiments that last half a day or less.
> >
> > ·         For under $20, you can have that Slender Cluster up all day
> long, up to ten (10) hours, for whatever demonstrations or experiments you
> want it for.
> >
> >
> >
> > As always, feedback is encouraged.
> >
> >
> >
> > Thanks,
> >
> >
> >
> > Kenneth Brotman
> >
> >
> >
> > From: Anthony Grasso [mailto:anthony.gra...@gmail.com]
> > Sent: Sunday, January 21, 2018 3:57 PM
> > To: user
> > Subject: Re: Slender Cassandra Cluster Project
> >
> >
> >
> > Hi Kenneth,
> >
> >
> >
> > Fantastic idea!
> >
> >
> >
> > One thing that came to mind from my reading of the proposed setup was
> rack awareness of each node. Given that the proposed setup contains three
> DCs, I assume that each node will be made rack aware? If not, consider
> defining three racks for each DC and placing two nodes in each rack. This
> will ensure that all the nodes in a single rack contain at most one replica
> of the data.
> >
> >
> >
> > Regards,
> >
> > Anthony
> >
> >
> >
> > On 17 January 2018 at 11:24, Kenneth Brotman
> <kenbrot...@yahoo.com.invalid> wrote:
> >
> > Sure.  That takes the project from awesome to 10X awesome.  I absolutely
> would be willing to do that.  Thanks Kurt!
> >
> >
> >
> > Regarding your comment on the keyspaces, I agree.  There should be a few
> simple examples one way or the other that can be duplicated and observed,
> and then an example to duplicate and play with that has a nice real world
> mix, with some keyspaces that replicate over only a subset of DC’s and some
> that replicate to all DC’s.
> >
> >
> >
> > Kenneth Brotman
> >
> >
> >
> > From: kurt greaves [mailto:k...@instaclustr.com]
> > Sent: Tuesday, January 16, 2018 1:31 PM
> > To: User
> > Subject: Re: Slender Cassandra Cluster Project
> >
> >
> >
> > Sounds like a great idea. Probably would be valuable to add to the
> official docs as an example set up if you're willing.
> >
> >
> >
> > Only thing I'd add is that you should have keyspaces that replicate over
> only a subset of DC's, plus one/some replicated to all DC's
> >
> >
> >
> > On 17 Jan. 2018 03:26, "Kenneth Brotman" <kenbrot...@yahoo.com.invalid>
> wrote:
> >
> > I’ve begun working on a reference project intended to provide guidance
> on configuring and operating a modest Cassandra cluster of about 18 nodes
> suitable for the economic study, demonstration, experimentation and testing
> of a Cassandra cluster.
> >
> >
> >
> > The slender cluster would be designed to be as inexpensive as possible
> while still using real world hardware in order to lower the cost to those
> with limited initial resources. Sorry no Raspberry Pi’s for this project.
> >
> >
> >
> > There would be an on-premises version and a cloud version.  Guidance
> would be provided on configuring the cluster, on demonstrating key
> Cassandra behaviors, on files sizes, capacity to use with the Slender
> Cassandra Cluster, and so on.
> >
> >
> >
> > Why about eighteen nodes? I tried to figure out what the minimum number
> of nodes needed for Cassandra to be Cassandra is?  Here were my
> considerations:
> >
> >
> >
> > •             A user wouldn’t run Cassandra in just one data center; so
> at least two datacenters.
> >
> > •             A user probably would want a third data center available
> for analytics.
> >
> > •             There needs to be enough nodes for enough parallelism to
> observe Cassandra’s distributed nature.
> >
> > •             The cluster should have enough nodes that one gets a sense
> of the need for cluster wide management tools to do things like repairs,
> snapshots and cluster monitoring.
> >
> > •             The cluster should be able to demonstrate a RF=3 with
> local quorum.  If replicated in all three data centers, one write would
> impact half the 18 nodes, 3 datacenters X 3 nodes per data center = 9 nodes
> of 18 nodes  If replicated in two of the data centers, one write would
> still impact one third of the 18 nodes, 2 DC’s X 3 nodes per DC = 6 of the
> 18 nodes.
> >
> >
> >
> > So eighteen seems like the minimum number of nodes needed.  That’s six
> nodes in each of three data centers.
> >
> >
> >
> > Before I get too carried away with this project, I’m looking for some
> feedback on whether this project would indeed be helpful to others? Also,
> should the project be changed in any way?
> >
> >
> >
> > It’s always a pleasure to connect with the Cassandra users’ community.
> Thanks for all the hard work, the expertise, the civil dialog.
> >
> >
> >
> > Kenneth Brotman
> >
> >
> >
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org
> For additional commands, e-mail: user-h...@cassandra.apache.org
>
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org
> For additional commands, e-mail: user-h...@cassandra.apache.org
>
>

Reply via email to