Yes, I'd prefer Terraform too. On 2018-01-31 06:32:21, Michael Mior <mm...@uwaterloo.ca> wrote: > While whatever format this comes out in would be helpful, you might want to > consider Terraform. 1Password recently published a blog post on their > experience with Terraform vs. CloudFormation. > > https://blog.agilebits.com/2018/01/25/terraforming-1password/ > > -- > Michael Mior > mm...@apache.org > > 2018-01-31 2:34 GMT-05:00 Kenneth Brotman <kenbrot...@yahoo.com.invalid>: > > > Hi Yuri, > > > > If possible I will do everything with AWS Cloudformation. I'm working on > > it now. Nothing published yet. > > > > Kenneth Brotman > > > > -----Original Message----- > > From: Yuri Subach [mailto:ysub...@gmail.com] > > Sent: Tuesday, January 30, 2018 7:02 PM > > To: user@cassandra.apache.org > > Subject: RE: Slender Cassandra Cluster Project > > > > Hi Kenneth, > > > > I like this project idea! > > > > A couple of questions: > > - What tools are you going to use for AWS cluster setup? > > - Do you have anything published already (github)? > > > > On 2018-01-22 22:42:11, Kenneth Brotman <kenbrot...@yahoo.com.INVALID> > > wrote: > > > Thanks Anthony! I’ve made a note to include that information in the > > documentation. You’re right. It won’t work as intended unless that is > > configured properly. > > > > > > > > > > > > I’m also favoring a couple other guidelines for Slender Cassandra: > > > > > > 1. SSD’s only, no spinning disks > > > > > > 2. At least two cores per node > > > > > > > > > > > > For AWS, I’m favoring the c3.large on Linux. It’s available in these > > regions: US-East, US-West and US-West2. The specifications are listed as: > > > > > > · Two (2) vCPU’s > > > > > > · 3.7 Gib Memory > > > > > > · Two (2) 16 GB SSD’s > > > > > > · Moderate I/O > > > > > > > > > > > > It’s going to be hard to beat the inexpensive cost of operating a > > Slender Cluster on demand in the cloud – and it fits a lot of the use cases > > well: > > > > > > > > > > > > · For under a $100 a month, in current pricing for EC2 > > instances, you can operate an eighteen (18) node Slender Cluster for five > > (5) hours a day, ten (10) days a month. That’s fine for demonstrations, > > teaching or experiments that last half a day or less. > > > > > > · For under $20, you can have that Slender Cluster up all day > > long, up to ten (10) hours, for whatever demonstrations or experiments you > > want it for. > > > > > > > > > > > > As always, feedback is encouraged. > > > > > > > > > > > > Thanks, > > > > > > > > > > > > Kenneth Brotman > > > > > > > > > > > > From: Anthony Grasso [mailto:anthony.gra...@gmail.com] > > > Sent: Sunday, January 21, 2018 3:57 PM > > > To: user > > > Subject: Re: Slender Cassandra Cluster Project > > > > > > > > > > > > Hi Kenneth, > > > > > > > > > > > > Fantastic idea! > > > > > > > > > > > > One thing that came to mind from my reading of the proposed setup was > > rack awareness of each node. Given that the proposed setup contains three > > DCs, I assume that each node will be made rack aware? If not, consider > > defining three racks for each DC and placing two nodes in each rack. This > > will ensure that all the nodes in a single rack contain at most one replica > > of the data. > > > > > > > > > > > > Regards, > > > > > > Anthony > > > > > > > > > > > > On 17 January 2018 at 11:24, Kenneth Brotman > > <kenbrot...@yahoo.com.invalid> wrote: > > > > > > Sure. That takes the project from awesome to 10X awesome. I absolutely > > would be willing to do that. Thanks Kurt! > > > > > > > > > > > > Regarding your comment on the keyspaces, I agree. There should be a few > > simple examples one way or the other that can be duplicated and observed, > > and then an example to duplicate and play with that has a nice real world > > mix, with some keyspaces that replicate over only a subset of DC’s and some > > that replicate to all DC’s. > > > > > > > > > > > > Kenneth Brotman > > > > > > > > > > > > From: kurt greaves [mailto:k...@instaclustr.com] > > > Sent: Tuesday, January 16, 2018 1:31 PM > > > To: User > > > Subject: Re: Slender Cassandra Cluster Project > > > > > > > > > > > > Sounds like a great idea. Probably would be valuable to add to the > > official docs as an example set up if you're willing. > > > > > > > > > > > > Only thing I'd add is that you should have keyspaces that replicate over > > only a subset of DC's, plus one/some replicated to all DC's > > > > > > > > > > > > On 17 Jan. 2018 03:26, "Kenneth Brotman" <kenbrot...@yahoo.com.invalid> > > wrote: > > > > > > I’ve begun working on a reference project intended to provide guidance > > on configuring and operating a modest Cassandra cluster of about 18 nodes > > suitable for the economic study, demonstration, experimentation and testing > > of a Cassandra cluster. > > > > > > > > > > > > The slender cluster would be designed to be as inexpensive as possible > > while still using real world hardware in order to lower the cost to those > > with limited initial resources. Sorry no Raspberry Pi’s for this project. > > > > > > > > > > > > There would be an on-premises version and a cloud version. Guidance > > would be provided on configuring the cluster, on demonstrating key > > Cassandra behaviors, on files sizes, capacity to use with the Slender > > Cassandra Cluster, and so on. > > > > > > > > > > > > Why about eighteen nodes? I tried to figure out what the minimum number > > of nodes needed for Cassandra to be Cassandra is? Here were my > > considerations: > > > > > > > > > > > > • A user wouldn’t run Cassandra in just one data center; so > > at least two datacenters. > > > > > > • A user probably would want a third data center available > > for analytics. > > > > > > • There needs to be enough nodes for enough parallelism to > > observe Cassandra’s distributed nature. > > > > > > • The cluster should have enough nodes that one gets a sense > > of the need for cluster wide management tools to do things like repairs, > > snapshots and cluster monitoring. > > > > > > • The cluster should be able to demonstrate a RF=3 with > > local quorum. If replicated in all three data centers, one write would > > impact half the 18 nodes, 3 datacenters X 3 nodes per data center = 9 nodes > > of 18 nodes If replicated in two of the data centers, one write would > > still impact one third of the 18 nodes, 2 DC’s X 3 nodes per DC = 6 of the > > 18 nodes. > > > > > > > > > > > > So eighteen seems like the minimum number of nodes needed. That’s six > > nodes in each of three data centers. > > > > > > > > > > > > Before I get too carried away with this project, I’m looking for some > > feedback on whether this project would indeed be helpful to others? Also, > > should the project be changed in any way? > > > > > > > > > > > > It’s always a pleasure to connect with the Cassandra users’ community. > > Thanks for all the hard work, the expertise, the civil dialog. > > > > > > > > > > > > Kenneth Brotman > > > > > > > > > > > > > --------------------------------------------------------------------- > > To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org > > For additional commands, e-mail: user-h...@cassandra.apache.org > > > > > > --------------------------------------------------------------------- > > To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org > > For additional commands, e-mail: user-h...@cassandra.apache.org > > > >
--------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org For additional commands, e-mail: user-h...@cassandra.apache.org