Yes, but getting at why you think EBS is going down is the real point. New GM in 2011. Very different product. 35:40 in the video
-- Jeff Jirsa > On Jan 31, 2016, at 9:57 PM, Eric Plowe <eric.pl...@gmail.com> wrote: > > Jeff, > > If EBS goes down, then EBS Gp2 will go down as well, no? I'm not discounting > EBS, but prior outages are worrisome. > >> On Sunday, January 31, 2016, Jeff Jirsa <jeff.ji...@crowdstrike.com> wrote: >> Free to choose what you'd like, but EBS outages were also addressed in that >> video (second half, discussion by Dennis Opacki). 2016 EBS isn't the same as >> 2011 EBS. >> >> -- >> Jeff Jirsa >> >> >>> On Jan 31, 2016, at 8:27 PM, Eric Plowe <eric.pl...@gmail.com> wrote: >>> >>> Thank you all for the suggestions. I'm torn between GP2 vs Ephemeral. GP2 >>> after testing is a viable contender for our workload. The only worry I have >>> is EBS outages, which have happened. >>> >>>> On Sunday, January 31, 2016, Jeff Jirsa <jeff.ji...@crowdstrike.com> wrote: >>>> Also in that video - it's long but worth watching >>>> >>>> We tested up to 1M reads/second as well, blowing out page cache to ensure >>>> we weren't "just" reading from memory >>>> >>>> >>>> >>>> -- >>>> Jeff Jirsa >>>> >>>> >>>>> On Jan 31, 2016, at 9:52 AM, Jack Krupansky <jack.krupan...@gmail.com> >>>>> wrote: >>>>> >>>>> How about reads? Any differences between read-intensive and >>>>> write-intensive workloads? >>>>> >>>>> -- Jack Krupansky >>>>> >>>>>> On Sun, Jan 31, 2016 at 3:13 AM, Jeff Jirsa <jeff.ji...@crowdstrike.com> >>>>>> wrote: >>>>>> Hi John, >>>>>> >>>>>> We run using 4T GP2 volumes, which guarantee 10k iops. Even at 1M writes >>>>>> per second on 60 nodes, we didn’t come close to hitting even 50% >>>>>> utilization (10k is more than enough for most workloads). PIOPS is not >>>>>> necessary. >>>>>> >>>>>> >>>>>> >>>>>> From: John Wong >>>>>> Reply-To: "user@cassandra.apache.org" >>>>>> Date: Saturday, January 30, 2016 at 3:07 PM >>>>>> To: "user@cassandra.apache.org" >>>>>> Subject: Re: EC2 storage options for C* >>>>>> >>>>>> For production I'd stick with ephemeral disks (aka instance storage) if >>>>>> you have running a lot of transaction. >>>>>> However, for regular small testing/qa cluster, or something you know you >>>>>> want to reload often, EBS is definitely good enough and we haven't had >>>>>> issues 99%. The 1% is kind of anomaly where we have flush blocked. >>>>>> >>>>>> But Jeff, kudo that you are able to use EBS. I didn't go through the >>>>>> video, do you actually use PIOPS or just standard GP2 in your production >>>>>> cluster? >>>>>> >>>>>>> On Sat, Jan 30, 2016 at 1:28 PM, Bryan Cheng <br...@blockcypher.com> >>>>>>> wrote: >>>>>>> Yep, that motivated my question "Do you have any idea what kind of disk >>>>>>> performance you need?". If you need the performance, its hard to beat >>>>>>> ephemeral SSD in RAID 0 on EC2, and its a solid, battle tested >>>>>>> configuration. If you don't, though, EBS GP2 will save a _lot_ of >>>>>>> headache. >>>>>>> >>>>>>> Personally, on small clusters like ours (12 nodes), we've found our >>>>>>> choice of instance dictated much more by the balance of price, CPU, and >>>>>>> memory. We're using GP2 SSD and we find that for our patterns the disk >>>>>>> is rarely the bottleneck. YMMV, of course. >>>>>>> >>>>>>>> On Fri, Jan 29, 2016 at 7:32 PM, Jeff Jirsa >>>>>>>> <jeff.ji...@crowdstrike.com> wrote: >>>>>>>> If you have to ask that question, I strongly recommend m4 or c4 >>>>>>>> instances with GP2 EBS. When you don’t care about replacing a node >>>>>>>> because of an instance failure, go with i2+ephemerals. Until then, GP2 >>>>>>>> EBS is capable of amazing things, and greatly simplifies life. >>>>>>>> >>>>>>>> We gave a talk on this topic at both Cassandra Summit and AWS >>>>>>>> re:Invent: https://www.youtube.com/watch?v=1R-mgOcOSd4 It’s very much >>>>>>>> a viable option, despite any old documents online that say otherwise. >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> From: Eric Plowe >>>>>>>> Reply-To: "user@cassandra.apache.org" >>>>>>>> Date: Friday, January 29, 2016 at 4:33 PM >>>>>>>> To: "user@cassandra.apache.org" >>>>>>>> Subject: EC2 storage options for C* >>>>>>>> >>>>>>>> My company is planning on rolling out a C* cluster in EC2. We are >>>>>>>> thinking about going with ephemeral SSDs. The question is this: Should >>>>>>>> we put two in RAID 0 or just go with one? We currently run a cluster >>>>>>>> in our data center with 2 250gig Samsung 850 EVO's in RAID 0 and we >>>>>>>> are happy with the performance we are seeing thus far. >>>>>>>> >>>>>>>> Thanks! >>>>>>>> >>>>>>>> Eric
smime.p7s
Description: S/MIME cryptographic signature