[jira] [Updated] (CASSANDRA-14765) Evaluate Recovery Time on Single Token Cluster Test

2019-04-10 Thread Dinesh Joshi (JIRA)


 [ 
https://issues.apache.org/jira/browse/CASSANDRA-14765?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dinesh Joshi updated CASSANDRA-14765:
-
Resolution: Done
Status: Resolved  (was: Open)

> Evaluate Recovery Time on Single Token Cluster Test
> ---
>
> Key: CASSANDRA-14765
> URL: https://issues.apache.org/jira/browse/CASSANDRA-14765
> Project: Cassandra
>  Issue Type: Sub-task
>  Components: Legacy/Streaming and Messaging
>Reporter: Joseph Lynch
>Assignee: Sumanth Pasupuleti
>Priority: Normal
> Attachments: 3.0.17-4.0.x-Streaming.png, cassandra_streaming.png, 
> image-2018-11-06-13-34-33-108.png, streaming_benchmarking.patch
>
>
> *Setup:*
>  * Cassandra: 6 (2*3 rack) node i3.8xlarge AWS instance (32 cpu cores, 240GB 
> ram) running cassandra trunk with Jason's 14503 changes vs the same footprint 
> running 3.0.17
>  * One datacenter, single tokens
>  * No compression, encryption, or coalescing turned on
> *Test #1:*
> ndbench loaded ~150GB of data per node into a LCS table. Then we killed a 
> node and let a new node stream. With a single token this should be a worst 
> case recovery scenario (only  a few peers to stream from).
> *Result:*
> As the table used LCS and we didn't not have encryption on, the zero copy 
> transfer was used via CASSANDRA-14556. We recovered *150GB in 5 minutes,* 
> going at a consistent rate of about 3 gigabit per second. Theoretically we 
> should be able to get 10 gigabit, but this is still something like an 
> estimated 16x improvement over 3.0.x. We're still running the 3.0.x test for 
> a hard comparison.
> *Follow Ups:*
> We need to get more rigorous measurements (over more terminations), as well 
> as finishing the 3.0.x test. [~sumanth.pasupuleti] and [~djoshi3] are driving 
> this.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: commits-unsubscr...@cassandra.apache.org
For additional commands, e-mail: commits-h...@cassandra.apache.org



[jira] [Updated] (CASSANDRA-14765) Evaluate Recovery Time on Single Token Cluster Test

2019-04-09 Thread Sumanth Pasupuleti (JIRA)


 [ 
https://issues.apache.org/jira/browse/CASSANDRA-14765?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sumanth Pasupuleti updated CASSANDRA-14765:
---
Attachment: cassandra_streaming.png

> Evaluate Recovery Time on Single Token Cluster Test
> ---
>
> Key: CASSANDRA-14765
> URL: https://issues.apache.org/jira/browse/CASSANDRA-14765
> Project: Cassandra
>  Issue Type: Sub-task
>  Components: Legacy/Streaming and Messaging
>Reporter: Joseph Lynch
>Assignee: Sumanth Pasupuleti
>Priority: Normal
> Attachments: 3.0.17-4.0.x-Streaming.png, cassandra_streaming.png, 
> image-2018-11-06-13-34-33-108.png, streaming_benchmarking.patch
>
>
> *Setup:*
>  * Cassandra: 6 (2*3 rack) node i3.8xlarge AWS instance (32 cpu cores, 240GB 
> ram) running cassandra trunk with Jason's 14503 changes vs the same footprint 
> running 3.0.17
>  * One datacenter, single tokens
>  * No compression, encryption, or coalescing turned on
> *Test #1:*
> ndbench loaded ~150GB of data per node into a LCS table. Then we killed a 
> node and let a new node stream. With a single token this should be a worst 
> case recovery scenario (only  a few peers to stream from).
> *Result:*
> As the table used LCS and we didn't not have encryption on, the zero copy 
> transfer was used via CASSANDRA-14556. We recovered *150GB in 5 minutes,* 
> going at a consistent rate of about 3 gigabit per second. Theoretically we 
> should be able to get 10 gigabit, but this is still something like an 
> estimated 16x improvement over 3.0.x. We're still running the 3.0.x test for 
> a hard comparison.
> *Follow Ups:*
> We need to get more rigorous measurements (over more terminations), as well 
> as finishing the 3.0.x test. [~sumanth.pasupuleti] and [~djoshi3] are driving 
> this.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: commits-unsubscr...@cassandra.apache.org
For additional commands, e-mail: commits-h...@cassandra.apache.org



[jira] [Updated] (CASSANDRA-14765) Evaluate Recovery Time on Single Token Cluster Test

2019-04-09 Thread Sumanth Pasupuleti (JIRA)


 [ 
https://issues.apache.org/jira/browse/CASSANDRA-14765?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sumanth Pasupuleti updated CASSANDRA-14765:
---
Attachment: streaming_benchmarking.patch

> Evaluate Recovery Time on Single Token Cluster Test
> ---
>
> Key: CASSANDRA-14765
> URL: https://issues.apache.org/jira/browse/CASSANDRA-14765
> Project: Cassandra
>  Issue Type: Sub-task
>  Components: Legacy/Streaming and Messaging
>Reporter: Joseph Lynch
>Assignee: Sumanth Pasupuleti
>Priority: Normal
> Attachments: 3.0.17-4.0.x-Streaming.png, 
> image-2018-11-06-13-34-33-108.png, streaming_benchmarking.patch
>
>
> *Setup:*
>  * Cassandra: 6 (2*3 rack) node i3.8xlarge AWS instance (32 cpu cores, 240GB 
> ram) running cassandra trunk with Jason's 14503 changes vs the same footprint 
> running 3.0.17
>  * One datacenter, single tokens
>  * No compression, encryption, or coalescing turned on
> *Test #1:*
> ndbench loaded ~150GB of data per node into a LCS table. Then we killed a 
> node and let a new node stream. With a single token this should be a worst 
> case recovery scenario (only  a few peers to stream from).
> *Result:*
> As the table used LCS and we didn't not have encryption on, the zero copy 
> transfer was used via CASSANDRA-14556. We recovered *150GB in 5 minutes,* 
> going at a consistent rate of about 3 gigabit per second. Theoretically we 
> should be able to get 10 gigabit, but this is still something like an 
> estimated 16x improvement over 3.0.x. We're still running the 3.0.x test for 
> a hard comparison.
> *Follow Ups:*
> We need to get more rigorous measurements (over more terminations), as well 
> as finishing the 3.0.x test. [~sumanth.pasupuleti] and [~djoshi3] are driving 
> this.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: commits-unsubscr...@cassandra.apache.org
For additional commands, e-mail: commits-h...@cassandra.apache.org



[jira] [Updated] (CASSANDRA-14765) Evaluate Recovery Time on Single Token Cluster Test

2018-11-15 Thread C. Scott Andreas (JIRA)


 [ 
https://issues.apache.org/jira/browse/CASSANDRA-14765?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

C. Scott Andreas updated CASSANDRA-14765:
-
Component/s: Streaming and Messaging

> Evaluate Recovery Time on Single Token Cluster Test
> ---
>
> Key: CASSANDRA-14765
> URL: https://issues.apache.org/jira/browse/CASSANDRA-14765
> Project: Cassandra
>  Issue Type: Sub-task
>  Components: Streaming and Messaging
>Reporter: Joseph Lynch
>Assignee: Sumanth Pasupuleti
>Priority: Major
> Attachments: 3.0.17-4.0.x-Streaming.png, 
> image-2018-11-06-13-34-33-108.png
>
>
> *Setup:*
>  * Cassandra: 6 (2*3 rack) node i3.8xlarge AWS instance (32 cpu cores, 240GB 
> ram) running cassandra trunk with Jason's 14503 changes vs the same footprint 
> running 3.0.17
>  * One datacenter, single tokens
>  * No compression, encryption, or coalescing turned on
> *Test #1:*
> ndbench loaded ~150GB of data per node into a LCS table. Then we killed a 
> node and let a new node stream. With a single token this should be a worst 
> case recovery scenario (only  a few peers to stream from).
> *Result:*
> As the table used LCS and we didn't not have encryption on, the zero copy 
> transfer was used via CASSANDRA-14556. We recovered *150GB in 5 minutes,* 
> going at a consistent rate of about 3 gigabit per second. Theoretically we 
> should be able to get 10 gigabit, but this is still something like an 
> estimated 16x improvement over 3.0.x. We're still running the 3.0.x test for 
> a hard comparison.
> *Follow Ups:*
> We need to get more rigorous measurements (over more terminations), as well 
> as finishing the 3.0.x test. [~sumanth.pasupuleti] and [~djoshi3] are driving 
> this.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: commits-unsubscr...@cassandra.apache.org
For additional commands, e-mail: commits-h...@cassandra.apache.org



[jira] [Updated] (CASSANDRA-14765) Evaluate Recovery Time on Single Token Cluster Test

2018-11-06 Thread Joseph Lynch (JIRA)


 [ 
https://issues.apache.org/jira/browse/CASSANDRA-14765?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Joseph Lynch updated CASSANDRA-14765:
-
Attachment: 3.0.17-4.0.x-Streaming.png

> Evaluate Recovery Time on Single Token Cluster Test
> ---
>
> Key: CASSANDRA-14765
> URL: https://issues.apache.org/jira/browse/CASSANDRA-14765
> Project: Cassandra
>  Issue Type: Sub-task
>Reporter: Joseph Lynch
>Assignee: Sumanth Pasupuleti
>Priority: Major
> Attachments: 3.0.17-4.0.x-Streaming.png, 
> image-2018-11-06-13-34-33-108.png
>
>
> *Setup:*
>  * Cassandra: 6 (2*3 rack) node i3.8xlarge AWS instance (32 cpu cores, 240GB 
> ram) running cassandra trunk with Jason's 14503 changes vs the same footprint 
> running 3.0.17
>  * One datacenter, single tokens
>  * No compression, encryption, or coalescing turned on
> *Test #1:*
> ndbench loaded ~150GB of data per node into a LCS table. Then we killed a 
> node and let a new node stream. With a single token this should be a worst 
> case recovery scenario (only  a few peers to stream from).
> *Result:*
> As the table used LCS and we didn't not have encryption on, the zero copy 
> transfer was used via CASSANDRA-14556. We recovered *150GB in 5 minutes,* 
> going at a consistent rate of about 3 gigabit per second. Theoretically we 
> should be able to get 10 gigabit, but this is still something like an 
> estimated 16x improvement over 3.0.x. We're still running the 3.0.x test for 
> a hard comparison.
> *Follow Ups:*
> We need to get more rigorous measurements (over more terminations), as well 
> as finishing the 3.0.x test. [~sumanth.pasupuleti] and [~djoshi3] are driving 
> this.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: commits-unsubscr...@cassandra.apache.org
For additional commands, e-mail: commits-h...@cassandra.apache.org



[jira] [Updated] (CASSANDRA-14765) Evaluate Recovery Time on Single Token Cluster Test

2018-11-06 Thread Joseph Lynch (JIRA)


 [ 
https://issues.apache.org/jira/browse/CASSANDRA-14765?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Joseph Lynch updated CASSANDRA-14765:
-
Attachment: image-2018-11-06-13-34-33-108.png

> Evaluate Recovery Time on Single Token Cluster Test
> ---
>
> Key: CASSANDRA-14765
> URL: https://issues.apache.org/jira/browse/CASSANDRA-14765
> Project: Cassandra
>  Issue Type: Sub-task
>Reporter: Joseph Lynch
>Assignee: Sumanth Pasupuleti
>Priority: Major
> Attachments: image-2018-11-06-13-34-33-108.png
>
>
> *Setup:*
>  * Cassandra: 6 (2*3 rack) node i3.8xlarge AWS instance (32 cpu cores, 240GB 
> ram) running cassandra trunk with Jason's 14503 changes vs the same footprint 
> running 3.0.17
>  * One datacenter, single tokens
>  * No compression, encryption, or coalescing turned on
> *Test #1:*
> ndbench loaded ~150GB of data per node into a LCS table. Then we killed a 
> node and let a new node stream. With a single token this should be a worst 
> case recovery scenario (only  a few peers to stream from).
> *Result:*
> As the table used LCS and we didn't not have encryption on, the zero copy 
> transfer was used via CASSANDRA-14556. We recovered *150GB in 5 minutes,* 
> going at a consistent rate of about 3 gigabit per second. Theoretically we 
> should be able to get 10 gigabit, but this is still something like an 
> estimated 16x improvement over 3.0.x. We're still running the 3.0.x test for 
> a hard comparison.
> *Follow Ups:*
> We need to get more rigorous measurements (over more terminations), as well 
> as finishing the 3.0.x test. [~sumanth.pasupuleti] and [~djoshi3] are driving 
> this.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: commits-unsubscr...@cassandra.apache.org
For additional commands, e-mail: commits-h...@cassandra.apache.org



[jira] [Updated] (CASSANDRA-14765) Evaluate Recovery Time on Single Token Cluster Test

2018-09-19 Thread Dinesh Joshi (JIRA)


 [ 
https://issues.apache.org/jira/browse/CASSANDRA-14765?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dinesh Joshi updated CASSANDRA-14765:
-
Description: 
*Setup:*
 * Cassandra: 6 (2*3 rack) node i3.8xlarge AWS instance (32 cpu cores, 240GB 
ram) running cassandra trunk with Jason's 14503 changes vs the same footprint 
running 3.0.17
 * One datacenter, single tokens
 * No compression, encryption, or coalescing turned on

*Test #1:*

ndbench loaded ~150GB of data per node into a LCS table. Then we killed a node 
and let a new node stream. With a single token this should be a worst case 
recovery scenario (only  a few peers to stream from).

*Result:*

As the table used LCS and we didn't not have encryption on, the zero copy 
transfer was used via CASSANDRA-14556. We recovered *150GB in 5 minutes,* going 
at a consistent rate of about 3 gigabit per second. Theoretically we should be 
able to get 10 gigabit, but this is still something like an estimated 16x 
improvement over 3.0.x. We're still running the 3.0.x test for a hard 
comparison.

*Follow Ups:*
We need to get more rigorous measurements (over more terminations), as well as 
finishing the 3.0.x test. [~sumanth.pasupuleti] and [~djoshi3] are driving this.

  was:
*Setup:*
 * Cassandra: 6 (2*3 rack) node i3.8xlarge AWS instance (32 cpu cores, 240GB 
ram) running cassandra trunk with Jason's 14503 changes vs the same footprint 
running 3.0.17
 * One datacenter, single tokens
 * No compression, encryption, or coalescing turned on

*Test #1:*

ndbench loaded ~150GB of data per node into a LCS table. Then we killed a node 
and let a new node stream. With a single token this should be a worst case 
recovery scenario (only  a few peers to stream from).

*Result:*

As the table used LCS and we didn't not have encryption on, the zero copy 
transfer was used via CASSANDRA-14556. We recovered *150GB in 5 minutes,* going 
at a consistent rate of about 3 gigabit per second. Theoretically we should be 
able to get 10 gigabit, but this is still something like an estimated 16x 
improvement over 3.0.x. We're still running the 3.0.x test for a hard 
comparison.

*Follow Ups:*
We need to get more rigorous measurements (over more terminations), as well as 
finishing the 3.0.x test. [~sumanth.pasupuleti] and [~djoshi] are driving this.


> Evaluate Recovery Time on Single Token Cluster Test
> ---
>
> Key: CASSANDRA-14765
> URL: https://issues.apache.org/jira/browse/CASSANDRA-14765
> Project: Cassandra
>  Issue Type: Sub-task
>Reporter: Joseph Lynch
>Priority: Major
>
> *Setup:*
>  * Cassandra: 6 (2*3 rack) node i3.8xlarge AWS instance (32 cpu cores, 240GB 
> ram) running cassandra trunk with Jason's 14503 changes vs the same footprint 
> running 3.0.17
>  * One datacenter, single tokens
>  * No compression, encryption, or coalescing turned on
> *Test #1:*
> ndbench loaded ~150GB of data per node into a LCS table. Then we killed a 
> node and let a new node stream. With a single token this should be a worst 
> case recovery scenario (only  a few peers to stream from).
> *Result:*
> As the table used LCS and we didn't not have encryption on, the zero copy 
> transfer was used via CASSANDRA-14556. We recovered *150GB in 5 minutes,* 
> going at a consistent rate of about 3 gigabit per second. Theoretically we 
> should be able to get 10 gigabit, but this is still something like an 
> estimated 16x improvement over 3.0.x. We're still running the 3.0.x test for 
> a hard comparison.
> *Follow Ups:*
> We need to get more rigorous measurements (over more terminations), as well 
> as finishing the 3.0.x test. [~sumanth.pasupuleti] and [~djoshi3] are driving 
> this.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: commits-unsubscr...@cassandra.apache.org
For additional commands, e-mail: commits-h...@cassandra.apache.org