git commit: Revert remove decompression code in CompressedSequentialWriter
Updated Branches: refs/heads/cassandra-2.0 cfa0e580e - 232906d7e Revert remove decompression code in CompressedSequentialWriter This reverts commit 9644f098967d6ba9154725148a383b3502ef62c2. Project: http://git-wip-us.apache.org/repos/asf/cassandra/repo Commit: http://git-wip-us.apache.org/repos/asf/cassandra/commit/232906d7 Tree: http://git-wip-us.apache.org/repos/asf/cassandra/tree/232906d7 Diff: http://git-wip-us.apache.org/repos/asf/cassandra/diff/232906d7 Branch: refs/heads/cassandra-2.0 Commit: 232906d7e01587c32739c61cd72079715c0abb3c Parents: cfa0e58 Author: Marcus Eriksson marc...@spotify.com Authored: Sat Oct 19 08:48:43 2013 +0200 Committer: Marcus Eriksson marc...@spotify.com Committed: Sat Oct 19 08:50:53 2013 +0200 -- .../io/compress/CompressedSequentialWriter.java | 12 1 file changed, 12 insertions(+) -- http://git-wip-us.apache.org/repos/asf/cassandra/blob/232906d7/src/java/org/apache/cassandra/io/compress/CompressedSequentialWriter.java -- diff --git a/src/java/org/apache/cassandra/io/compress/CompressedSequentialWriter.java b/src/java/org/apache/cassandra/io/compress/CompressedSequentialWriter.java index da4307e..54b990f 100644 --- a/src/java/org/apache/cassandra/io/compress/CompressedSequentialWriter.java +++ b/src/java/org/apache/cassandra/io/compress/CompressedSequentialWriter.java @@ -21,6 +21,7 @@ import java.io.EOFException; import java.io.File; import java.io.IOException; import java.util.zip.Adler32; +import java.util.zip.CRC32; import java.util.zip.Checksum; import org.apache.cassandra.io.FSReadError; @@ -192,6 +193,17 @@ public class CompressedSequentialWriter extends SequentialWriter { out.seek(chunkOffset); out.readFully(compressed.buffer, 0, chunkSize); + +try +{ +// repopulate buffer +compressor.uncompress(compressed.buffer, 0, chunkSize, buffer, 0); +} +catch (IOException e) +{ +throw new CorruptBlockException(getPath(), chunkOffset, chunkSize); +} + checksum.update(compressed.buffer, 0, chunkSize); if (out.readInt() != (int) checksum.getValue())
[2/2] git commit: Merge branch 'cassandra-2.0' into trunk
Merge branch 'cassandra-2.0' into trunk Project: http://git-wip-us.apache.org/repos/asf/cassandra/repo Commit: http://git-wip-us.apache.org/repos/asf/cassandra/commit/656d3240 Tree: http://git-wip-us.apache.org/repos/asf/cassandra/tree/656d3240 Diff: http://git-wip-us.apache.org/repos/asf/cassandra/diff/656d3240 Branch: refs/heads/trunk Commit: 656d3240f7488be7cbc05fab7e9dd45616c5cd27 Parents: 3020555 232906d Author: Marcus Eriksson marc...@spotify.com Authored: Sat Oct 19 08:51:54 2013 +0200 Committer: Marcus Eriksson marc...@spotify.com Committed: Sat Oct 19 08:51:54 2013 +0200 -- .../io/compress/CompressedSequentialWriter.java | 12 1 file changed, 12 insertions(+) --
[1/2] git commit: Revert remove decompression code in CompressedSequentialWriter
Updated Branches: refs/heads/trunk 302055583 - 656d3240f Revert remove decompression code in CompressedSequentialWriter This reverts commit 9644f098967d6ba9154725148a383b3502ef62c2. Project: http://git-wip-us.apache.org/repos/asf/cassandra/repo Commit: http://git-wip-us.apache.org/repos/asf/cassandra/commit/232906d7 Tree: http://git-wip-us.apache.org/repos/asf/cassandra/tree/232906d7 Diff: http://git-wip-us.apache.org/repos/asf/cassandra/diff/232906d7 Branch: refs/heads/trunk Commit: 232906d7e01587c32739c61cd72079715c0abb3c Parents: cfa0e58 Author: Marcus Eriksson marc...@spotify.com Authored: Sat Oct 19 08:48:43 2013 +0200 Committer: Marcus Eriksson marc...@spotify.com Committed: Sat Oct 19 08:50:53 2013 +0200 -- .../io/compress/CompressedSequentialWriter.java | 12 1 file changed, 12 insertions(+) -- http://git-wip-us.apache.org/repos/asf/cassandra/blob/232906d7/src/java/org/apache/cassandra/io/compress/CompressedSequentialWriter.java -- diff --git a/src/java/org/apache/cassandra/io/compress/CompressedSequentialWriter.java b/src/java/org/apache/cassandra/io/compress/CompressedSequentialWriter.java index da4307e..54b990f 100644 --- a/src/java/org/apache/cassandra/io/compress/CompressedSequentialWriter.java +++ b/src/java/org/apache/cassandra/io/compress/CompressedSequentialWriter.java @@ -21,6 +21,7 @@ import java.io.EOFException; import java.io.File; import java.io.IOException; import java.util.zip.Adler32; +import java.util.zip.CRC32; import java.util.zip.Checksum; import org.apache.cassandra.io.FSReadError; @@ -192,6 +193,17 @@ public class CompressedSequentialWriter extends SequentialWriter { out.seek(chunkOffset); out.readFully(compressed.buffer, 0, chunkSize); + +try +{ +// repopulate buffer +compressor.uncompress(compressed.buffer, 0, chunkSize, buffer, 0); +} +catch (IOException e) +{ +throw new CorruptBlockException(getPath(), chunkOffset, chunkSize); +} + checksum.update(compressed.buffer, 0, chunkSize); if (out.readInt() != (int) checksum.getValue())
[jira] [Commented] (CASSANDRA-6176) remove decompression code in CompressedSequentialWriter
[ https://issues.apache.org/jira/browse/CASSANDRA-6176?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13799817#comment-13799817 ] Marcus Eriksson commented on CASSANDRA-6176: no, but now it is remove decompression code in CompressedSequentialWriter --- Key: CASSANDRA-6176 URL: https://issues.apache.org/jira/browse/CASSANDRA-6176 Project: Cassandra Issue Type: Improvement Components: Core Reporter: Dave Brosius Assignee: Dave Brosius Priority: Trivial Fix For: 2.0.2 Attachments: 6176.txt Changes for https://issues.apache.org/jira/browse/CASSANDRA-5862 have removed the need to decompress chunks during CompressedSequentialWriter.resetAndTruncate. remove code. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (CASSANDRA-6218) Reduce WAN traffic while doing repairs
[ https://issues.apache.org/jira/browse/CASSANDRA-6218?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13799818#comment-13799818 ] Vijay commented on CASSANDRA-6218: -- Wont it be simpler to just forward (Similar to write forwards) to the difference of (A, B) and (C, D) to each other (after initial repair) than initiating another repair again between (A, B) and (C, D) in step 3? Another possible option: Consider (DC1: A, B, C and DC2: X, Y, Z) Start Merkel tree comparison between all the nodes, once the differences is identified: Stream within the DC and then across the DC using a proxy or a forwarder node picked. (A, B, C to X) and then (X, Y, Z to A) Now both the DC has all the inconsistent data hence they can stream the ranges which where identified as inconsistent Reduce WAN traffic while doing repairs -- Key: CASSANDRA-6218 URL: https://issues.apache.org/jira/browse/CASSANDRA-6218 Project: Cassandra Issue Type: Improvement Components: Core Reporter: sankalp kohli Priority: Minor The way we send out data that does not match over WAN can be improved. Example: Say there are four nodes(A,B,C,D) which are replica of a range we are repairing. A, B is in DC1 and C,D is in DC2. If A does not have the data which other replicas have, then we will have following streams 1) A to B and back 2) A to C and back(Goes over WAN) 3) A to D and back(Goes over WAN) One of the ways of doing it to reduce WAN traffic is this. 1) Repair A and B only with each other and C and D with each other starting at same time t. 2) Once these repairs have finished, A,B and C,D are in sync with respect to time t. 3) Now run a repair between A and C, the streams which are exchanged as a result of the diff will also be streamed to B and D via A and C(C and D behaves like a proxy to the streams). For a replication of DC1:2,DC2:2, the WAN traffic will get reduced by 50% and even more for higher replication factors. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Comment Edited] (CASSANDRA-6218) Reduce WAN traffic while doing repairs
[ https://issues.apache.org/jira/browse/CASSANDRA-6218?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13799818#comment-13799818 ] Vijay edited comment on CASSANDRA-6218 at 10/19/13 7:04 AM: Wont it be simpler to just forward (Similar to write forwards) to the difference of (A, B) and (C, D) to each other (after initial repair) than initiating another repair again between (A, B) and (C, D) in step 3? Another possible option: Consider (DC1: A, B, C and DC2: X, Y, Z) Start Merkel tree comparison between all the nodes, once the differences is identified: Stream within the DC and then across the DC using a proxy or a forwarder node picked. (A, B, C to X) and then (X, Y, Z to A) Now both the DC's have consistent data hence the proxy/forwarder can stream the ranges which where identified as inconsistent in the Merkel comparison was (Author: vijay2...@yahoo.com): Wont it be simpler to just forward (Similar to write forwards) to the difference of (A, B) and (C, D) to each other (after initial repair) than initiating another repair again between (A, B) and (C, D) in step 3? Another possible option: Consider (DC1: A, B, C and DC2: X, Y, Z) Start Merkel tree comparison between all the nodes, once the differences is identified: Stream within the DC and then across the DC using a proxy or a forwarder node picked. (A, B, C to X) and then (X, Y, Z to A) Now both the DC has all the inconsistent data hence they can stream the ranges which where identified as inconsistent Reduce WAN traffic while doing repairs -- Key: CASSANDRA-6218 URL: https://issues.apache.org/jira/browse/CASSANDRA-6218 Project: Cassandra Issue Type: Improvement Components: Core Reporter: sankalp kohli Priority: Minor The way we send out data that does not match over WAN can be improved. Example: Say there are four nodes(A,B,C,D) which are replica of a range we are repairing. A, B is in DC1 and C,D is in DC2. If A does not have the data which other replicas have, then we will have following streams 1) A to B and back 2) A to C and back(Goes over WAN) 3) A to D and back(Goes over WAN) One of the ways of doing it to reduce WAN traffic is this. 1) Repair A and B only with each other and C and D with each other starting at same time t. 2) Once these repairs have finished, A,B and C,D are in sync with respect to time t. 3) Now run a repair between A and C, the streams which are exchanged as a result of the diff will also be streamed to B and D via A and C(C and D behaves like a proxy to the streams). For a replication of DC1:2,DC2:2, the WAN traffic will get reduced by 50% and even more for higher replication factors. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (CASSANDRA-6142) Remove multithreaded compaction
[ https://issues.apache.org/jira/browse/CASSANDRA-6142?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=1370#comment-1370 ] Jonathan Ellis commented on CASSANDRA-6142: --- Pushed fixes for these to the same branch. (They are indeed existing bugs in LCR.) Remove multithreaded compaction --- Key: CASSANDRA-6142 URL: https://issues.apache.org/jira/browse/CASSANDRA-6142 Project: Cassandra Issue Type: Bug Components: Core Reporter: Jonathan Ellis Assignee: Jonathan Ellis Priority: Minor Fix For: 2.1 There is at best a very small sweet spot for multithreaded compaction (ParallelCompactionIterable). For large rows, we stall the pipeline and fall back to a single LCR pass. For small rows, the overhead of the coordination outweighs the benefits of parallelization (45s to compact 2x1M stress rows with multithreading enabled, vs 35 with it disabled). -- This message was sent by Atlassian JIRA (v6.1#6144)
[01/12] git commit: free metadata on close for CSF
Updated Branches: refs/heads/cassandra-1.2 e983ef13b - abe1395cb refs/heads/cassandra-2.0 232906d7e - 58014d303 refs/heads/trunk 656d3240f - d32f1eb21 free metadata on close for CSF Project: http://git-wip-us.apache.org/repos/asf/cassandra/repo Commit: http://git-wip-us.apache.org/repos/asf/cassandra/commit/6f0b8800 Tree: http://git-wip-us.apache.org/repos/asf/cassandra/tree/6f0b8800 Diff: http://git-wip-us.apache.org/repos/asf/cassandra/diff/6f0b8800 Branch: refs/heads/cassandra-2.0 Commit: 6f0b88002ecef88e31074356125f28f81a22a98a Parents: 16bb898 Author: Jonathan Ellis jbel...@apache.org Authored: Sat Oct 19 23:39:42 2013 +0100 Committer: Jonathan Ellis jbel...@apache.org Committed: Sat Oct 19 23:39:42 2013 +0100 -- src/java/org/apache/cassandra/io/util/CompressedSegmentedFile.java | 1 + 1 file changed, 1 insertion(+) -- http://git-wip-us.apache.org/repos/asf/cassandra/blob/6f0b8800/src/java/org/apache/cassandra/io/util/CompressedSegmentedFile.java -- diff --git a/src/java/org/apache/cassandra/io/util/CompressedSegmentedFile.java b/src/java/org/apache/cassandra/io/util/CompressedSegmentedFile.java index f41b1aa..d0ea3fd 100644 --- a/src/java/org/apache/cassandra/io/util/CompressedSegmentedFile.java +++ b/src/java/org/apache/cassandra/io/util/CompressedSegmentedFile.java @@ -57,5 +57,6 @@ public class CompressedSegmentedFile extends SegmentedFile implements ICompresse public void cleanup() { +metadata.close(); } }
[04/12] git commit: assert that we're not using the size of a freed memory region (a sign of reference-counting bugs)
assert that we're not using the size of a freed memory region (a sign of reference-counting bugs) Project: http://git-wip-us.apache.org/repos/asf/cassandra/repo Commit: http://git-wip-us.apache.org/repos/asf/cassandra/commit/0c65759e Tree: http://git-wip-us.apache.org/repos/asf/cassandra/tree/0c65759e Diff: http://git-wip-us.apache.org/repos/asf/cassandra/diff/0c65759e Branch: refs/heads/trunk Commit: 0c65759ef4cb91bbca60881c9e3aae39ce9ddc2f Parents: 6f0b880 Author: Jonathan Ellis jbel...@apache.org Authored: Sat Oct 19 23:43:10 2013 +0100 Committer: Jonathan Ellis jbel...@apache.org Committed: Sat Oct 19 23:43:10 2013 +0100 -- src/java/org/apache/cassandra/io/util/Memory.java | 1 + 1 file changed, 1 insertion(+) -- http://git-wip-us.apache.org/repos/asf/cassandra/blob/0c65759e/src/java/org/apache/cassandra/io/util/Memory.java -- diff --git a/src/java/org/apache/cassandra/io/util/Memory.java b/src/java/org/apache/cassandra/io/util/Memory.java index 91c601d..f276190 100644 --- a/src/java/org/apache/cassandra/io/util/Memory.java +++ b/src/java/org/apache/cassandra/io/util/Memory.java @@ -156,6 +156,7 @@ public class Memory public long size() { +assert peer != 0; return size; }
[03/12] git commit: assert that we're not using the size of a freed memory region (a sign of reference-counting bugs)
assert that we're not using the size of a freed memory region (a sign of reference-counting bugs) Project: http://git-wip-us.apache.org/repos/asf/cassandra/repo Commit: http://git-wip-us.apache.org/repos/asf/cassandra/commit/0c65759e Tree: http://git-wip-us.apache.org/repos/asf/cassandra/tree/0c65759e Diff: http://git-wip-us.apache.org/repos/asf/cassandra/diff/0c65759e Branch: refs/heads/cassandra-2.0 Commit: 0c65759ef4cb91bbca60881c9e3aae39ce9ddc2f Parents: 6f0b880 Author: Jonathan Ellis jbel...@apache.org Authored: Sat Oct 19 23:43:10 2013 +0100 Committer: Jonathan Ellis jbel...@apache.org Committed: Sat Oct 19 23:43:10 2013 +0100 -- src/java/org/apache/cassandra/io/util/Memory.java | 1 + 1 file changed, 1 insertion(+) -- http://git-wip-us.apache.org/repos/asf/cassandra/blob/0c65759e/src/java/org/apache/cassandra/io/util/Memory.java -- diff --git a/src/java/org/apache/cassandra/io/util/Memory.java b/src/java/org/apache/cassandra/io/util/Memory.java index 91c601d..f276190 100644 --- a/src/java/org/apache/cassandra/io/util/Memory.java +++ b/src/java/org/apache/cassandra/io/util/Memory.java @@ -156,6 +156,7 @@ public class Memory public long size() { +assert peer != 0; return size; }
[12/12] git commit: merge
merge Project: http://git-wip-us.apache.org/repos/asf/cassandra/repo Commit: http://git-wip-us.apache.org/repos/asf/cassandra/commit/d32f1eb2 Tree: http://git-wip-us.apache.org/repos/asf/cassandra/tree/d32f1eb2 Diff: http://git-wip-us.apache.org/repos/asf/cassandra/diff/d32f1eb2 Branch: refs/heads/trunk Commit: d32f1eb2128296b2077db6bffa226535852c47b9 Parents: 0a8f4ba 656d324 Author: Jonathan Ellis jbel...@apache.org Authored: Sun Oct 20 02:09:18 2013 +0100 Committer: Jonathan Ellis jbel...@apache.org Committed: Sun Oct 20 02:09:18 2013 +0100 -- NEWS.txt| 12 +++- .../io/compress/CompressedSequentialWriter.java | 12 .../cassandra/service/StorageService.java | 12 ++-- .../service/LeaveAndBootstrapTest.java | 72 ++-- .../org/apache/cassandra/service/MoveTest.java | 64 - .../cassandra/tools/SSTableExportTest.java | 4 +- 6 files changed, 99 insertions(+), 77 deletions(-) -- http://git-wip-us.apache.org/repos/asf/cassandra/blob/d32f1eb2/NEWS.txt -- diff --cc NEWS.txt index f8ef5ff,519ed55..1dd2a0e --- a/NEWS.txt +++ b/NEWS.txt @@@ -38,12 -38,6 +38,12 @@@ New feature - Compaction history and stats are now saved to system keyspace (system.compaction_history table). You can access historiy via new 'nodetool compactionhistory' command or CQL. - - Added a new consistenct level, LOCAL_ONE, that forces all CL.ONE operations to ++- Added a new consistency level, LOCAL_ONE, that forces all CL.ONE operations to + execute only in the local datacenter. +- New replace_address to supplant the (now removed) replace_token and + replace_node workflows to replace a dead node in place. Works like the + old options, but takes the IP address of the node to be replaced. + 2.0.1 =
[02/12] git commit: free metadata on close for CSF
free metadata on close for CSF Project: http://git-wip-us.apache.org/repos/asf/cassandra/repo Commit: http://git-wip-us.apache.org/repos/asf/cassandra/commit/6f0b8800 Tree: http://git-wip-us.apache.org/repos/asf/cassandra/tree/6f0b8800 Diff: http://git-wip-us.apache.org/repos/asf/cassandra/diff/6f0b8800 Branch: refs/heads/trunk Commit: 6f0b88002ecef88e31074356125f28f81a22a98a Parents: 16bb898 Author: Jonathan Ellis jbel...@apache.org Authored: Sat Oct 19 23:39:42 2013 +0100 Committer: Jonathan Ellis jbel...@apache.org Committed: Sat Oct 19 23:39:42 2013 +0100 -- src/java/org/apache/cassandra/io/util/CompressedSegmentedFile.java | 1 + 1 file changed, 1 insertion(+) -- http://git-wip-us.apache.org/repos/asf/cassandra/blob/6f0b8800/src/java/org/apache/cassandra/io/util/CompressedSegmentedFile.java -- diff --git a/src/java/org/apache/cassandra/io/util/CompressedSegmentedFile.java b/src/java/org/apache/cassandra/io/util/CompressedSegmentedFile.java index f41b1aa..d0ea3fd 100644 --- a/src/java/org/apache/cassandra/io/util/CompressedSegmentedFile.java +++ b/src/java/org/apache/cassandra/io/util/CompressedSegmentedFile.java @@ -57,5 +57,6 @@ public class CompressedSegmentedFile extends SegmentedFile implements ICompresse public void cleanup() { +metadata.close(); } }
[07/12] git commit: merge from 1.2
merge from 1.2 Project: http://git-wip-us.apache.org/repos/asf/cassandra/repo Commit: http://git-wip-us.apache.org/repos/asf/cassandra/commit/74d63ba5 Tree: http://git-wip-us.apache.org/repos/asf/cassandra/tree/74d63ba5 Diff: http://git-wip-us.apache.org/repos/asf/cassandra/diff/74d63ba5 Branch: refs/heads/cassandra-2.0 Commit: 74d63ba566ef783243802f2ab0c453e8c2440574 Parents: 0c65759 fa15524 Author: Jonathan Ellis jbel...@apache.org Authored: Sun Oct 20 00:21:48 2013 +0100 Committer: Jonathan Ellis jbel...@apache.org Committed: Sun Oct 20 00:21:48 2013 +0100 -- CHANGES.txt | 1 + NEWS.txt | 1 + .../org/apache/cassandra/hadoop/ColumnFamilyRecordReader.java | 3 +++ 3 files changed, 5 insertions(+) -- http://git-wip-us.apache.org/repos/asf/cassandra/blob/74d63ba5/CHANGES.txt -- diff --cc CHANGES.txt index f4d6980,70bb919..41e885e --- a/CHANGES.txt +++ b/CHANGES.txt @@@ -1,29 -1,8 +1,30 @@@ -1.2.12 +2.0.2 + * Update FailureDetector to use nanontime (CASSANDRA-4925) + * Fix FileCacheService regressions (CASSANDRA-6149) + * Never return WriteTimeout for CL.ANY (CASSANDRA-6032) + * Fix race conditions in bulk loader (CASSANDRA-6129) + * Add configurable metrics reporting (CASSANDRA-4430) + * drop queries exceeding a configurable number of tombstones (CASSANDRA-6117) + * Track and persist sstable read activity (CASSANDRA-5515) + * Fixes for speculative retry (CASSANDRA-5932) + * Improve memory usage of metadata min/max column names (CASSANDRA-6077) + * Fix thrift validation refusing row markers on CQL3 tables (CASSANDRA-6081) + * Fix insertion of collections with CAS (CASSANDRA-6069) + * Correctly send metadata on SELECT COUNT (CASSANDRA-6080) + * Track clients' remote addresses in ClientState (CASSANDRA-6070) + * Create snapshot dir if it does not exist when migrating + leveled manifest (CASSANDRA-6093) + * make sequential nodetool repair the default (CASSANDRA-5950) + * Add more hooks for compaction strategy implementations (CASSANDRA-6111) + * Fix potential NPE on composite 2ndary indexes (CASSANDRA-6098) + * Delete can potentially be skipped in batch (CASSANDRA-6115) + * Allow alter keyspace on system_traces (CASSANDRA-6016) + * Disallow empty column names in cql (CASSANDRA-6136) + * Use Java7 file-handling APIs and fix file moving on Windows (CASSANDRA-5383) + * Save compaction history to system keyspace (CASSANDRA-5078) + * Fix NPE if StorageService.getOperationMode() is executed before full startup (CASSANDRA-6166) +Merged from 1.2: + * (Hadoop) Require CFRR batchSize to be at least 2 (CASSANDRA-6114) - - -1.2.11 * Add a warning for small LCS sstable size (CASSANDRA-6191) * Add ability to list specific KS/CF combinations in nodetool cfstats (CASSANDRA-4191) * Mark CF clean if a mutation raced the drop and got it marked dirty http://git-wip-us.apache.org/repos/asf/cassandra/blob/74d63ba5/NEWS.txt -- diff --cc NEWS.txt index 4bffa77,d803f02..7bd0d63 --- a/NEWS.txt +++ b/NEWS.txt @@@ -32,17 -25,14 +32,18 @@@ New feature replace_node workflows to replace a dead node in place. Works like the old options, but takes the IP address of the node to be replaced. ++ +2.0.1 += Upgrading - -- ColumnFamilyMetrics#sstablesPerReadHistogram switched from uniform sampling - to biased-to-last-five-minutes sampling. +- The default memtable allocation has changed from 1/3 of heap to 1/4 + of heap. Also, default (single-partition) read and write timeouts + have been reduced from 10s to 5s and 2s, respectively. -1.2.10 -== +2.0.0 += Upgrading - http://git-wip-us.apache.org/repos/asf/cassandra/blob/74d63ba5/src/java/org/apache/cassandra/hadoop/ColumnFamilyRecordReader.java --
[09/12] git commit: Merge branch 'cassandra-2.0' into trunk
Merge branch 'cassandra-2.0' into trunk Project: http://git-wip-us.apache.org/repos/asf/cassandra/repo Commit: http://git-wip-us.apache.org/repos/asf/cassandra/commit/0a8f4bab Tree: http://git-wip-us.apache.org/repos/asf/cassandra/tree/0a8f4bab Diff: http://git-wip-us.apache.org/repos/asf/cassandra/diff/0a8f4bab Branch: refs/heads/trunk Commit: 0a8f4bab76ac9ab57e79e20b0fdc4f2d0b3b8bfc Parents: 6c37934 74d63ba Author: Jonathan Ellis jbel...@apache.org Authored: Sun Oct 20 00:22:51 2013 +0100 Committer: Jonathan Ellis jbel...@apache.org Committed: Sun Oct 20 00:22:51 2013 +0100 -- CHANGES.txt | 1 + NEWS.txt | 1 + .../org/apache/cassandra/hadoop/ColumnFamilyRecordReader.java | 3 +++ .../org/apache/cassandra/io/util/CompressedSegmentedFile.java | 1 + src/java/org/apache/cassandra/io/util/Memory.java | 1 + 5 files changed, 7 insertions(+) -- http://git-wip-us.apache.org/repos/asf/cassandra/blob/0a8f4bab/CHANGES.txt -- http://git-wip-us.apache.org/repos/asf/cassandra/blob/0a8f4bab/NEWS.txt --
[06/12] git commit: Require CFRR batchSize to be at least 2 patch by Alex Liu and jbellis for CASSANDRA-6114
Require CFRR batchSize to be at least 2 patch by Alex Liu and jbellis for CASSANDRA-6114 Project: http://git-wip-us.apache.org/repos/asf/cassandra/repo Commit: http://git-wip-us.apache.org/repos/asf/cassandra/commit/fa155248 Tree: http://git-wip-us.apache.org/repos/asf/cassandra/tree/fa155248 Diff: http://git-wip-us.apache.org/repos/asf/cassandra/diff/fa155248 Branch: refs/heads/cassandra-2.0 Commit: fa155248eeb88fabe2a67fd9b591bc3401c1bae6 Parents: 558483d Author: Jonathan Ellis jbel...@apache.org Authored: Sun Oct 20 00:18:58 2013 +0100 Committer: Jonathan Ellis jbel...@apache.org Committed: Sun Oct 20 00:18:58 2013 +0100 -- CHANGES.txt | 4 .../org/apache/cassandra/hadoop/ColumnFamilyRecordReader.java| 3 +++ 2 files changed, 7 insertions(+) -- http://git-wip-us.apache.org/repos/asf/cassandra/blob/fa155248/CHANGES.txt -- diff --git a/CHANGES.txt b/CHANGES.txt index 87be6fa..70bb919 100644 --- a/CHANGES.txt +++ b/CHANGES.txt @@ -1,3 +1,7 @@ +1.2.12 + * (Hadoop) Require CFRR batchSize to be at least 2 (CASSANDRA-6114) + + 1.2.11 * Add a warning for small LCS sstable size (CASSANDRA-6191) * Add ability to list specific KS/CF combinations in nodetool cfstats (CASSANDRA-4191) http://git-wip-us.apache.org/repos/asf/cassandra/blob/fa155248/src/java/org/apache/cassandra/hadoop/ColumnFamilyRecordReader.java -- diff --git a/src/java/org/apache/cassandra/hadoop/ColumnFamilyRecordReader.java b/src/java/org/apache/cassandra/hadoop/ColumnFamilyRecordReader.java index 701260a..6846356 100644 --- a/src/java/org/apache/cassandra/hadoop/ColumnFamilyRecordReader.java +++ b/src/java/org/apache/cassandra/hadoop/ColumnFamilyRecordReader.java @@ -144,6 +144,9 @@ public class ColumnFamilyRecordReader extends RecordReaderByteBuffer, SortedMap cfName = ConfigHelper.getInputColumnFamily(conf); consistencyLevel = ConsistencyLevel.valueOf(ConfigHelper.getReadConsistencyLevel(conf)); keyspace = ConfigHelper.getInputKeyspace(conf); + +if (batchSize 2) +throw new IllegalArgumentException(Minimum batchSize is 2. Suggested batchSize is 100 or more); try {
[11/12] git commit: merge
merge Project: http://git-wip-us.apache.org/repos/asf/cassandra/repo Commit: http://git-wip-us.apache.org/repos/asf/cassandra/commit/58014d30 Tree: http://git-wip-us.apache.org/repos/asf/cassandra/tree/58014d30 Diff: http://git-wip-us.apache.org/repos/asf/cassandra/diff/58014d30 Branch: refs/heads/cassandra-2.0 Commit: 58014d3038621a9bd732724d226b377917fd62e7 Parents: 74d63ba 232906d Author: Jonathan Ellis jbel...@apache.org Authored: Sun Oct 20 02:08:52 2013 +0100 Committer: Jonathan Ellis jbel...@apache.org Committed: Sun Oct 20 02:08:52 2013 +0100 -- NEWS.txt| 12 +++- .../io/compress/CompressedSequentialWriter.java | 12 .../org/apache/cassandra/service/StorageService.java| 12 ++-- 3 files changed, 29 insertions(+), 7 deletions(-) -- http://git-wip-us.apache.org/repos/asf/cassandra/blob/58014d30/NEWS.txt -- diff --cc NEWS.txt index 7bd0d63,9d90ea7..69ab4fd --- a/NEWS.txt +++ b/NEWS.txt @@@ -26,12 -26,6 +26,12 @@@ New feature - Compaction history and stats are now saved to system keyspace (system.compaction_history table). You can access historiy via new 'nodetool compactionhistory' command or CQL. - - Added a new consistenct level, LOCAL_ONE, that forces all CL.ONE operations to ++- Added a new consistency level, LOCAL_ONE, that forces all CL.ONE operations to + execute only in the local datacenter. +- New replace_address to supplant the (now removed) replace_token and + replace_node workflows to replace a dead node in place. Works like the + old options, but takes the IP address of the node to be replaced. + 2.0.1 =
[08/12] git commit: merge from 1.2
merge from 1.2 Project: http://git-wip-us.apache.org/repos/asf/cassandra/repo Commit: http://git-wip-us.apache.org/repos/asf/cassandra/commit/74d63ba5 Tree: http://git-wip-us.apache.org/repos/asf/cassandra/tree/74d63ba5 Diff: http://git-wip-us.apache.org/repos/asf/cassandra/diff/74d63ba5 Branch: refs/heads/trunk Commit: 74d63ba566ef783243802f2ab0c453e8c2440574 Parents: 0c65759 fa15524 Author: Jonathan Ellis jbel...@apache.org Authored: Sun Oct 20 00:21:48 2013 +0100 Committer: Jonathan Ellis jbel...@apache.org Committed: Sun Oct 20 00:21:48 2013 +0100 -- CHANGES.txt | 1 + NEWS.txt | 1 + .../org/apache/cassandra/hadoop/ColumnFamilyRecordReader.java | 3 +++ 3 files changed, 5 insertions(+) -- http://git-wip-us.apache.org/repos/asf/cassandra/blob/74d63ba5/CHANGES.txt -- diff --cc CHANGES.txt index f4d6980,70bb919..41e885e --- a/CHANGES.txt +++ b/CHANGES.txt @@@ -1,29 -1,8 +1,30 @@@ -1.2.12 +2.0.2 + * Update FailureDetector to use nanontime (CASSANDRA-4925) + * Fix FileCacheService regressions (CASSANDRA-6149) + * Never return WriteTimeout for CL.ANY (CASSANDRA-6032) + * Fix race conditions in bulk loader (CASSANDRA-6129) + * Add configurable metrics reporting (CASSANDRA-4430) + * drop queries exceeding a configurable number of tombstones (CASSANDRA-6117) + * Track and persist sstable read activity (CASSANDRA-5515) + * Fixes for speculative retry (CASSANDRA-5932) + * Improve memory usage of metadata min/max column names (CASSANDRA-6077) + * Fix thrift validation refusing row markers on CQL3 tables (CASSANDRA-6081) + * Fix insertion of collections with CAS (CASSANDRA-6069) + * Correctly send metadata on SELECT COUNT (CASSANDRA-6080) + * Track clients' remote addresses in ClientState (CASSANDRA-6070) + * Create snapshot dir if it does not exist when migrating + leveled manifest (CASSANDRA-6093) + * make sequential nodetool repair the default (CASSANDRA-5950) + * Add more hooks for compaction strategy implementations (CASSANDRA-6111) + * Fix potential NPE on composite 2ndary indexes (CASSANDRA-6098) + * Delete can potentially be skipped in batch (CASSANDRA-6115) + * Allow alter keyspace on system_traces (CASSANDRA-6016) + * Disallow empty column names in cql (CASSANDRA-6136) + * Use Java7 file-handling APIs and fix file moving on Windows (CASSANDRA-5383) + * Save compaction history to system keyspace (CASSANDRA-5078) + * Fix NPE if StorageService.getOperationMode() is executed before full startup (CASSANDRA-6166) +Merged from 1.2: + * (Hadoop) Require CFRR batchSize to be at least 2 (CASSANDRA-6114) - - -1.2.11 * Add a warning for small LCS sstable size (CASSANDRA-6191) * Add ability to list specific KS/CF combinations in nodetool cfstats (CASSANDRA-4191) * Mark CF clean if a mutation raced the drop and got it marked dirty http://git-wip-us.apache.org/repos/asf/cassandra/blob/74d63ba5/NEWS.txt -- diff --cc NEWS.txt index 4bffa77,d803f02..7bd0d63 --- a/NEWS.txt +++ b/NEWS.txt @@@ -32,17 -25,14 +32,18 @@@ New feature replace_node workflows to replace a dead node in place. Works like the old options, but takes the IP address of the node to be replaced. ++ +2.0.1 += Upgrading - -- ColumnFamilyMetrics#sstablesPerReadHistogram switched from uniform sampling - to biased-to-last-five-minutes sampling. +- The default memtable allocation has changed from 1/3 of heap to 1/4 + of heap. Also, default (single-partition) read and write timeouts + have been reduced from 10s to 5s and 2s, respectively. -1.2.10 -== +2.0.0 += Upgrading - http://git-wip-us.apache.org/repos/asf/cassandra/blob/74d63ba5/src/java/org/apache/cassandra/hadoop/ColumnFamilyRecordReader.java --
[10/12] git commit: Require CFRR batchSize to be at least 2 patch by Alex Liu and jbellis for CASSANDRA-6114
Require CFRR batchSize to be at least 2 patch by Alex Liu and jbellis for CASSANDRA-6114 Project: http://git-wip-us.apache.org/repos/asf/cassandra/repo Commit: http://git-wip-us.apache.org/repos/asf/cassandra/commit/abe1395c Tree: http://git-wip-us.apache.org/repos/asf/cassandra/tree/abe1395c Diff: http://git-wip-us.apache.org/repos/asf/cassandra/diff/abe1395c Branch: refs/heads/cassandra-1.2 Commit: abe1395cbc29b21856d06b4bb3857fa7ae95eb18 Parents: e983ef1 Author: Jonathan Ellis jbel...@apache.org Authored: Sun Oct 20 00:18:58 2013 +0100 Committer: Jonathan Ellis jbel...@apache.org Committed: Sun Oct 20 02:08:08 2013 +0100 -- CHANGES.txt | 4 .../org/apache/cassandra/hadoop/ColumnFamilyRecordReader.java| 3 +++ 2 files changed, 7 insertions(+) -- http://git-wip-us.apache.org/repos/asf/cassandra/blob/abe1395c/CHANGES.txt -- diff --git a/CHANGES.txt b/CHANGES.txt index 87be6fa..70bb919 100644 --- a/CHANGES.txt +++ b/CHANGES.txt @@ -1,3 +1,7 @@ +1.2.12 + * (Hadoop) Require CFRR batchSize to be at least 2 (CASSANDRA-6114) + + 1.2.11 * Add a warning for small LCS sstable size (CASSANDRA-6191) * Add ability to list specific KS/CF combinations in nodetool cfstats (CASSANDRA-4191) http://git-wip-us.apache.org/repos/asf/cassandra/blob/abe1395c/src/java/org/apache/cassandra/hadoop/ColumnFamilyRecordReader.java -- diff --git a/src/java/org/apache/cassandra/hadoop/ColumnFamilyRecordReader.java b/src/java/org/apache/cassandra/hadoop/ColumnFamilyRecordReader.java index 701260a..6846356 100644 --- a/src/java/org/apache/cassandra/hadoop/ColumnFamilyRecordReader.java +++ b/src/java/org/apache/cassandra/hadoop/ColumnFamilyRecordReader.java @@ -144,6 +144,9 @@ public class ColumnFamilyRecordReader extends RecordReaderByteBuffer, SortedMap cfName = ConfigHelper.getInputColumnFamily(conf); consistencyLevel = ConsistencyLevel.valueOf(ConfigHelper.getReadConsistencyLevel(conf)); keyspace = ConfigHelper.getInputKeyspace(conf); + +if (batchSize 2) +throw new IllegalArgumentException(Minimum batchSize is 2. Suggested batchSize is 100 or more); try {
[jira] [Updated] (CASSANDRA-6135) Add beforeChange Notification to Gossiper State.
[ https://issues.apache.org/jira/browse/CASSANDRA-6135?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jonathan Ellis updated CASSANDRA-6135: -- Reviewer: Brandon Williams (was: Tyler Hobbs) Add beforeChange Notification to Gossiper State. Key: CASSANDRA-6135 URL: https://issues.apache.org/jira/browse/CASSANDRA-6135 Project: Cassandra Issue Type: New Feature Reporter: Benjamin Coverston Assignee: Sergio Bossa Attachments: 0001-New-Gossiper-notification-to-IEndpointStateChangeSub.patch, 0002-CASSANDRA-6135.diff, CASSANDRA-6135-V3.patch We would like an internal notification to be fired before state changes happen so we can intercept them, and in some cases defer them. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (CASSANDRA-6201) Expose compression metadata in cfstats
[ https://issues.apache.org/jira/browse/CASSANDRA-6201?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13800054#comment-13800054 ] Jonathan Ellis commented on CASSANDRA-6201: --- Approach looks good, but you should reference the sstable (and skip it if reference failse) before checking the metadata since otherwise you could include the size of a just-freed segment. Expose compression metadata in cfstats -- Key: CASSANDRA-6201 URL: https://issues.apache.org/jira/browse/CASSANDRA-6201 Project: Cassandra Issue Type: New Feature Components: Tools Reporter: Nicolas Favre-Felix Priority: Minor Labels: jmx, nodetool Attachments: cassandra-trunk-6201.patch It would be useful to expose how much off-heap memory is allocated for compression metadata. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (CASSANDRA-5818) Duplicated error messages on directory creation error at startup
[ https://issues.apache.org/jira/browse/CASSANDRA-5818?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13800055#comment-13800055 ] Jonathan Ellis commented on CASSANDRA-5818: --- How does that look to you, Mikhail? Duplicated error messages on directory creation error at startup Key: CASSANDRA-5818 URL: https://issues.apache.org/jira/browse/CASSANDRA-5818 Project: Cassandra Issue Type: Bug Reporter: Michaël Figuière Assignee: koray sariteke Priority: Trivial Fix For: 2.1 Attachments: patch.diff, trunk-5818.patch When I start Cassandra without the appropriate OS access rights to the default Cassandra directories, I get a flood of {{ERROR}} messages at startup, whereas one per directory would be more appropriate. See bellow: {code} ERROR 13:37:39,792 Failed to create /var/lib/cassandra/data/system/schema_triggers directory ERROR 13:37:39,797 Failed to create /var/lib/cassandra/data/system/schema_triggers directory ERROR 13:37:39,798 Failed to create /var/lib/cassandra/data/system/schema_triggers directory ERROR 13:37:39,798 Failed to create /var/lib/cassandra/data/system/schema_triggers directory ERROR 13:37:39,799 Failed to create /var/lib/cassandra/data/system/schema_triggers directory ERROR 13:37:39,800 Failed to create /var/lib/cassandra/data/system/batchlog directory ERROR 13:37:39,801 Failed to create /var/lib/cassandra/data/system/batchlog directory ERROR 13:37:39,801 Failed to create /var/lib/cassandra/data/system/batchlog directory ERROR 13:37:39,802 Failed to create /var/lib/cassandra/data/system/batchlog directory ERROR 13:37:39,802 Failed to create /var/lib/cassandra/data/system/peer_events directory ERROR 13:37:39,803 Failed to create /var/lib/cassandra/data/system/peer_events directory ERROR 13:37:39,803 Failed to create /var/lib/cassandra/data/system/peer_events directory ERROR 13:37:39,804 Failed to create /var/lib/cassandra/data/system/compactions_in_progress directory ERROR 13:37:39,805 Failed to create /var/lib/cassandra/data/system/compactions_in_progress directory ERROR 13:37:39,805 Failed to create /var/lib/cassandra/data/system/compactions_in_progress directory ERROR 13:37:39,806 Failed to create /var/lib/cassandra/data/system/compactions_in_progress directory ERROR 13:37:39,807 Failed to create /var/lib/cassandra/data/system/compactions_in_progress directory ERROR 13:37:39,808 Failed to create /var/lib/cassandra/data/system/hints directory ERROR 13:37:39,809 Failed to create /var/lib/cassandra/data/system/hints directory ERROR 13:37:39,809 Failed to create /var/lib/cassandra/data/system/hints directory ERROR 13:37:39,811 Failed to create /var/lib/cassandra/data/system/hints directory ERROR 13:37:39,811 Failed to create /var/lib/cassandra/data/system/hints directory ERROR 13:37:39,812 Failed to create /var/lib/cassandra/data/system/schema_keyspaces directory ERROR 13:37:39,812 Failed to create /var/lib/cassandra/data/system/schema_keyspaces directory ERROR 13:37:39,813 Failed to create /var/lib/cassandra/data/system/schema_keyspaces directory ERROR 13:37:39,814 Failed to create /var/lib/cassandra/data/system/schema_keyspaces directory ERROR 13:37:39,814 Failed to create /var/lib/cassandra/data/system/schema_keyspaces directory ERROR 13:37:39,815 Failed to create /var/lib/cassandra/data/system/range_xfers directory ERROR 13:37:39,816 Failed to create /var/lib/cassandra/data/system/range_xfers directory ERROR 13:37:39,817 Failed to create /var/lib/cassandra/data/system/range_xfers directory ERROR 13:37:39,817 Failed to create /var/lib/cassandra/data/system/schema_columnfamilies directory ERROR 13:37:39,818 Failed to create /var/lib/cassandra/data/system/schema_columnfamilies directory ERROR 13:37:39,818 Failed to create /var/lib/cassandra/data/system/schema_columnfamilies directory ERROR 13:37:39,820 Failed to create /var/lib/cassandra/data/system/schema_columnfamilies directory ERROR 13:37:39,821 Failed to create /var/lib/cassandra/data/system/schema_columnfamilies directory ERROR 13:37:39,821 Failed to create /var/lib/cassandra/data/system/schema_columnfamilies directory ERROR 13:37:39,822 Failed to create /var/lib/cassandra/data/system/schema_columnfamilies directory ERROR 13:37:39,822 Failed to create /var/lib/cassandra/data/system/schema_columnfamilies directory ERROR 13:37:39,823 Failed to create /var/lib/cassandra/data/system/schema_columnfamilies directory ERROR 13:37:39,824 Failed to create /var/lib/cassandra/data/system/schema_columnfamilies directory ERROR 13:37:39,824 Failed to create /var/lib/cassandra/data/system/schema_columnfamilies directory ERROR 13:37:39,825 Failed to
[jira] [Commented] (CASSANDRA-6114) Pig with widerows=true and batch size = 1 works incorrectly
[ https://issues.apache.org/jira/browse/CASSANDRA-6114?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13800056#comment-13800056 ] Jonathan Ellis commented on CASSANDRA-6114: --- I think batchSize==1 is broken for non-wide-rows as well. Updated to throw InvalidArgument for either and committed. Pig with widerows=true and batch size = 1 works incorrectly --- Key: CASSANDRA-6114 URL: https://issues.apache.org/jira/browse/CASSANDRA-6114 Project: Cassandra Issue Type: Bug Components: Hadoop Reporter: Alex Liu Assignee: Alex Liu Priority: Minor Fix For: 1.2.12, 2.0.2 Attachments: 6114.txt If I run the demo pig scripts, I end up with a column family with 6 fairly wide rows. If I load and dump those rows with widerows=true or set the cassandra.range.batch.size=1, the dump returns the correct values. However, if I set both of those, it does not. So in the case of a batch size of 1, wide rows support is broken. So it's relatively simple to reproduce from the demo data: {code} grunt SET cassandra.range.batch.size 1 grunt rows = LOAD 'cassandra://PigDemo/Scores' using CassandraStorage(); grunt dump rows; ... (sylvain,{(4,),(7,),(10,),(21,),(24,),(46,),(47,),(49,),(51,),(52,),(67,),(68,),(72,),(73,),(82,),(83,),(86,),(98,),(101,),(105,),(108,),(112,),(114,),(124,),(125,),(136,),(139,),(145,),(150,),(151,),(153,),(165,),(167,),(171,),(178,),(182,),(202,),(211,),(212,),(215,),(226,),(237,),(242,),(243,),(255,),(261,),(273,),(282,),(300,),(307,),(308,),(311,),(312,),(313,),(316,),(317,),(332,),(337,),(338,),(348,),(355,),(360,),(361,),(373,),(375,),(377,),(384,),(401,),(404,),(412,),(418,),(429,),(436,),(441,),(451,),(453,),(461,),(473,),(478,),(483,),(485,),(486,),(489,),(509,),(511,),(516,),(517,),(521,),(536,),(541,),(543,),(545,),(550,),(583,),(587,),(592,),(611,),(613,),(622,),(625,),(627,),(633,),(648,),(649,),(651,),(659,),(665,),(668,),(670,),(672,),(679,),(688,),(692,),(700,),(703,),(707,),(709,),(730,),(731,),(738,),(740,),(744,),(750,),(759,),(764,),(766,),(768,),(774,),(776,),(778,),(779,),(788,),(795,),(796,),(813,),(821,),(825,),(830,),(831,),(835,),(843,),(846,),(847,),(848,),(851,),(862,),(863,),(872,),(878,),(881,),(883,),(884,),(888,),(905,),(906,),(916,),(921,),(926,),(928,),(944,),(946,),(947,),(952,),(954,),(972,),(973,),(974,),(976,),(978,),(982,),(991,)}) (brandon,{(6,),(7,),(14,),(15,),(25,),(36,),(37,),(38,),(46,),(53,),(57,),(65,),(74,),(75,),(84,),(91,),(104,),(120,),(128,),(137,),(148,),(159,),(171,),(174,),(176,),(179,),(183,),(192,),(195,),(201,),(205,),(210,),(216,),(222,),(223,),(243,),(255,),(264,),(271,),(287,),(290,),(308,),(309,),(326,),(343,),(347,),(356,),(359,),(360,),(363,),(367,),(368,),(378,),(398,),(400,),(402,),(410,),(412,),(419,),(427,),(429,),(447,),(449,),(462,),(464,),(468,),(470,),(472,),(480,),(482,),(506,),(511,),(520,),(521,),(522,),(524,),(535,),(548,),(553,),(565,),(569,),(571,),(573,),(575,),(583,),(584,),(595,),(597,),(606,),(608,),(634,),(646,),(650,),(654,),(667,),(673,),(677,),(686,),(690,),(692,),(713,),(715,),(721,),(723,),(736,),(737,),(752,),(753,),(758,),(759,),(764,),(766,),(767,),(776,),(778,),(786,),(812,),(816,),(818,),(823,),(826,),(832,),(838,),(842,),(860,),(873,),(879,),(918,),(919,),(935,),(941,),(942,),(948,),(956,),(961,),(966,),(973,),(974,),(977,),(979,),(983,),(984,),(986,),(995,),(997,)}) (jake,{(1,),(7,),(10,),(14,),(29,),(52,),(54,),(65,),(67,),(78,),(82,),(83,),(89,),(97,),(100,),(115,),(126,),(140,),(141,),(145,),(214,),(221,),(230,),(231,),(232,),(241,),(245,),(247,),(265,),(266,),(269,),(271,),(282,),(286,),(288,),(299,),(316,),(323,),(331,),(332,),(335,),(338,),(348,),(353,),(355,),(364,),(367,),(371,),(379,),(398,),(409,),(420,),(428,),(429,),(439,),(443,),(450,),(454,),(467,),(477,),(482,),(488,),(490,),(502,),(503,),(512,),(520,),(521,),(535,),(536,),(541,),(548,),(552,),(557,),(560,),(596,),(600,),(604,),(606,),(611,),(613,),(621,),(624,),(630,),(635,),(641,),(647,),(655,),(660,),(665,),(674,),(676,),(690,),(693,),(694,),(704,),(719,),(720,),(724,),(731,),(749,),(751,),(763,),(765,),(767,),(771,),(779,),(782,),(784,),(789,),(793,),(797,),(798,),(801,),(802,),(806,),(820,),(825,),(839,),(845,),(848,),(856,),(865,),(866,),(867,),(870,),(876,),(887,),(891,),(901,),(905,),(908,),(922,),(929,),(944,),(960,),(964,),(980,),(988,),(996,)})
[jira] [Commented] (CASSANDRA-4809) Allow restoring specific column families from archived commitlog
[ https://issues.apache.org/jira/browse/CASSANDRA-4809?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13800053#comment-13800053 ] Jonathan Ellis commented on CASSANDRA-4809: --- - Comment doesn't match code -- says empty string array but code is array of empty string. I think there are existing examples of provide ks+cf, otherwise default to everything. Suggest checking that out; the approach is a bit clunky. - Besides splitting the parameter, should trim leading + trailing spaces. Guarantee that will be a common form of user confusion otherwise. Allow restoring specific column families from archived commitlog Key: CASSANDRA-4809 URL: https://issues.apache.org/jira/browse/CASSANDRA-4809 Project: Cassandra Issue Type: Improvement Affects Versions: 1.2.0 Reporter: Nick Bailey Assignee: Lyuben Todorov Labels: lhf Fix For: 2.0.2 Attachments: 4809.patch, 4809_v2.patch, 4809_v3.patch Currently you can only restore the entire contents of a commit log archive. It would be useful to specify the keyspaces/column families you want to restore from an archived commitlog. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Commented] (CASSANDRA-6109) Consider coldness in STCS compaction
[ https://issues.apache.org/jira/browse/CASSANDRA-6109?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13800057#comment-13800057 ] Jonathan Ellis commented on CASSANDRA-6109: --- What if we just added a bucket filter that said, SSTables representing less than X% of the reads will not be bucketed? Straightforward to tune and I can't think of any really pathological cases, other than where size-tiering just doesn't put hot overlapping sstables in the same bucket. (Which I think is out of scope to solve here -- we need cardinality estimation to fix that.) Consider coldness in STCS compaction Key: CASSANDRA-6109 URL: https://issues.apache.org/jira/browse/CASSANDRA-6109 Project: Cassandra Issue Type: New Feature Components: Core Reporter: Jonathan Ellis Assignee: Tyler Hobbs Fix For: 2.0.2 Attachments: 6109-v1.patch, 6109-v2.patch I see two options: # Don't compact cold sstables at all # Compact cold sstables only if there is nothing more important to compact The latter is better if you have cold data that may become hot again... but it's confusing if you have a workload such that you can't keep up with *all* compaction, but you can keep up with hot sstable. (Compaction backlog stat becomes useless since we fall increasingly behind.) -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Updated] (CASSANDRA-6196) Add compaction, compression to cqlsh tab completion for CREATE TABLE
[ https://issues.apache.org/jira/browse/CASSANDRA-6196?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jonathan Ellis updated CASSANDRA-6196: -- Reviewer: Brandon Williams (was: Aleksey Yeschenko) Add compaction, compression to cqlsh tab completion for CREATE TABLE Key: CASSANDRA-6196 URL: https://issues.apache.org/jira/browse/CASSANDRA-6196 Project: Cassandra Issue Type: Bug Components: Tools Reporter: Jonathan Ellis Assignee: Mikhail Stepura Priority: Minor Fix For: 2.0.2 Attachments: cassandra-2.0-6196.patch -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Updated] (CASSANDRA-6048) Add the ability to use multiple indexes in a single query
[ https://issues.apache.org/jira/browse/CASSANDRA-6048?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Alex Liu updated CASSANDRA-6048: Attachment: (was: 6048-trunk.txt) Add the ability to use multiple indexes in a single query - Key: CASSANDRA-6048 URL: https://issues.apache.org/jira/browse/CASSANDRA-6048 Project: Cassandra Issue Type: New Feature Components: Core Reporter: Alex Liu Assignee: Alex Liu Fix For: 2.1 Attachments: 6048-1.2-branch.txt, 6048-trunk.txt Existing data filtering uses the following algorithm {code} 1. find best selective predicate based on the smallest mean columns count 2. fetch rows for the best selective predicate predicate, then filter the data based on other predicates left. {code} So potentially we could improve the performance by {code} 1. joining multiple predicates then do the data filtering for other predicates. 2. fine tune the best predicate selection algorithm {code} For multiple predicate join, it could improve performance if one predicate has many entries and another predicate has a very few of entries. It means a few index CF read, join the row keys, fetch rows then filter other predicates Another approach is to have index on multiple columns. -- This message was sent by Atlassian JIRA (v6.1#6144)
[jira] [Updated] (CASSANDRA-6048) Add the ability to use multiple indexes in a single query
[ https://issues.apache.org/jira/browse/CASSANDRA-6048?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Alex Liu updated CASSANDRA-6048: Attachment: 6048-trunk.txt Add the ability to use multiple indexes in a single query - Key: CASSANDRA-6048 URL: https://issues.apache.org/jira/browse/CASSANDRA-6048 Project: Cassandra Issue Type: New Feature Components: Core Reporter: Alex Liu Assignee: Alex Liu Fix For: 2.1 Attachments: 6048-1.2-branch.txt, 6048-trunk.txt Existing data filtering uses the following algorithm {code} 1. find best selective predicate based on the smallest mean columns count 2. fetch rows for the best selective predicate predicate, then filter the data based on other predicates left. {code} So potentially we could improve the performance by {code} 1. joining multiple predicates then do the data filtering for other predicates. 2. fine tune the best predicate selection algorithm {code} For multiple predicate join, it could improve performance if one predicate has many entries and another predicate has a very few of entries. It means a few index CF read, join the row keys, fetch rows then filter other predicates Another approach is to have index on multiple columns. -- This message was sent by Atlassian JIRA (v6.1#6144)