Rok Doltar created CASSANDRA-13770: -------------------------------------- Summary: AssertionError: Lower bound INCL_START_BOUND during select by index Key: CASSANDRA-13770 URL: https://issues.apache.org/jira/browse/CASSANDRA-13770 Project: Cassandra Issue Type: Bug Environment: Cassandra 3.11 (cassandra.noarch 3.11.0-1), CentOS Linux release 7.3.1611 (Core)
Reporter: Rok Doltar We are getting the following error: DEBUG [Native-Transport-Requests-1] 2017-08-17 07:47:01,815 ReadCallback.java:132 - Failed; received 0 of 1 responses WARN [ReadStage-2] 2017-08-17 07:47:01,816 AbstractLocalAwareExecutorService.java:167 - Uncaught exception on thread Thread[ReadStage-2,5,main]: {} java.lang.AssertionError: Lower bound [INCL_START_BOUND(0028354338333835414433363737373137344544303642413442444246344544443932334538463946340000283836453642373436354546423435334544363636443236344644313935333032363338314542363200, ab570080-831f-11e7-a81f-417b646547c3, , 1x) ]is bigger than first returned value [Row: partition_key=0028354338333835414433363737373137344544303642413442444246344544443932334538463946340000283836453642373436354546423435334544363636443236344644313935333032363338314542363200, version=null, file_path=null, file_name=null | ] for sstable /var/lib/cassandra/data/catalog/file-aa90a340831f11e7aca2ed895c1dab3f/.idx_file_path_hash/mc-51-big-Data.db at org.apache.cassandra.db.rows.UnfilteredRowIteratorWithLowerBound.computeNext(UnfilteredRowIteratorWithLowerBound.java:124) ~[apache-cassandra-3.11.0.jar:3.11.0] at org.apache.cassandra.db.rows.UnfilteredRowIteratorWithLowerBound.computeNext(UnfilteredRowIteratorWithLowerBound.java:47) ~[apache-cassandra-3.11.0.jar:3.11.0] at org.apache.cassandra.utils.AbstractIterator.hasNext(AbstractIterator.java:47) ~[apache-cassandra-3.11.0.jar:3.11.0] at org.apache.cassandra.utils.MergeIterator$Candidate.advance(MergeIterator.java:374) ~[apache-cassandra-3.11.0.jar:3.11.0] at org.apache.cassandra.utils.MergeIterator$ManyToOne.advance(MergeIterator.java:186) ~[apache-cassandra-3.11.0.jar:3.11.0] at org.apache.cassandra.utils.MergeIterator$ManyToOne.computeNext(MergeIterator.java:155) ~[apache-cassandra-3.11.0.jar:3.11.0] at org.apache.cassandra.utils.AbstractIterator.hasNext(AbstractIterator.java:47) ~[apache-cassandra-3.11.0.jar:3.11.0] at org.apache.cassandra.db.rows.UnfilteredRowIterators$UnfilteredRowMergeIterator.computeNext(UnfilteredRowIterators.java:500) ~[apache-cassandra-3.11.0.jar:3.11.0] at org.apache.cassandra.db.rows.UnfilteredRowIterators$UnfilteredRowMergeIterator.computeNext(UnfilteredRowIterators.java:360) ~[apache-cassandra-3.11.0.jar:3.11.0] at org.apache.cassandra.utils.AbstractIterator.hasNext(AbstractIterator.java:47) ~[apache-cassandra-3.11.0.jar:3.11.0] at org.apache.cassandra.db.rows.UnfilteredRowIterator.isEmpty(UnfilteredRowIterator.java:67) ~[apache-cassandra-3.11.0.jar:3.11.0] at org.apache.cassandra.db.SinglePartitionReadCommand.withSSTablesIterated(SinglePartitionReadCommand.java:695) ~[apache-cassandra-3.11.0.jar:3.11.0] at org.apache.cassandra.db.SinglePartitionReadCommand.queryMemtableAndDiskInternal(SinglePartitionReadCommand.java:639) ~[apache-cassandra-3.11.0.jar:3.11.0] at org.apache.cassandra.db.SinglePartitionReadCommand.queryMemtableAndDisk(SinglePartitionReadCommand.java:514) ~[apache-cassandra-3.11.0.jar:3.11.0] at org.apache.cassandra.index.internal.CassandraIndexSearcher.queryIndex(CassandraIndexSearcher.java:81) ~[apache-cassandra-3.11.0.jar:3.11.0] at org.apache.cassandra.index.internal.CassandraIndexSearcher.search(CassandraIndexSearcher.java:63) ~[apache-cassandra-3.11.0.jar:3.11.0] at org.apache.cassandra.db.ReadCommand.executeLocally(ReadCommand.java:408) ~[apache-cassandra-3.11.0.jar:3.11.0] at org.apache.cassandra.service.StorageProxy$LocalReadRunnable.runMayThrow(StorageProxy.java:1882) ~[apache-cassandra-3.11.0.jar:3.11.0] at org.apache.cassandra.service.StorageProxy$DroppableRunnable.run(StorageProxy.java:2587) ~[apache-cassandra-3.11.0.jar:3.11.0] at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) ~[na:1.8.0_141] at org.apache.cassandra.concurrent.AbstractLocalAwareExecutorService$FutureTask.run(AbstractLocalAwareExecutorService.java:162) ~[apache-cassandra-3.11.0.jar:3.11.0] at org.apache.cassandra.concurrent.AbstractLocalAwareExecutorService$LocalSessionFutureTask.run(AbstractLocalAwareExecutorService.java:134) [apache-cassandra-3.11.0.jar:3.11.0] at org.apache.cassandra.concurrent.SEPWorker.run(SEPWorker.java:109) [apache-cassandra-3.11.0.jar:3.11.0] at java.lang.Thread.run(Thread.java:748) [na:1.8.0_141] The related table is: CREATE TABLE catalog.file ( path_hash text, file_hash text, version timeuuid, file_path text, file_name text, allocations_size bigint, change_time timestamp, creation_time timestamp, dacl frozen<acl>, ea_size bigint, end_of_file bigint, file_attributes bigint, file_id blob, group_sid frozen<sid>, host text static, last_access_time timestamp, last_write_time timestamp, owner_sid frozen<sid>, share text static, PRIMARY KEY ((path_hash, file_hash), version, file_path, file_name) ) WITH CLUSTERING ORDER BY (version DESC, file_path ASC, file_name ASC) AND bloom_filter_fp_chance = 0.01 AND caching = {'keys': 'ALL', 'rows_per_partition': 'NONE'} AND comment = '' AND compaction = {'class': 'org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy', 'max_threshold': '32', 'min_threshold': '4'} AND compression = {'chunk_length_in_kb': '64', 'class': 'org.apache.cassandra.io.compress.LZ4Compressor'} AND crc_check_chance = 1.0 AND dclocal_read_repair_chance = 0.1 AND default_time_to_live = 0 AND gc_grace_seconds = 864000 AND max_index_interval = 2048 AND memtable_flush_period_in_ms = 0 AND min_index_interval = 128 AND read_repair_chance = 0.0 AND speculative_retry = '99PERCENTILE'; CREATE INDEX idx_file_path_hash ON catalog.file (path_hash); This happens always when we are querying by the idx_file_path_hash index and always for the same path_hash "5C8385AD36777174ED06BA4BDBF4EDD923E8F9F4": cqlsh> select * from catalog.file where path_hash='5C8385AD36777174ED06BA4BDBF4EDD923E8F9F4'; ReadFailure: Error from server: code=1300 [Replica(s) failed to execute read] message="Operation failed - received 0 responses and 1 failures" info={'failures': 1, 'received_responses': 0, 'required_responses': 1, 'consistency': 'ONE'} If querying without using index the data is displayed just fine: cqlsh> select * from catalog.file LIMIT 1; path_hash | file_hash | version | file_path | file_name | host | share | allocations_size | change_time | creation_time | dacl | ea_size | end_of_file | file_attributes | file_id | group_sid | last_access_time | last_write_time | owner_sid ------------------------------------------+------------------------------------------+--------------------------------------+----------------------------------+------------------------------------+--------------+-------+------------------+---------------------------------+---------------------------------+---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+---------+-------------+-----------------+--------------------+-----------------------------------------------------------------------------------------------------------------------+---------------------------------+---------------------------------+----------------------------------------------------------------------------------------------------------------------- 5C8385AD36777174ED06BA4BDBF4EDD923E8F9F4 | 86E6B7465EFB453ED666D264FD1953026381EB62 | ab570080-831f-11e7-a81f-417b646547c3 | 2015_VSIcon/SchemaObjectProperty | SchemaObjectProperty_16x_24.bmp_13 | 10.17.62.151 | rokd | 12288 | 2017-07-12 11:51:20.159000+0000 | 2017-07-12 11:51:20.151000+0000 | {revision: 2, aces: [{ace_type: 0, ace_flags: {16}, ace_size: 0, access_mask: null, sid: {revision: 1, sid_identifier_authority: 0x000000000005, sub_authorities: [21, 769239019, 917752761, 3061700898, 500]}}, {ace_type: 0, ace_flags: {16}, ace_size: 0, access_mask: null, sid: {revision: 1, sid_identifier_authority: 0x000000000005, sub_authorities: [32, 544]}}, {ace_type: 0, ace_flags: {16}, ace_size: 0, access_mask: null, sid: {revision: 1, sid_identifier_authority: 0x000000000005, sub_authorities: [32, 545]}}]} | 0 | 822 | 33 | 0xd5a2000000000000 | {revision: 1, sid_identifier_authority: 0x000000000005, sub_authorities: [21, 769239019, 917752761, 3061700898, 513]} | 2017-07-12 11:51:20.151000+0000 | 2016-01-08 09:50:34.000000+0000 | {revision: 1, sid_identifier_authority: 0x000000000005, sub_authorities: [21, 769239019, 917752761, 3061700898, 500]} Rebuiding index doesn't help: # nodetool rebuild_index catalog file idx_file_path_hash -- This message was sent by Atlassian JIRA (v6.4.14#64029) --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@cassandra.apache.org For additional commands, e-mail: commits-h...@cassandra.apache.org