Re: Repair corrupt SSTable from power outage?

2015-10-06 Thread John Anderson
I'm running 2.1.9

On Fri, Oct 2, 2015 at 8:40 AM, Tyler Hobbs  wrote:

> What version of Cassandra are you running?  It sounds like the disk
> failure policy is incorrectly being applied to scrub, which kind of defeats
> the purpose of it.  I'd recommend opening a ticket (
> https://issues.apache.org/jira/browse/CASSANDRA) with the information you
> posted here.
>
> On Fri, Oct 2, 2015 at 3:01 AM, George Sigletos 
> wrote:
>
>> I'm also facing problems regarding corrupt sstables and also couldn't run
>> sstablescrub successfully.
>>
>> I restarted my nodes with disk failure policy "best_effort", then I run
>> the "nodetool scrub "
>>
>> Once done I removed the corrupt tables manually and started repair
>>
>>
>> On Thu, Oct 1, 2015 at 7:27 PM, John Anderson  wrote:
>>
>>> I have a 25 node cluster and we lost power on one of the racks last
>>> night and now 6 of our nodes will not start up and we are getting the
>>> following error:
>>>
>>> INFO  [main] 2015-10-01 10:19:22,111 CassandraDaemon.java:122 - JMX is
>>> enabled to receive remote connections on port: 7199
>>> INFO  [main] 2015-10-01 10:19:22,124 CacheService.java:111 -
>>> Initializing key cache with capacity of 100 MBs.
>>> INFO  [main] 2015-10-01 10:19:22,129 CacheService.java:133 -
>>> Initializing row cache with capacity of 0 MBs
>>> INFO  [main] 2015-10-01 10:19:22,133 CacheService.java:150 -
>>> Initializing counter cache with capacity of 50 MBs
>>> INFO  [main] 2015-10-01 10:19:22,135 CacheService.java:161 - Scheduling
>>> counter cache save to every 7200 seconds (going to save all keys).
>>> INFO  [main] 2015-10-01 10:19:22,211 ColumnFamilyStore.java:363 -
>>> Initializing system.sstable_activity
>>> INFO  [SSTableBatchOpen:1] 2015-10-01 10:19:22,639
>>> SSTableReader.java:478 - Opening
>>> /mnt/cassandra/data/data/system/sstable_activity-5a1ff267ace03f128563cfae6103c65e/system-sstable_activity-ka-45
>>> (805 bytes)
>>> ERROR [SSTableBatchOpen:1] 2015-10-01 10:19:22,657 FileUtils.java:447 -
>>> Exiting forcefully due to file system exception on startup, disk failure
>>> policy "stop"
>>> org.apache.cassandra.io.sstable.CorruptSSTableException:
>>> java.io.EOFException
>>> at
>>> org.apache.cassandra.io.compress.CompressionMetadata.(CompressionMetadata.java:131)
>>> ~[apache-cassandra-2.1.9.jar:2.1.9]
>>> at
>>> org.apache.cassandra.io.compress.CompressionMetadata.create(CompressionMetadata.java:85)
>>> ~[apache-cassandra-2.1.9.jar:2.1.9]
>>> at
>>> org.apache.cassandra.io.util.CompressedSegmentedFile$Builder.metadata(CompressedSegmentedFile.java:79)
>>> ~[apache-cassandra-2.1.9.jar:2.1.9]
>>> at
>>> org.apache.cassandra.io.util.CompressedPoolingSegmentedFile$Builder.complete(CompressedPoolingSegmentedFile.java:72)
>>> ~[apache-cassandra-2.1.9.jar:2.1.9]
>>> at
>>> org.apache.cassandra.io.util.SegmentedFile$Builder.complete(SegmentedFile.java:168)
>>> ~[apache-cassandra-2.1.9.jar:2.1.9]
>>> at
>>> org.apache.cassandra.io.sstable.SSTableReader.load(SSTableReader.java:752)
>>> ~[apache-cassandra-2.1.9.jar:2.1.9]
>>> at
>>> org.apache.cassandra.io.sstable.SSTableReader.load(SSTableReader.java:703)
>>> ~[apache-cassandra-2.1.9.jar:2.1.9]
>>> at
>>> org.apache.cassandra.io.sstable.SSTableReader.open(SSTableReader.java:491)
>>> ~[apache-cassandra-2.1.9.jar:2.1.9]
>>> at
>>> org.apache.cassandra.io.sstable.SSTableReader.open(SSTableReader.java:387)
>>> ~[apache-cassandra-2.1.9.jar:2.1.9]
>>> at
>>> org.apache.cassandra.io.sstable.SSTableReader$4.run(SSTableReader.java:534)
>>> ~[apache-cassandra-2.1.9.jar:2.1.9]
>>> at
>>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>>> [na:1.8.0_60]
>>> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>>> [na:1.8.0_60]
>>> at
>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>>> [na:1.8.0_60]
>>> at
>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>>> [na:1.8.0_60]
>>> at java.lang.Thread.run(Thread.java:745) [na:1.8.0_60]
>>> Caused by: java.io.EOFException: null
>>> at
>>> java.io.DataInputStream.readUnsignedShort(DataInputStream.java:340)
>>> ~[na:1.8.0_60]
>>> at java.io.DataInputStream.readUTF(DataInputStream.java:589)
>>> ~[na:1.8.0_60]
>>> at java.io.DataInputStream.readUTF(DataInputStream.java:564)
>>> ~[na:1.8.0_60]
>>> at
>>> org.apache.cassandra.io.compress.CompressionMetadata.(CompressionMetadata.java:106)
>>> ~[apache-cassandra-2.1.9.jar:2.1.9]
>>> ... 14 common frames omitted
>>>
>>>
>>> I found some people recommending scrubbing the sstable so I attempted
>>> that and got the following error:
>>>
>>> bin/sstablescrub system sstable_activity -v
>>>
>>>
>>> ERROR 17:26:03 Exiting forcefully due to file system exception on
>>> startup, disk failure policy "stop"
>>> org.apache.cassandra.io.sstable.CorruptSSTableException:
>>> java.io.EOFE

Re: Repair corrupt SSTable from power outage?

2015-10-02 Thread Tyler Hobbs
What version of Cassandra are you running?  It sounds like the disk failure
policy is incorrectly being applied to scrub, which kind of defeats the
purpose of it.  I'd recommend opening a ticket (
https://issues.apache.org/jira/browse/CASSANDRA) with the information you
posted here.

On Fri, Oct 2, 2015 at 3:01 AM, George Sigletos 
wrote:

> I'm also facing problems regarding corrupt sstables and also couldn't run
> sstablescrub successfully.
>
> I restarted my nodes with disk failure policy "best_effort", then I run
> the "nodetool scrub "
>
> Once done I removed the corrupt tables manually and started repair
>
>
> On Thu, Oct 1, 2015 at 7:27 PM, John Anderson  wrote:
>
>> I have a 25 node cluster and we lost power on one of the racks last night
>> and now 6 of our nodes will not start up and we are getting the following
>> error:
>>
>> INFO  [main] 2015-10-01 10:19:22,111 CassandraDaemon.java:122 - JMX is
>> enabled to receive remote connections on port: 7199
>> INFO  [main] 2015-10-01 10:19:22,124 CacheService.java:111 - Initializing
>> key cache with capacity of 100 MBs.
>> INFO  [main] 2015-10-01 10:19:22,129 CacheService.java:133 - Initializing
>> row cache with capacity of 0 MBs
>> INFO  [main] 2015-10-01 10:19:22,133 CacheService.java:150 - Initializing
>> counter cache with capacity of 50 MBs
>> INFO  [main] 2015-10-01 10:19:22,135 CacheService.java:161 - Scheduling
>> counter cache save to every 7200 seconds (going to save all keys).
>> INFO  [main] 2015-10-01 10:19:22,211 ColumnFamilyStore.java:363 -
>> Initializing system.sstable_activity
>> INFO  [SSTableBatchOpen:1] 2015-10-01 10:19:22,639 SSTableReader.java:478
>> - Opening
>> /mnt/cassandra/data/data/system/sstable_activity-5a1ff267ace03f128563cfae6103c65e/system-sstable_activity-ka-45
>> (805 bytes)
>> ERROR [SSTableBatchOpen:1] 2015-10-01 10:19:22,657 FileUtils.java:447 -
>> Exiting forcefully due to file system exception on startup, disk failure
>> policy "stop"
>> org.apache.cassandra.io.sstable.CorruptSSTableException:
>> java.io.EOFException
>> at
>> org.apache.cassandra.io.compress.CompressionMetadata.(CompressionMetadata.java:131)
>> ~[apache-cassandra-2.1.9.jar:2.1.9]
>> at
>> org.apache.cassandra.io.compress.CompressionMetadata.create(CompressionMetadata.java:85)
>> ~[apache-cassandra-2.1.9.jar:2.1.9]
>> at
>> org.apache.cassandra.io.util.CompressedSegmentedFile$Builder.metadata(CompressedSegmentedFile.java:79)
>> ~[apache-cassandra-2.1.9.jar:2.1.9]
>> at
>> org.apache.cassandra.io.util.CompressedPoolingSegmentedFile$Builder.complete(CompressedPoolingSegmentedFile.java:72)
>> ~[apache-cassandra-2.1.9.jar:2.1.9]
>> at
>> org.apache.cassandra.io.util.SegmentedFile$Builder.complete(SegmentedFile.java:168)
>> ~[apache-cassandra-2.1.9.jar:2.1.9]
>> at
>> org.apache.cassandra.io.sstable.SSTableReader.load(SSTableReader.java:752)
>> ~[apache-cassandra-2.1.9.jar:2.1.9]
>> at
>> org.apache.cassandra.io.sstable.SSTableReader.load(SSTableReader.java:703)
>> ~[apache-cassandra-2.1.9.jar:2.1.9]
>> at
>> org.apache.cassandra.io.sstable.SSTableReader.open(SSTableReader.java:491)
>> ~[apache-cassandra-2.1.9.jar:2.1.9]
>> at
>> org.apache.cassandra.io.sstable.SSTableReader.open(SSTableReader.java:387)
>> ~[apache-cassandra-2.1.9.jar:2.1.9]
>> at
>> org.apache.cassandra.io.sstable.SSTableReader$4.run(SSTableReader.java:534)
>> ~[apache-cassandra-2.1.9.jar:2.1.9]
>> at
>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>> [na:1.8.0_60]
>> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>> [na:1.8.0_60]
>> at
>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>> [na:1.8.0_60]
>> at
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>> [na:1.8.0_60]
>> at java.lang.Thread.run(Thread.java:745) [na:1.8.0_60]
>> Caused by: java.io.EOFException: null
>> at
>> java.io.DataInputStream.readUnsignedShort(DataInputStream.java:340)
>> ~[na:1.8.0_60]
>> at java.io.DataInputStream.readUTF(DataInputStream.java:589)
>> ~[na:1.8.0_60]
>> at java.io.DataInputStream.readUTF(DataInputStream.java:564)
>> ~[na:1.8.0_60]
>> at
>> org.apache.cassandra.io.compress.CompressionMetadata.(CompressionMetadata.java:106)
>> ~[apache-cassandra-2.1.9.jar:2.1.9]
>> ... 14 common frames omitted
>>
>>
>> I found some people recommending scrubbing the sstable so I attempted
>> that and got the following error:
>>
>> bin/sstablescrub system sstable_activity -v
>>
>>
>> ERROR 17:26:03 Exiting forcefully due to file system exception on
>> startup, disk failure policy "stop"
>> org.apache.cassandra.io.sstable.CorruptSSTableException:
>> java.io.EOFException
>> at
>> org.apache.cassandra.io.compress.CompressionMetadata.(CompressionMetadata.java:131)
>> ~[apache-cassandra-2.1.9.jar:2.1.9]
>> at
>> org.apache.cassandra.io.

Re: Repair corrupt SSTable from power outage?

2015-10-02 Thread George Sigletos
I'm also facing problems regarding corrupt sstables and also couldn't run
sstablescrub successfully.

I restarted my nodes with disk failure policy "best_effort", then I run the
"nodetool scrub "

Once done I removed the corrupt tables manually and started repair


On Thu, Oct 1, 2015 at 7:27 PM, John Anderson  wrote:

> I have a 25 node cluster and we lost power on one of the racks last night
> and now 6 of our nodes will not start up and we are getting the following
> error:
>
> INFO  [main] 2015-10-01 10:19:22,111 CassandraDaemon.java:122 - JMX is
> enabled to receive remote connections on port: 7199
> INFO  [main] 2015-10-01 10:19:22,124 CacheService.java:111 - Initializing
> key cache with capacity of 100 MBs.
> INFO  [main] 2015-10-01 10:19:22,129 CacheService.java:133 - Initializing
> row cache with capacity of 0 MBs
> INFO  [main] 2015-10-01 10:19:22,133 CacheService.java:150 - Initializing
> counter cache with capacity of 50 MBs
> INFO  [main] 2015-10-01 10:19:22,135 CacheService.java:161 - Scheduling
> counter cache save to every 7200 seconds (going to save all keys).
> INFO  [main] 2015-10-01 10:19:22,211 ColumnFamilyStore.java:363 -
> Initializing system.sstable_activity
> INFO  [SSTableBatchOpen:1] 2015-10-01 10:19:22,639 SSTableReader.java:478
> - Opening
> /mnt/cassandra/data/data/system/sstable_activity-5a1ff267ace03f128563cfae6103c65e/system-sstable_activity-ka-45
> (805 bytes)
> ERROR [SSTableBatchOpen:1] 2015-10-01 10:19:22,657 FileUtils.java:447 -
> Exiting forcefully due to file system exception on startup, disk failure
> policy "stop"
> org.apache.cassandra.io.sstable.CorruptSSTableException:
> java.io.EOFException
> at
> org.apache.cassandra.io.compress.CompressionMetadata.(CompressionMetadata.java:131)
> ~[apache-cassandra-2.1.9.jar:2.1.9]
> at
> org.apache.cassandra.io.compress.CompressionMetadata.create(CompressionMetadata.java:85)
> ~[apache-cassandra-2.1.9.jar:2.1.9]
> at
> org.apache.cassandra.io.util.CompressedSegmentedFile$Builder.metadata(CompressedSegmentedFile.java:79)
> ~[apache-cassandra-2.1.9.jar:2.1.9]
> at
> org.apache.cassandra.io.util.CompressedPoolingSegmentedFile$Builder.complete(CompressedPoolingSegmentedFile.java:72)
> ~[apache-cassandra-2.1.9.jar:2.1.9]
> at
> org.apache.cassandra.io.util.SegmentedFile$Builder.complete(SegmentedFile.java:168)
> ~[apache-cassandra-2.1.9.jar:2.1.9]
> at
> org.apache.cassandra.io.sstable.SSTableReader.load(SSTableReader.java:752)
> ~[apache-cassandra-2.1.9.jar:2.1.9]
> at
> org.apache.cassandra.io.sstable.SSTableReader.load(SSTableReader.java:703)
> ~[apache-cassandra-2.1.9.jar:2.1.9]
> at
> org.apache.cassandra.io.sstable.SSTableReader.open(SSTableReader.java:491)
> ~[apache-cassandra-2.1.9.jar:2.1.9]
> at
> org.apache.cassandra.io.sstable.SSTableReader.open(SSTableReader.java:387)
> ~[apache-cassandra-2.1.9.jar:2.1.9]
> at
> org.apache.cassandra.io.sstable.SSTableReader$4.run(SSTableReader.java:534)
> ~[apache-cassandra-2.1.9.jar:2.1.9]
> at
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> [na:1.8.0_60]
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> [na:1.8.0_60]
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> [na:1.8.0_60]
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> [na:1.8.0_60]
> at java.lang.Thread.run(Thread.java:745) [na:1.8.0_60]
> Caused by: java.io.EOFException: null
> at
> java.io.DataInputStream.readUnsignedShort(DataInputStream.java:340)
> ~[na:1.8.0_60]
> at java.io.DataInputStream.readUTF(DataInputStream.java:589)
> ~[na:1.8.0_60]
> at java.io.DataInputStream.readUTF(DataInputStream.java:564)
> ~[na:1.8.0_60]
> at
> org.apache.cassandra.io.compress.CompressionMetadata.(CompressionMetadata.java:106)
> ~[apache-cassandra-2.1.9.jar:2.1.9]
> ... 14 common frames omitted
>
>
> I found some people recommending scrubbing the sstable so I attempted that
> and got the following error:
>
> bin/sstablescrub system sstable_activity -v
>
>
> ERROR 17:26:03 Exiting forcefully due to file system exception on startup,
> disk failure policy "stop"
> org.apache.cassandra.io.sstable.CorruptSSTableException:
> java.io.EOFException
> at
> org.apache.cassandra.io.compress.CompressionMetadata.(CompressionMetadata.java:131)
> ~[apache-cassandra-2.1.9.jar:2.1.9]
> at
> org.apache.cassandra.io.compress.CompressionMetadata.create(CompressionMetadata.java:85)
> ~[apache-cassandra-2.1.9.jar:2.1.9]
> at
> org.apache.cassandra.io.util.CompressedSegmentedFile$Builder.metadata(CompressedSegmentedFile.java:79)
> ~[apache-cassandra-2.1.9.jar:2.1.9]
> at
> org.apache.cassandra.io.util.CompressedPoolingSegmentedFile$Builder.complete(CompressedPoolingSegmentedFile.java:72)
> ~[apache-cassandra-2.1.9.jar:2.1.9]
> at
> org.apache.

Repair corrupt SSTable from power outage?

2015-10-01 Thread John Anderson
I have a 25 node cluster and we lost power on one of the racks last night
and now 6 of our nodes will not start up and we are getting the following
error:

INFO  [main] 2015-10-01 10:19:22,111 CassandraDaemon.java:122 - JMX is
enabled to receive remote connections on port: 7199
INFO  [main] 2015-10-01 10:19:22,124 CacheService.java:111 - Initializing
key cache with capacity of 100 MBs.
INFO  [main] 2015-10-01 10:19:22,129 CacheService.java:133 - Initializing
row cache with capacity of 0 MBs
INFO  [main] 2015-10-01 10:19:22,133 CacheService.java:150 - Initializing
counter cache with capacity of 50 MBs
INFO  [main] 2015-10-01 10:19:22,135 CacheService.java:161 - Scheduling
counter cache save to every 7200 seconds (going to save all keys).
INFO  [main] 2015-10-01 10:19:22,211 ColumnFamilyStore.java:363 -
Initializing system.sstable_activity
INFO  [SSTableBatchOpen:1] 2015-10-01 10:19:22,639 SSTableReader.java:478 -
Opening
/mnt/cassandra/data/data/system/sstable_activity-5a1ff267ace03f128563cfae6103c65e/system-sstable_activity-ka-45
(805 bytes)
ERROR [SSTableBatchOpen:1] 2015-10-01 10:19:22,657 FileUtils.java:447 -
Exiting forcefully due to file system exception on startup, disk failure
policy "stop"
org.apache.cassandra.io.sstable.CorruptSSTableException:
java.io.EOFException
at
org.apache.cassandra.io.compress.CompressionMetadata.(CompressionMetadata.java:131)
~[apache-cassandra-2.1.9.jar:2.1.9]
at
org.apache.cassandra.io.compress.CompressionMetadata.create(CompressionMetadata.java:85)
~[apache-cassandra-2.1.9.jar:2.1.9]
at
org.apache.cassandra.io.util.CompressedSegmentedFile$Builder.metadata(CompressedSegmentedFile.java:79)
~[apache-cassandra-2.1.9.jar:2.1.9]
at
org.apache.cassandra.io.util.CompressedPoolingSegmentedFile$Builder.complete(CompressedPoolingSegmentedFile.java:72)
~[apache-cassandra-2.1.9.jar:2.1.9]
at
org.apache.cassandra.io.util.SegmentedFile$Builder.complete(SegmentedFile.java:168)
~[apache-cassandra-2.1.9.jar:2.1.9]
at
org.apache.cassandra.io.sstable.SSTableReader.load(SSTableReader.java:752)
~[apache-cassandra-2.1.9.jar:2.1.9]
at
org.apache.cassandra.io.sstable.SSTableReader.load(SSTableReader.java:703)
~[apache-cassandra-2.1.9.jar:2.1.9]
at
org.apache.cassandra.io.sstable.SSTableReader.open(SSTableReader.java:491)
~[apache-cassandra-2.1.9.jar:2.1.9]
at
org.apache.cassandra.io.sstable.SSTableReader.open(SSTableReader.java:387)
~[apache-cassandra-2.1.9.jar:2.1.9]
at
org.apache.cassandra.io.sstable.SSTableReader$4.run(SSTableReader.java:534)
~[apache-cassandra-2.1.9.jar:2.1.9]
at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
[na:1.8.0_60]
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
[na:1.8.0_60]
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
[na:1.8.0_60]
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
[na:1.8.0_60]
at java.lang.Thread.run(Thread.java:745) [na:1.8.0_60]
Caused by: java.io.EOFException: null
at
java.io.DataInputStream.readUnsignedShort(DataInputStream.java:340)
~[na:1.8.0_60]
at java.io.DataInputStream.readUTF(DataInputStream.java:589)
~[na:1.8.0_60]
at java.io.DataInputStream.readUTF(DataInputStream.java:564)
~[na:1.8.0_60]
at
org.apache.cassandra.io.compress.CompressionMetadata.(CompressionMetadata.java:106)
~[apache-cassandra-2.1.9.jar:2.1.9]
... 14 common frames omitted


I found some people recommending scrubbing the sstable so I attempted that
and got the following error:

bin/sstablescrub system sstable_activity -v


ERROR 17:26:03 Exiting forcefully due to file system exception on startup,
disk failure policy "stop"
org.apache.cassandra.io.sstable.CorruptSSTableException:
java.io.EOFException
at
org.apache.cassandra.io.compress.CompressionMetadata.(CompressionMetadata.java:131)
~[apache-cassandra-2.1.9.jar:2.1.9]
at
org.apache.cassandra.io.compress.CompressionMetadata.create(CompressionMetadata.java:85)
~[apache-cassandra-2.1.9.jar:2.1.9]
at
org.apache.cassandra.io.util.CompressedSegmentedFile$Builder.metadata(CompressedSegmentedFile.java:79)
~[apache-cassandra-2.1.9.jar:2.1.9]
at
org.apache.cassandra.io.util.CompressedPoolingSegmentedFile$Builder.complete(CompressedPoolingSegmentedFile.java:72)
~[apache-cassandra-2.1.9.jar:2.1.9]
at
org.apache.cassandra.io.util.SegmentedFile$Builder.complete(SegmentedFile.java:168)
~[apache-cassandra-2.1.9.jar:2.1.9]
at
org.apache.cassandra.io.sstable.SSTableReader.load(SSTableReader.java:752)
~[apache-cassandra-2.1.9.jar:2.1.9]
at
org.apache.cassandra.io.sstable.SSTableReader.load(SSTableReader.java:703)
~[apache-cassandra-2.1.9.jar:2.1.9]
at
org.apache.cassandra.io.sstable.SSTableReader.open(SSTableReader.java:491)
~[apache-cassandra-2.1.9.jar:2.1.9]
at
org.apache.cassandra.io.sstable.SSTable