rustyrazorblade commented on code in PR #3606:
URL: https://github.com/apache/cassandra/pull/3606#discussion_r1850646455
##########
src/java/org/apache/cassandra/io/util/CompressedChunkReader.java:
##########
@@ -117,8 +141,23 @@ public void readChunk(long position, ByteBuffer
uncompressed)
{
ByteBuffer compressed = bufferHolder.getBuffer(length);
- if (channel.read(compressed, chunk.offset) != length)
- throw new CorruptBlockException(channel.filePath(),
chunk);
+ if (readAheadBuffer != null && readAheadBuffer.hasBuffer())
+ {
+ int copied = 0;
+ while (copied < length) {
Review Comment:
Hmm... hard for me to say. I was previously under the impression that there
wasn't a good reason to use a larger size for LZ4 because it had some upper
limit on what it would actually compress, but when I looked at the docs it gave
me the impression that it can handle fairly large buffers, so maybe there's a
good reason to use it. If you've got some fairly large partitions and you're
only slicing them, there could be a benefit to it. I lean towards no warning,
but if you feel strongly about it I'm not opposed. I don't know what action a
user would take if they have a valid use case for the large buffer, and there
shouldn't be any real perf loss from not hitting the internal RA code path if
the reads are large enough.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]