k8ika0s commented on PR #48195:
URL: https://github.com/apache/arrow/pull/48195#issuecomment-3568424755

   The patch tackles the same corner of the bit-stream utilities, but the 
handling on big-endian ends up taking a pretty different route. In this PR, the 
generic unpack paths stay active on BE, with the cached-word machinery still 
steering most reads. That works fine on little-endian, but those helpers lean 
on assumptions about word layout that are trickier to uphold on BE even with 
the surrounding byte-swaps.
   
   In my version, the BE code path steps around those assumptions entirely: VLQ 
parsing pulls straight from the underlying buffer, and the bulk bit extraction 
uses the simpler, portable reader rather than the wide 32/64-bit fast paths. It 
costs a few cycles, but it keeps the behavior identical across hosts without 
depending on how the cached words line up.
   
   Nothing here looks wildly off, but you can see the philosophical split: this 
PR keeps the optimized hot paths alive everywhere, while the alternative 
narrows the surface area on BE so the byte order never has a chance to get 
involved.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to