[ https://issues.apache.org/jira/browse/ARROW-462?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Wes McKinney updated ARROW-462: ------------------------------- Description: We use a hash table to extract unique values and dictionary indices. There may be an opportunity to consolidate common code from the dictionary encoding implementation implemented in parquet-cpp (but the dictionary indices will not be run-length encoded in Arrow): https://github.com/apache/parquet-cpp/blob/master/src/parquet/encodings/dictionary-encoding.h This functionality also needs to permit encoding split across multiple record batches -- so the hash table would be a stateful entity, and we can continue to hash more chunks of data to dictionary-encode multiple arrays with a shared dictionary at the end. was: We use a hash table to extract unique values and dictionary indices. There may be an opportunity to consolidate common code from the dictionary encoding implementation implemented in parquet-cpp (but the dictionary indices will not be run-length encoded in Arrow): https://github.com/apache/parquet-cpp/blob/master/src/parquet/encodings/dictionary-encoding.h > [C++] Implement in-memory conversions between non-nested primitive types and > DictionaryArray equivalent > ------------------------------------------------------------------------------------------------------- > > Key: ARROW-462 > URL: https://issues.apache.org/jira/browse/ARROW-462 > Project: Apache Arrow > Issue Type: New Feature > Components: C++ > Reporter: Wes McKinney > > We use a hash table to extract unique values and dictionary indices. There > may be an opportunity to consolidate common code from the dictionary encoding > implementation implemented in parquet-cpp (but the dictionary indices will > not be run-length encoded in Arrow): > https://github.com/apache/parquet-cpp/blob/master/src/parquet/encodings/dictionary-encoding.h > This functionality also needs to permit encoding split across multiple record > batches -- so the hash table would be a stateful entity, and we can continue > to hash more chunks of data to dictionary-encode multiple arrays with a > shared dictionary at the end. -- This message was sent by Atlassian JIRA (v6.3.4#6332)