Hello all,I encountered a problem while using hive, please ask you. The
following is the specific situation.
Platform: hive on spark
Error: java.io.IOException: Too many bytes before delimiter: 2147483648
Description: When using small files of the same format for testing, there is no
problem; but for all data (about 140G), the above error is reported.
Table creation statement:
DROP TABLE IF EXISTS test;
CREATE EXTERNAL TABLE test (
json string
)
ROW FORMAT DELIMITED
FIELDS TERMINATED BY '\ r \ n'
STORED AS TEXTFILE
LOCATION 'hdfs: / userdata / test';
The big data platform I use is TDH Community Edition 5.0.2.
Any help is much appreciated!