diffent behavior for escape character backslash when bulk loading data seems there's diffent behavior for escape character between Mapreduce JsonBulkLoadTool and psql.py
if lines end with backslash \ , psql.py can load it without any error, Mapreduce JsonBulkLoadTool fails with error java.io.IOException: EOF whilst processing escape sequence, even with -ignore-error parameter if lines end with double backslash \\, psql.py can load it without any error, as \\ but Mapreduce JsonBulkLoadTool can load it differently , as \ so, the two loading tools work differently? and I think the way psql.py is more better. it can load data ignoring any errors, and load data as its original state. is there any method or parameter to make the Mapreduce loader work as psql.py ? -------------------------------------------------- echo \\\\>a.csv cat a.csv \\ hdfs dfs -rm a.csv hdfs dfs -put a.csv ...JsonBulkLoadTool -g -t a -i a.csv phoenixload -g -t a -i a.csv -- success select * from a; \ psql.py -t A zoo a.csv select * from a; \\ --------------------------------------------------