Jack Fan created ARROW-7841: ------------------------------- Summary: pyarrow release 0.16.0 breaks `libhdfs.so` loading mechanism Key: ARROW-7841 URL: https://issues.apache.org/jira/browse/ARROW-7841 Project: Apache Arrow Issue Type: Bug Components: Python Affects Versions: 0.16.0 Reporter: Jack Fan Fix For: 0.15.1
I have my env variable setup correctly according to the pyarrow README {code:java} $ ls $HADOOP_HOME/lib/native libhadoop.a libhadooppipes.a libhadoop.so libhadoop.so.1.0.0 libhadooputils.a libhdfs.a libhdfs.so libhdfs.so.0.0.0 {code} Use the following script to reproduce {code:java} import pyarrow pyarrow.hdfs.connect('hdfs://localhost'){code} With pyarrow version 0.15.1 it is fine. However, version 0.16.0 will give error {code:java} Traceback (most recent call last): File "<string>", line 2, in <module> File "/home/jackwindows/anaconda2/lib/python2.7/site-packages/pyarrow/hdfs.py", line 215, in connect extra_conf=extra_conf) File "/home/jackwindows/anaconda2/lib/python2.7/site-packages/pyarrow/hdfs.py", line 40, in __init__ self._connect(host, port, user, kerb_ticket, driver, extra_conf) File "pyarrow/io-hdfs.pxi", line 89, in pyarrow.lib.HadoopFileSystem._connect File "pyarrow/error.pxi", line 99, in pyarrow.lib.check_status IOError: Unable to load libhdfs: /opt/hadoop/latest/libhdfs.so: cannot open shared object file: No such file or directory {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)