[ https://issues.apache.org/jira/browse/ARROW-785?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16976721#comment-16976721 ]
albertoramon commented on ARROW-785: ------------------------------------ I saw this:(SparkSQL 2.4.4 PyArrow 0.15) The problem is Create table with INT columns (BIGINT works properly) SOL: Change INT to BIGINT works fine (I tried to use Double but didn't work) in create table In my case: these Parquet Files are from SSB benchmark {code:java} SELECT MAX(LO_CUSTKEY), MAX(LO_PARTKEY), MAX (LO_SUPPKEY) FROM SSB.LINEORDER; Returns: 29999 200000 2000 {code} In my Column_Types I Had, thus I need review my Python Code :) : {code:java} 'lo_custkey':'int64', 'lo_partkey':'int64', 'lo_suppkey':'int64',{code} > possible issue on writing parquet via pyarrow, subsequently read in Hive > ------------------------------------------------------------------------ > > Key: ARROW-785 > URL: https://issues.apache.org/jira/browse/ARROW-785 > Project: Apache Arrow > Issue Type: Bug > Components: Python > Reporter: Jeff Reback > Assignee: Wes McKinney > Priority: Minor > Fix For: 0.5.0 > > > details here: > http://stackoverflow.com/questions/43268872/parquet-creation-conversion-from-pandas-dataframe-to-pyarrow-table-not-working-f > This round trips in pandas->parquet->pandas just fine on released pandas > (0.19.2) and pyarrow (0.2). > OP stats that it is not readable in Hive however. -- This message was sent by Atlassian Jira (v8.3.4#803005)