[jira] [Commented] (HIVE-9854) OutofMemory while read ORCFile table

2016-06-19 Thread bin wang (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-9854?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15338941#comment-15338941
 ] 

bin wang commented on HIVE-9854:


how to fix this?

> OutofMemory while read ORCFile table
> 
>
> Key: HIVE-9854
> URL: https://issues.apache.org/jira/browse/HIVE-9854
> Project: Hive
>  Issue Type: Bug
>  Components: Serializers/Deserializers
>Affects Versions: 0.13.1
>Reporter: Liao, Xiaoge
>
> Log:
> Diagnostic Messages for this Task:
> Error: java.io.IOException: java.lang.reflect.InvocationTargetException
>   at 
> org.apache.hadoop.hive.io.HiveIOExceptionHandlerChain.handleRecordReaderCreationException(HiveIOExceptionHandlerChain.java:97)
>   at 
> org.apache.hadoop.hive.io.HiveIOExceptionHandlerUtil.handleRecordReaderCreationException(HiveIOExceptionHandlerUtil.java:57)
>   at 
> org.apache.hadoop.hive.shims.HadoopShimsSecure$CombineFileRecordReader.initNextRecordReader(HadoopShimsSecure.java:294)
>   at 
> org.apache.hadoop.hive.shims.HadoopShimsSecure$CombineFileRecordReader.(HadoopShimsSecure.java:241)
>   at 
> org.apache.hadoop.hive.shims.HadoopShimsSecure$CombineFileInputFormatShim.getRecordReader(HadoopShimsSecure.java:365)
>   at 
> org.apache.hadoop.hive.ql.io.CombineHiveInputFormat.getRecordReader(CombineHiveInputFormat.java:591)
>   at 
> org.apache.hadoop.mapred.MapTask$TrackedRecordReader.(MapTask.java:166)
>   at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:407)
>   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:340)
>   at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:160)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:396)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1438)
>   at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:155)
> Caused by: java.lang.reflect.InvocationTargetException
>   at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
>   at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
>   at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
>   at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
>   at 
> org.apache.hadoop.hive.shims.HadoopShimsSecure$CombineFileRecordReader.initNextRecordReader(HadoopShimsSecure.java:280)
>   ... 11 more
> Caused by: java.lang.OutOfMemoryError: Java heap space
>   at 
> org.apache.hadoop.hive.ql.io.orc.DynamicByteArray.grow(DynamicByteArray.java:64)
>   at 
> org.apache.hadoop.hive.ql.io.orc.DynamicByteArray.readAll(DynamicByteArray.java:142)
>   at 
> org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl$StringDictionaryTreeReader.startStripe(RecordReaderImpl.java:1547)
>   at 
> org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl$StringTreeReader.startStripe(RecordReaderImpl.java:1337)
>   at 
> org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl$StructTreeReader.startStripe(RecordReaderImpl.java:1825)
>   at 
> org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.readStripe(RecordReaderImpl.java:2537)
>   at 
> org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.advanceStripe(RecordReaderImpl.java:2950)
>   at 
> org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.advanceToNextRow(RecordReaderImpl.java:2992)
>   at 
> org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.(RecordReaderImpl.java:284)
>   at 
> org.apache.hadoop.hive.ql.io.orc.ReaderImpl.rowsOptions(ReaderImpl.java:480)
>   at 
> org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.createReaderFromFile(OrcInputFormat.java:214)
>   at 
> org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$OrcRecordReader.(OrcInputFormat.java:146)
>   at 
> org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.getRecordReader(OrcInputFormat.java:997)
>   at 
> org.apache.hadoop.hive.ql.io.CombineHiveRecordReader.(CombineHiveRecordReader.java:65)
>   ... 16 more
> FAILED: Execution Error, return code 2 from 
> org.apache.hadoop.hive.ql.exec.mr.MapRedTask
> MapReduce Jobs Launched: 
> Stage-Stage-1: Map: 105   Cumulative CPU: 656.39 sec   HDFS Read: 4040094761 
> HDFS Write: 139 FAIL
> Total MapReduce CPU Time Spent: 10 minutes 56 seconds 390 msec



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-12874) dynamic partition insert project wrong column

2016-01-14 Thread bin wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-12874?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

bin wang updated HIVE-12874:

Affects Version/s: (was: 0.14.1)
   1.1.0

> dynamic partition insert project wrong column
> -
>
> Key: HIVE-12874
> URL: https://issues.apache.org/jira/browse/HIVE-12874
> Project: Hive
>  Issue Type: Bug
>  Components: SQL
>Affects Versions: 1.1.0
> Environment: hive 1.1.0-cdh5.4.8
>Reporter: bin wang
>Assignee: Alan Gates
>
> We have two table as below:
> create table  test (
> id bigint comment ' id',
> )
> PARTITIONED BY(etl_dt string)
> STORED AS ORC;
> create table  test1 (
> id bigint
> start_time int,
> )
> PARTITIONED BY(etl_dt string)
> STORED AS ORC;
> we use sql like below to import rows from test1 to test:
> insert overwrite table test PARTITION(etl_dt)
> select id
> ,from_unixtime(start_time,'-MM-dd') as  etl_dt
>   
>  
> from test1
> where test1.etl_dt='2016-01-12';
> but it behave wrong, it use test1.etl_dt as the test's partition value, not 
> the 'etl_dt' in select.
> We think it's a bug, anyone to fix it? 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)