The data is okay, because the exact same Parquet directory is working fine
on the local drive, it's just not working when using HDFS.  I tried casting
as you said, but that ended up with the exact same problem.


Regards,


*Adam Gilmore*

Director of Technology

a...@pharmadata.net.au


+61 421 997 655 (Mobile)

1300 733 876 (AU)

+617 3171 9902 (Intl)


*PharmaData*

Data Intelligence Solutions for Pharmacy

www.PharmaData.net.au <http://www.pharmadata.net.au/>



[image: pharmadata-sig]



*Disclaimer*

This communication including any attachments may contain information that
is either confidential or otherwise protected from disclosure and is
intended solely for the use of the intended recipient. If you are not the
intended recipient please immediately notify the sender by e-mail and
delete the original transmission and its contents. Any unauthorised use,
dissemination, forwarding, printing, or copying of this communication
including any file attachments is prohibited. The recipient should check
this email and any attachments for viruses and other defects. The Company
disclaims any liability for loss or damage arising in any way from this
communication including any file attachments.

On Tue, Jan 6, 2015 at 9:49 AM, MapR <sth...@maprtech.com> wrote:

> Please try casting the colum data type. Also please verify that all the
> column data is satisfying your data type.
>
> Sudhakar Thota
> Sent from my iPhone
>
> > On Jan 5, 2015, at 5:56 AM, Adam Gilmore <dragoncu...@gmail.com> wrote:
> >
> > The actual stack trace is:
> >
> > 2015-01-05 13:48:27,356 [2b5569d5-3771-748d-1390-3a8930d02002:frag:1:12]
> > ERROR o.a.drill.exec.ops.FragmentContext - Fragment Context received
> > failure.
> > org.apache.drill.common.exceptions.DrillRuntimeException:
> > java.io.IOException: can not read class parquet.format.PageHeader: don't
> > know what type: 13
> >        at
> >
> org.apache.drill.exec.store.parquet.columnreaders.ParquetRecordReader.next(ParquetRecordReader.java:427)
> > ~[drill-java-exec-0.7.0-rebuffed.jar:0.7.0]
> >        at
> > org.apache.drill.exec.physical.impl.ScanBatch.next(ScanBatch.java:158)
> > ~[drill-java-exec-0.7.0-rebuffed.jar:0.7.0]
> >        at
> >
> org.apache.drill.exec.physical.impl.validate.IteratorValidatorBatchIterator.next(IteratorValidatorBatchIterator.java:118)
> > [drill-java-exec-0.7.0-rebuffed.jar:0.7.0]
> >        at
> >
> org.apache.drill.exec.record.AbstractRecordBatch.next(AbstractRecordBatch.java:99)
> > [drill-java-exec-0.7.0-rebuffed.jar:0.7.0]
> >        at
> >
> org.apache.drill.exec.record.AbstractRecordBatch.next(AbstractRecordBatch.java:89)
> > [drill-java-exec-0.7.0-rebuffed.jar:0.7.0]
> >        at
> >
> org.apache.drill.exec.physical.impl.aggregate.StreamingAggBatch.buildSchema(StreamingAggBatch.java:83)
> > [drill-java-exec-0.7.0-rebuffed.jar:0.7.0]
> >        at
> >
> org.apache.drill.exec.record.AbstractRecordBatch.next(AbstractRecordBatch.java:130)
> > [drill-java-exec-0.7.0-rebuffed.jar:0.7.0]
> >        at
> >
> org.apache.drill.exec.physical.impl.validate.IteratorValidatorBatchIterator.next(IteratorValidatorBatchIterator.java:118)
> > [drill-java-exec-0.7.0-rebuffed.jar:0.7.0]
> >        at
> >
> org.apache.drill.exec.physical.impl.BaseRootExec.next(BaseRootExec.java:67)
> > [drill-java-exec-0.7.0-rebuffed.jar:0.7.0]
> >        at
> >
> org.apache.drill.exec.physical.impl.SingleSenderCreator$SingleSenderRootExec.innerNext(SingleSenderCreator.java:97)
> > [drill-java-exec-0.7.0-rebuffed.jar:0.7.0]
> >        at
> >
> org.apache.drill.exec.physical.impl.BaseRootExec.next(BaseRootExec.java:57)
> > [drill-java-exec-0.7.0-rebuffed.jar:0.7.0]
> >        at
> >
> org.apache.drill.exec.work.fragment.FragmentExecutor.run(FragmentExecutor.java:114)
> > [drill-java-exec-0.7.0-rebuffed.jar:0.7.0]
> >        at
> >
> org.apache.drill.exec.work.WorkManager$RunnableWrapper.run(WorkManager.java:254)
> > [drill-java-exec-0.7.0-rebuffed.jar:0.7.0]
> >        at
> >
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> > [na:1.7.0_71]
> >        at
> >
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> > [na:1.7.0_71]
> >        at java.lang.Thread.run(Thread.java:745) [na:1.7.0_71]
> > Caused by: java.io.IOException: can not read class
> > parquet.format.PageHeader: don't know what type: 13
> >        at parquet.format.Util.read(Util.java:50)
> > ~[parquet-format-2.1.1-drill-r1.jar:na]
> >        at parquet.format.Util.readPageHeader(Util.java:26)
> > ~[parquet-format-2.1.1-drill-r1.jar:na]
> >        at
> >
> org.apache.drill.exec.store.parquet.ColumnDataReader.readPageHeader(ColumnDataReader.java:47)
> > ~[drill-java-exec-0.7.0-rebuffed.jar:0.7.0]
> >        at
> >
> org.apache.drill.exec.store.parquet.columnreaders.PageReader.next(PageReader.java:169)
> > ~[drill-java-exec-0.7.0-rebuffed.jar:0.7.0]
> >        at
> >
> org.apache.drill.exec.store.parquet.columnreaders.NullableColumnReader.processPages(NullableColumnReader.java:76)
> > ~[drill-java-exec-0.7.0-rebuffed.jar:0.7.0]
> >        at
> >
> org.apache.drill.exec.store.parquet.columnreaders.ParquetRecordReader.readAllFixedFields(ParquetRecordReader.java:366)
> > ~[drill-java-exec-0.7.0-rebuffed.jar:0.7.0]
> >        at
> >
> org.apache.drill.exec.store.parquet.columnreaders.ParquetRecordReader.next(ParquetRecordReader.java:409)
> > ~[drill-java-exec-0.7.0-rebuffed.jar:0.7.0]
> >        ... 15 common frames omitted
> > Caused by: parquet.org.apache.thrift.protocol.TProtocolException: don't
> > know what type: 13
> >        at
> >
> parquet.org.apache.thrift.protocol.TCompactProtocol.getTType(TCompactProtocol.java:806)
> > ~[parquet-format-2.1.1-drill-r1.jar:na]
> >        at
> >
> parquet.org.apache.thrift.protocol.TCompactProtocol.readListBegin(TCompactProtocol.java:536)
> > ~[parquet-format-2.1.1-drill-r1.jar:na]
> >        at
> >
> parquet.org.apache.thrift.protocol.TCompactProtocol.readSetBegin(TCompactProtocol.java:547)
> > ~[parquet-format-2.1.1-drill-r1.jar:na]
> >        at
> >
> parquet.org.apache.thrift.protocol.TProtocolUtil.skip(TProtocolUtil.java:128)
> > ~[parquet-format-2.1.1-drill-r1.jar:na]
> >        at
> >
> parquet.org.apache.thrift.protocol.TProtocolUtil.skip(TProtocolUtil.java:60)
> > ~[parquet-format-2.1.1-drill-r1.jar:na]
> >        at parquet.format.PageHeader.read(PageHeader.java:897)
> > ~[parquet-format-2.1.1-drill-r1.jar:na]
> >        at parquet.format.Util.read(Util.java:47)
> > ~[parquet-format-2.1.1-drill-r1.jar:na]
> >        ... 21 common frames omitted
> >
> >
> >> On Mon, Jan 5, 2015 at 6:26 PM, Adam Gilmore <dragoncu...@gmail.com>
> wrote:
> >>
> >> Hi all,
> >>
> >> I'm trying to do a really simple query on a parquet directory on HDFS.
> >>
> >> This works fine:
> >>
> >> select count(*) from hdfs.warehouse.saleparquet
> >>
> >> However, this fails:
> >>
> >> 0: jdbc:drill:local> select sum(sellprice) from
> hdfs.warehouse.saleparquet;
> >> Query failed: Query failed: Failure while running fragment., You tried
> to
> >> do a batch data read operation when you were in a state of STOP.  You
> can
> >> only do this type of operation when you are in a state of OK or
> >> OK_NEW_SCHEMA. [ 92fc8807-220b-466c-bbac-1f524d4251cb on
> >> ip-10-8-1-154.ap-southeast-2.compute.internal:31010 ]
> >> [ 92fc8807-220b-466c-bbac-1f524d4251cb on
> >> ip-10-8-1-154.ap-southeast-2.compute.internal:31010 ]
> >>
> >>
> >> Error: exception while executing query: Failure while executing query.
> >> (state=,code=0)
> >>
> >> Seems like a very simple query.
> >>
> >> Funnily enough, if I copy it off HDFS to the local system and run the
> same
> >> query against the local file, it works fine.  Just purely something to
> do
> >> with HDFS.
> >>
> >> Any ideas?  I'm running 0.7.
> >>
>

Reply via email to