https://issues.apache.org/jira/browse/DRILL-4317
On 26 Jan 2016, at 23:50, Abdel Hakim Deneche wrote:
This definitely looks like a bug, could you open a JIRA and share as
much
information as possible about the structure of the CSV file and the
number
of records.
On Tue, Jan 26, 2016 at 7:38 PM, Matt <bsg...@gmail.com> wrote:
The CTAS with fails with:
~~~
Error: SYSTEM ERROR: IllegalArgumentException: length: -260
(expected: >=
0)
Fragment 1:2
[Error Id: 1807615e-4385-4f85-8402-5900aaa568e9 on es07:31010]
(java.lang.IllegalArgumentException) length: -260 (expected: >= 0)
io.netty.buffer.AbstractByteBuf.checkIndex():1131
io.netty.buffer.PooledUnsafeDirectByteBuf.nioBuffer():344
io.netty.buffer.WrappedByteBuf.nioBuffer():727
io.netty.buffer.UnsafeDirectLittleEndian.nioBuffer():26
io.netty.buffer.DrillBuf.nioBuffer():356
org.apache.drill.exec.store.ParquetOutputRecordWriter$VarCharParquetConverter.writeField():1842
org.apache.drill.exec.store.EventBasedRecordWriter.write():62
org.apache.drill.exec.physical.impl.WriterRecordBatch.innerNext():106
org.apache.drill.exec.record.AbstractRecordBatch.next():162
org.apache.drill.exec.physical.impl.BaseRootExec.next():104
org.apache.drill.exec.physical.impl.SingleSenderCreator$SingleSenderRootExec.innerNext():93
org.apache.drill.exec.physical.impl.BaseRootExec.next():94
org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():256
org.apache.drill.exec.work.fragment.FragmentExecutor$1.run():250
java.security.AccessController.doPrivileged():-2
javax.security.auth.Subject.doAs():415
org.apache.hadoop.security.UserGroupInformation.doAs():1657
org.apache.drill.exec.work.fragment.FragmentExecutor.run():250
org.apache.drill.common.SelfCleaningRunnable.run():38
java.util.concurrent.ThreadPoolExecutor.runWorker():1145
java.util.concurrent.ThreadPoolExecutor$Worker.run():615
java.lang.Thread.run():745 (state=,code=0)
~~~
And a simple SELECT * fails with:
~~~
java.lang.IndexOutOfBoundsException: index: 547681, length: 1
(expected:
range(0, 547681))
at
io.netty.buffer.AbstractByteBuf.checkIndex(AbstractByteBuf.java:1134)
at
io.netty.buffer.PooledUnsafeDirectByteBuf.getBytes(PooledUnsafeDirectByteBuf.java:136)
at
io.netty.buffer.WrappedByteBuf.getBytes(WrappedByteBuf.java:289)
at
io.netty.buffer.UnsafeDirectLittleEndian.getBytes(UnsafeDirectLittleEndian.java:26)
at io.netty.buffer.DrillBuf.getBytes(DrillBuf.java:586)
at io.netty.buffer.DrillBuf.getBytes(DrillBuf.java:586)
at io.netty.buffer.DrillBuf.getBytes(DrillBuf.java:586)
at io.netty.buffer.DrillBuf.getBytes(DrillBuf.java:586)
at
org.apache.drill.exec.vector.VarCharVector$Accessor.get(VarCharVector.java:443)
at
org.apache.drill.exec.vector.accessor.VarCharAccessor.getBytes(VarCharAccessor.java:125)
at
org.apache.drill.exec.vector.accessor.VarCharAccessor.getString(VarCharAccessor.java:146)
at
org.apache.drill.exec.vector.accessor.VarCharAccessor.getObject(VarCharAccessor.java:136)
at
org.apache.drill.exec.vector.accessor.VarCharAccessor.getObject(VarCharAccessor.java:94)
at
org.apache.drill.exec.vector.accessor.BoundCheckingAccessor.getObject(BoundCheckingAccessor.java:148)
at
org.apache.drill.jdbc.impl.TypeConvertingSqlAccessor.getObject(TypeConvertingSqlAccessor.java:795)
at
org.apache.drill.jdbc.impl.AvaticaDrillSqlAccessor.getObject(AvaticaDrillSqlAccessor.java:179)
at
net.hydromatic.avatica.AvaticaResultSet.getObject(AvaticaResultSet.java:351)
at
org.apache.drill.jdbc.impl.DrillResultSetImpl.getObject(DrillResultSetImpl.java:420)
at sqlline.Rows$Row.<init>(Rows.java:157)
at sqlline.IncrementalRows.hasNext(IncrementalRows.java:63)
at
sqlline.TableOutputFormat$ResizingRowsProvider.next(TableOutputFormat.java:87)
at sqlline.TableOutputFormat.print(TableOutputFormat.java:118)
at sqlline.SqlLine.print(SqlLine.java:1593)
at sqlline.Commands.execute(Commands.java:852)
at sqlline.Commands.sql(Commands.java:751)
at sqlline.SqlLine.dispatch(SqlLine.java:746)
at sqlline.SqlLine.begin(SqlLine.java:621)
at sqlline.SqlLine.start(SqlLine.java:375)
at sqlline.SqlLine.main(SqlLine.java:268)
~~~
It also looks like if I run the SELECT from a bash shell as "sqlline
-u
... -f test.sql 2>&1 > test.out" upon the error the sqlline session
"locks
up". No errors spool to the out file and the Java thread can only be
terminated with a kill -9. It can be backgrounded with ^z, but won't
respond to a ^c.
On 26 Jan 2016, at 14:07, Abdel Hakim Deneche wrote:
It's an internal buffer index. Can you try enabling verbose errors
and run
the query again, this should provide us with more details about the
error.
You can enable verbose error by running the following before the
select *:
alter session set `exec.errors.verbose`=true;
thanks
On Tue, Jan 26, 2016 at 11:01 AM, Matt <bsg...@gmail.com> wrote:
Putting the "select * from
`/csv/customer/hourly/customer_201510170000.csv`;" in a local .sql
file,
and executing it with sqlline > /dev/null (to avoid a ton of
scrolling)
results in:
~~~
index: 418719, length: 2 (expected: range(0, 418719))
Aborting
command
set because "force" is false and command failed: "select * from
`/csv/customer/hourly/customer_201510170000.csv`;"
Closing: org.apache.drill.jdbc.impl.DrillConnectionImpl
~~~
Is that index a byte or line offset?
On 26 Jan 2016, at 12:55, Abdel Hakim Deneche wrote:
Does a select * on the same data also fail ?
On Tue, Jan 26, 2016 at 9:44 AM, Matt <bsg...@gmail.com> wrote:
Getting some errors when attempting to create Parquet files from
CSV
data,
and trying to determine if it is due to the format of the source
data.
Its a fairly simple format of
"datetime,key,key,key,numeric,numeric,numeric, ..." with 32 of
those
numeric columns in total.
The source data does contain a lot missing values for the numeric
columns,
and those are represented by as consecutive delimiters:
""datetime,key,key,key,numeric,,,,,,..."
Could this be causing the CTAS to fail with these types of
errors? Or
is
there another cause to look for?
~~~
Error: SYSTEM ERROR: IllegalArgumentException: length: -260
(expected:
=
0)
│·······························································
│·······························································
Fragment 1:2
~~~
--
Abdelhakim Deneche
Software Engineer
<http://www.mapr.com/>
Now Available - Free Hadoop On-Demand Training
<
http://www.mapr.com/training?utm_source=Email&utm_medium=Signature&utm_campaign=Free%20available
--
Abdelhakim Deneche
Software Engineer
<http://www.mapr.com/>
Now Available - Free Hadoop On-Demand Training
<
http://www.mapr.com/training?utm_source=Email&utm_medium=Signature&utm_campaign=Free%20available
--
Abdelhakim Deneche
Software Engineer
<http://www.mapr.com/>
Now Available - Free Hadoop On-Demand Training
<http://www.mapr.com/training?utm_source=Email&utm_medium=Signature&utm_campaign=Free%20available>