[ 
https://issues.apache.org/jira/browse/DRILL-1075?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14326276#comment-14326276
 ] 

Hema Kumar S edited comment on DRILL-1075 at 2/18/15 6:02 PM:
--------------------------------------------------------------

[~amitskatti]
I'm using cdh 4.5 and trying to store file in hdfs, (using create table), I 
added all the jars to drill classpath. I got below error. 

Java.lang.UnsupportedOperationException: This is supposed to be overridden by 
subclasses.
{quote}
 at 
com.google.protobuf.GeneratedMessage.getUnknownFields(GeneratedMessage.java:180)
 ~[protobuf-java-2.5.0.jar:na]
        at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$GetFileInfoRequestProto.getSerializedSize(ClientNamenodeProtocolProtos.java:30108)
 ~[hadoop-hdfs-2.0.0-cdh4.5.0.jar:na]
        at 
com.google.protobuf.AbstractMessageLite.toByteString(AbstractMessageLite.java:49)
 ~[protobuf-java-2.5.0.jar:na]
        at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.constructRpcRequest(ProtobufRpcEngine.java:149)
 ~[hadoop-common-2.0.0-cdh4.5.0.jar:na]
        at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:193)
 ~[hadoop-common-2.0.0-cdh4.5.0.jar:na]
        at com.sun.proxy.$Proxy33.getFileInfo(Unknown Source) ~[na:na]
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 
~[na:1.7.0_51]
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) 
~[na:1.7.0_51]
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 ~[na:1.7.0_51]
        at java.lang.reflect.Method.invoke(Method.java:606) ~[na:1.7.0_51]
        at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:164)
 ~[hadoop-common-2.0.0-cdh4.5.0.jar:na]
        at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:83)
 ~[hadoop-common-2.0.0-cdh4.5.0.jar:na]
        at com.sun.proxy.$Proxy33.getFileInfo(Unknown Source) ~[na:na]
        at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:629)
 ~[hadoop-hdfs-2.0.0-cdh4.5.0.jar:na]
        at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:1545) 
~[hadoop-hdfs-2.0.0-cdh4.5.0.jar:na]
        at 
org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:820)
 ~[hadoop-hdfs-2.0.0-cdh4.5.0.jar:na]
        at 
org.apache.drill.exec.store.dfs.shim.fallback.FallbackFileSystem.getFileStatus(FallbackFileSystem.java:84)
 ~[drill-java-exec-0.7.0-rebuffed.jar:0.7.0]
        at 
org.apache.drill.exec.store.dfs.FileSelection.create(FileSelection.java:135) 
~[drill-java-exec-0.7.0-rebuffed.jar:0.7.0]
        at 
org.apache.drill.exec.store.dfs.WorkspaceSchemaFactory.create(WorkspaceSchemaFactory.java:133)
 ~[drill-java-exec-0.7.0-rebuffed.jar:0.7.0]
{quote}


was (Author: hema.kumar):
[~amitskatti]
I'm trying to store file in hdfs, (using create table), I added all the jars to 
drill classpath. I got below error. 

Java.lang.UnsupportedOperationException: This is supposed to be overridden by 
subclasses.
{quote}
 at 
com.google.protobuf.GeneratedMessage.getUnknownFields(GeneratedMessage.java:180)
 ~[protobuf-java-2.5.0.jar:na]
        at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$GetFileInfoRequestProto.getSerializedSize(ClientNamenodeProtocolProtos.java:30108)
 ~[hadoop-hdfs-2.0.0-cdh4.5.0.jar:na]
        at 
com.google.protobuf.AbstractMessageLite.toByteString(AbstractMessageLite.java:49)
 ~[protobuf-java-2.5.0.jar:na]
        at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.constructRpcRequest(ProtobufRpcEngine.java:149)
 ~[hadoop-common-2.0.0-cdh4.5.0.jar:na]
        at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:193)
 ~[hadoop-common-2.0.0-cdh4.5.0.jar:na]
        at com.sun.proxy.$Proxy33.getFileInfo(Unknown Source) ~[na:na]
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 
~[na:1.7.0_51]
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) 
~[na:1.7.0_51]
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 ~[na:1.7.0_51]
        at java.lang.reflect.Method.invoke(Method.java:606) ~[na:1.7.0_51]
        at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:164)
 ~[hadoop-common-2.0.0-cdh4.5.0.jar:na]
        at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:83)
 ~[hadoop-common-2.0.0-cdh4.5.0.jar:na]
        at com.sun.proxy.$Proxy33.getFileInfo(Unknown Source) ~[na:na]
        at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:629)
 ~[hadoop-hdfs-2.0.0-cdh4.5.0.jar:na]
        at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:1545) 
~[hadoop-hdfs-2.0.0-cdh4.5.0.jar:na]
        at 
org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:820)
 ~[hadoop-hdfs-2.0.0-cdh4.5.0.jar:na]
        at 
org.apache.drill.exec.store.dfs.shim.fallback.FallbackFileSystem.getFileStatus(FallbackFileSystem.java:84)
 ~[drill-java-exec-0.7.0-rebuffed.jar:0.7.0]
        at 
org.apache.drill.exec.store.dfs.FileSelection.create(FileSelection.java:135) 
~[drill-java-exec-0.7.0-rebuffed.jar:0.7.0]
        at 
org.apache.drill.exec.store.dfs.WorkspaceSchemaFactory.create(WorkspaceSchemaFactory.java:133)
 ~[drill-java-exec-0.7.0-rebuffed.jar:0.7.0]
{quote}

> can not create hdfs as connection type in storage engine : server throws http 
> 500 error
> ---------------------------------------------------------------------------------------
>
>                 Key: DRILL-1075
>                 URL: https://issues.apache.org/jira/browse/DRILL-1075
>             Project: Apache Drill
>          Issue Type: Bug
>            Reporter: Vivian Summers
>            Assignee: Jacques Nadeau
>            Priority: Critical
>             Fix For: 0.4.0
>
>
> Server at 8047 throws:
> HTTP ERROR 500
> Problem accessing /storage/config/update. Reason:
>     Request failed.
> configure file:
> {
>   "type" : "file",
>   "enabled" : true,
>   "connection" : "hdfs:///",
>   "workspaces" : {
>     "root" : {
>       "location" : "/",
>       "writable" : false,
>       "storageformat" : null
>     },
>     "default" : {
>       "location" : "/user/root",
>       "writable" : true,
>       "storageformat" : null
>     },
>     "tmp" : {
>       "location" : "/tmp",
>       "writable" : true,
>       "storageformat" : "csv"
>     }
>   },
>   "formats" : {
>     "psv" : {
>       "type" : "text",
>       "extensions" : [ "tbl" ],
>       "delimiter" : "|"
>     },
>     "csv" : {
>       "type" : "text",
>       "extensions" : [ "csv" ],
>       "delimiter" : ","
>     },
>     "tsv" : {
>       "type" : "text",
>       "extensions" : [ "tsv" ],
>       "delimiter" : "\t"
>     },
>     "parquet" : {
>       "type" : "parquet"
>     },
>     "json" : {
>       "type" : "json"
>     }
>   }
> }



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to