Thanks Yin !

every thing else works great!

+1 (non-binding)

On Fri, Jun 5, 2015 at 2:11 PM, Yin Huai <yh...@databricks.com> wrote:

> Hi Bobby,
>
> sqlContext.table("test.test1") is not officially supported in 1.3. For
> now, please use the "use database" as a workaround. We will add it.
>
> Thanks,
>
> Yin
>
> On Fri, Jun 5, 2015 at 12:18 PM, Bobby Chowdary <
> bobby.chowdar...@gmail.com> wrote:
>
>> Not sure if its a blocker but there might be a minor issue with hive
>> context, there is also a work around
>>
>> *Works:*
>>
>> from pyspark.sql import HiveContext
>>
>> sqlContext = HiveContext(sc)
>> df = sqlContext.sql("select * from test.test1")
>>
>> *Does not Work:*
>>
>>  df = sqlContext.table("test.test1")
>>
>> Py4JJavaError: An error occurred while calling o260.table. : 
>> org.apache.spark.sql.catalyst.analysis.NoSuchTableException     at 
>> org.apache.spark.sql.hive.client.ClientInterface$anonfun$getTable$1.apply(ClientInterface.scala:112)
>>      at 
>> org.apache.spark.sql.hive.client.ClientInterface$anonfun$getTable$1.apply(ClientInterface.scala:112)
>>      at scala.Option.getOrElse(Option.scala:120)     at 
>> org.apache.spark.sql.hive.client.ClientInterface$class.getTable(ClientInterface.scala:112)
>>      at 
>> org.apache.spark.sql.hive.client.ClientWrapper.getTable(ClientWrapper.scala:58)
>>      at 
>> org.apache.spark.sql.hive.HiveMetastoreCatalog.lookupRelation(HiveMetastoreCatalog.scala:227)
>>      at 
>> org.apache.spark.sql.hive.HiveContext$anon$2.org$apache$spark$sql$catalyst$analysis$OverrideCatalog$super$lookupRelation(HiveContext.scala:370)
>>      at 
>> org.apache.spark.sql.catalyst.analysis.OverrideCatalog$anonfun$lookupRelation$3.apply(Catalog.scala:165)
>>      at 
>> org.apache.spark.sql.catalyst.analysis.OverrideCatalog$anonfun$lookupRelation$3.apply(Catalog.scala:165)
>>      at scala.Option.getOrElse(Option.scala:120)     at 
>> org.apache.spark.sql.catalyst.analysis.OverrideCatalog$class.lookupRelation(Catalog.scala:165)
>>      at 
>> org.apache.spark.sql.hive.HiveContext$anon$2.lookupRelation(HiveContext.scala:370)
>>      at org.apache.spark.sql.SQLContext.table(SQLContext.scala:754)     at 
>> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)     at 
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>      at 
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>      at java.lang.reflect.Method.invoke(Method.java:497)     at 
>> py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:231)     at 
>> py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:379)     at 
>> py4j.Gateway.invoke(Gateway.java:259)     at 
>> py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:133)     at 
>> py4j.commands.CallCommand.execute(CallCommand.java:79)     at 
>> py4j.GatewayConnection.run(GatewayConnection.java:207)     at 
>> java.lang.Thread.run(Thread.java:745)  (<class 
>> 'py4j.protocol.Py4JJavaError'>, Py4JJavaError(u'An error occurred while 
>> calling o260.table.\n', JavaObject id=o262), <traceback object at 0x2e248c0>)
>>
>> How ever which i swtich db context it works
>>
>> *Works:*
>>
>>  sqlContext.sql("use test")
>>  df = sqlContext.table("test1")
>>
>> Bulit on Mac OSX  JDK6for Mapr Distribution and Running on CentOS 7.0 JDK8
>>
>> make-distribution.sh --tgz -Pmapr4  -Phive -Pnetlib-lgpl -Phive-thriftserver
>>
>> didn’t have this issue in RC3 and tried it on scala as well.
>>
>> Thanks
>> Bobby
>> ​
>>
>
>

Reply via email to