BTW, do we have to register JdbcDialect for every Spark/SQL context, or
once for a Spark server?
On Sun, Jul 23, 2017 at 2:26 AM, Luqman Ghani wrote:
> I have found the solution for this error. I have to register a JdbcDialect
> for Drill as mentioned in the following post on
I have found the solution for this error. I have to register a JdbcDialect
for Drill as mentioned in the following post on SO:
https://stackoverflow.com/questions/35476076/integrating-spark-sql-and-apache-drill-through-jdbc
Thanks
On Sun, Jul 23, 2017 at 2:10 AM, Luqman Ghani
I have done that, but Spark is encompassing my query with same clause:
SELECT "CustomerID", etc FROM ( my query from table) so same error.
On Sun, Jul 23, 2017 at 2:02 AM, ayan guha wrote:
> You can formulate a query in dbtable clause in jdbc reader.
>
> On Sun, 23 Jul 2017
Hi - and my thanks to you and Gerard. Only late hour in the night can explain
how I could possibly miss this.
Cheers!
Lukasz
On 22/07/2017 10:48, yohann jardin wrote:
Hello Lukasz,
You can just:
val pairRdd = javapairrdd.rdd();
Then pairRdd will be of type RDD>, with K
You can formulate a query in dbtable clause in jdbc reader.
On Sun, 23 Jul 2017 at 6:43 am, Luqman Ghani wrote:
> Hi,
>
> I'm working on integrating Apache Drill with Apache Spark with Drill's
> JDBC driver. I'm trying a simple select * from table from Drill through
>
Hi,
I'm working on integrating Apache Drill with Apache Spark with Drill's JDBC
driver. I'm trying a simple select * from table from Drill through
spark.sqlContext.load via jdbc driver. I'm running the following code in
Spark Shell:
> ./bin/spark-shell --driver-class-path
The Dataset.join(right: Dataset[_], joinExprs: Column) API can use any
arbitrary expression so you can use UDF for join.
The problem with all non-equality joins is that they use
BroadcastNestedLoopJoin or equivalent, that is an (M X N) nested-loop
which will be unusable for medium/large
Hi everyone,
My environment is PySpark with Spark 2.0.0.
I'm using spark to load data from a large number of files into a Spark
dataframe with fields say field1 to field10. While loading my data I have
ensured that records are partitioned by field1 and field2(without using
partitionBy). This
Normally a family of joins (left, right outter, inner) are performed on two
dataframes using columns for the comparison ie left("acol") ===
ight("acol") . the comparison operator of the "left" dataframe does
something internally and produces a column that i assume is used by the
join.
What I want
On Saturday 22 July 2017 01:31 PM, kant kodali wrote:
Is there a way to run Spark SQL through REST?
There is spark-jobserver
(https://github.com/spark-jobserver/spark-jobserver). It does more than
just REST API (like long running SparkContext).
regards
--
Sumedh Wale
SnappyData
There's Livi but it's pretty resource intensive.
I know it's not helpful but my company has developed its own and I am trying to
Open Source it.
Looks like there are quite a few companies who had the need and custom build.
jg
> On Jul 22, 2017, at 04:01, kant kodali
Hello Lukasz,
You can just:
val pairRdd = javapairrdd.rdd();
Then pairRdd will be of type RDD>, with K being
com.vividsolutions.jts.geom.Polygon, and V being
java.util.HashSet[com.vividsolutions.jts.geom.Polygon]
If you really want to continue with Java objects:
val
Please unsubscribe me
Is there a way to run Spark SQL through REST?
14 matches
Mail list logo