t;u...@moosheimer.com> wrote:
> Dear devs,
>
> is it possible to use Spark 2.0.2 Scala 2.11 and consume messages from kafka
> server 0.10.0.2 running on Scala 2.10?
> I tried this the last two days by using createDirectStream and can't get no
> message out of kafka?!
>
> I'm usin
Dear devs,
is it possible to use Spark 2.0.2 Scala 2.11 and consume messages from
kafka server 0.10.0.2 running on Scala 2.10?
I tried this the last two days by using createDirectStream and can't get
no message out of kafka?!
I'm using HDP 2.5.3 running kafka_2.10-0.10.0.2.5.3.0-37 and Spark
, Karamba <phantom...@web.de> wrote:
> |Hi, I am looking for an assembly for Spark 2.2.0 with Scala 2.11. I
> can't find one in MVN Repository. Moreover, "org.apache.spark" %%
> "spark-streaming-kafka-0-10_2.11" % "2.1.0 shows that even sbt does not
> find o
|Hi, I am looking for an assembly for Spark 2.2.0 with Scala 2.11. I
can't find one in MVN Repository. Moreover, "org.apache.spark" %%
"spark-streaming-kafka-0-10_2.11" % "2.1.0 shows that even sbt does not
find one: [error] (*:update) sbt.ResolveException: unresolved
dep
Hi folks,
Our project is a mess of scala 2.10 and 2.11, so I tried to switch
everything to 2.11.
I had some exasperating errors like this :
java.lang.NoClassDefFoundError:
org/apache/spark/sql/execution/datasources/DDLParser
at org.apache.spark.sql.SQLContext.(SQLContext.scala:208)
at
,
The instructions for building spark against scala-2.11 indicate using
-Dspark-2.11. When I look in the pom.xml I find a profile named
'spark-2.11' but nothing that would indicate I should set a property. The
sbt build seems to need the -Dscala-2.11 property set. Finally build/mvn
does a simple
I've used the instructions and it worked fine.
Can you post exactly what you're doing, and what it fails with? Or are you
just trying to understand how it works?
2015-08-24 15:48 GMT-04:00 Lanny Ripple la...@spotright.com:
Hello,
The instructions for building spark against scala-2.11
...@spotright.com wrote:
Hello,
The instructions for building spark against scala-2.11 indicate using
-Dspark-2.11. When I look in the pom.xml I find a profile named
'spark-2.11' but nothing that would indicate I should set a property. The
sbt build seems to need the -Dscala-2.11 property set. Finally
Hello,
The instructions for building spark against scala-2.11 indicate using
-Dspark-2.11. When I look in the pom.xml I find a profile named
'spark-2.11' but nothing that would indicate I should set a property. The
sbt build seems to need the -Dscala-2.11 property set. Finally build/mvn
does
You were building against 1.4.x, right ?
In master branch, switch-to-scala-2.11.sh is gone. There is scala-2.11
profile.
FYI
On Sun, Aug 16, 2015 at 11:12 AM, Stephen Boesch java...@gmail.com wrote:
I am building spark with the following options - most notably the
**scala-2.11**:
.
In 1.4 it is change-scala-version.sh 2.11
But the problem was it is a -Dscala-211 not a -P. I misread the doc's.
2015-08-17 14:17 GMT-07:00 Ted Yu yuzhih...@gmail.com:
You were building against 1.4.x, right ?
In master branch, switch-to-scala-2.11.sh is gone. There is scala-2.11
I am building spark with the following options - most notably the
**scala-2.11**:
. dev/switch-to-scala-2.11.sh
mvn -Phive -Pyarn -Phadoop-2.6 -Dhadoop2.6.2 -Pscala-2.11 -DskipTests
-Dmaven.javadoc.skip=true clean package
The build goes pretty far but fails in one of the minor modules
some serious stability problems simply trying to run the Spark 1.3 Scala 2.11 REPL. Most of the time it fails to load and spews a torrent of compiler assertion failures, etc. See attached.spark@dp-cluster-master-node-001:~/spark/bin$ spark-shell
Spark Command: java -cp
/opt/spark/conf:/opt/spark/lib
on 2.11 is here:
http://spark.apache.org/docs/latest/building-spark.html#building-for-scala-211
Also, I'm experiencing some serious stability problems simply trying to run
the Spark 1.3 Scala 2.11 REPL. Most of the time it fails to load and spews a
torrent of compiler assertion failures, etc. See
to run
the Spark 1.3 Scala 2.11 REPL. Most of the time it fails to load and spews a
torrent of compiler assertion failures, etc. See attached.
Unfortunately, it appears the Spark 1.3 Scala 2.11 REPL is simply not ready
for production use. I was going to file a bug, but it seems clear
/latest/building-spark.html#building-for-scala-211
Also, I'm experiencing some serious stability problems simply trying to run
the Spark 1.3 Scala 2.11 REPL. Most of the time it fails to load and spews a
torrent of compiler assertion failures, etc. See attached.
Unfortunately, it appears
trying to run
the Spark 1.3 Scala 2.11 REPL. Most of the time it fails to load and spews a
torrent of compiler assertion failures, etc. See attached.
Unfortunately, it appears the Spark 1.3 Scala 2.11 REPL is simply not ready
for production use. I was going to file a bug, but it seems
-2988
Any help in getting this working would be much appreciated!
Thanks
Alex
On Thu, Apr 9, 2015 at 11:32 AM, Prashant Sharma scrapco...@gmail.com
wrote:
You are right this needs to be done. I can work on it soon, I was not
sure if there is any one even using scala 2.11 spark repl. Actually
was not sure
if there is any one even using scala 2.11 spark repl. Actually there is a
patch in scala 2.10 shell to support adding jars (Lost the JIRA ID), which
has to be ported for scala 2.11 too. If however, you(or anyone else) are
planning to work, I can help you ?
Prashant Sharma
On Thu
You are right this needs to be done. I can work on it soon, I was not sure
if there is any one even using scala 2.11 spark repl. Actually there is a
patch in scala 2.10 shell to support adding jars (Lost the JIRA ID), which
has to be ported for scala 2.11 too. If however, you(or anyone else
20 matches
Mail list logo