Hi,
I am facing issues with the high number of active jobs showing in UI. I am
using spark structure streaming to read data from Solace and writing back
to Kafka.
After 24-28Hr, I see active jobs start accumulating and it happens on
random intervals after 24-28hr. And also I can see many active tas
Java is very much supported in Spark. In our open source project, we
haven’t done spark connect yet but we do a lot of transformations, ML and
graph stuff using Java with Spark. Never faced the language barrier.
Cheers,
Sonal
https://github.com/zinggAI/zingg
On Sat, 5 Apr 2025 at 4:42 PM, Ángel
I think you have more limitations using Spark Connect than Spark from Java.
I used RDD, registered UDFs, ... from Java without any problems.
El sáb, 5 abr 2025, 9:50, tim wade escribió:
> Hello
>
> I only know Java programming. If I use Java to communicate with the
> Spark API and submit tasks t
Hello
I only know Java programming. If I use Java to communicate with the
Spark API and submit tasks to Spark API from Java, I'm not sure what
disadvantages this might have. I see other people writing tasks in
Scala, then compiling them and submitting to Spark using spark-submit.
Thanks.
O
I think I did that some years ago in Spark 2.4 on a Hortonworks cluster
with SSL and Kerberos enabled. It worked, but never went into production.
El vie, 4 abr 2025, 9:54, tim wade escribió:
> Hello
>
> I am just newbie to spark. I am programming with Java mainly, knowing
> scala very bit.
>
> C