uthor will in no case be liable for any monetary damages arising from such
> loss, damage or destruction.
>
>
>
> On Thu, 2 Feb 2023 at 17:26, Harut Martirosyan <mailto:harut.martiros...@gmail.com>> wrote:
>> Generally, the problem is that I don’t find a way to automatically
Generally, the problem is that I don’t find a way to automatically create a
JDBC table in the JDBC database when I want to insert data into it using Spark
SQL only, not DataFrames API.
> On 2 Feb 2023, at 21:22, Harut Martirosyan
> wrote:
>
> Hi, thanks for the reply.
>
>
ble for any monetary damages arising from such
> loss, damage or destruction.
>
>
>
> On Wed, 1 Feb 2023 at 19:33, Harut Martirosyan <mailto:harut.martiros...@gmail.com>> wrote:
>> I have a resultset (defined in SQL), and I want to insert it into my
that using pure SQL (no python/scala/java)?
I am trying to use INSERT OVERWRITE DIRECTORY with JDBC file format
(according to the documentation) but as expected this functionality is
available only for File-based storage systems.
--
RGRDZ Harut
Hello.
We’re successfully exporting technical metrics to prometheus using built-in
capabilities of Spark 3, but we need to add custom business metrics as well
using python. Seems like there’s no documentation for that.
Thanks.
Hi.
I'm running spark on YaRN without CGroups turned on, and have 2 questions:
1. Does anyone of spark/yarn guarantee that my spark tasks won't eat up more
CPU cores than I've assigned? (I assume there are no guarantees, correct?)
2. What is the effect of setting --executor-cores when submitting
things to the stack like Storm or Flume, but
from other side, it really takes much more resources on same streaming than
our previous setup with Flume, especially if we have multiple destinations
(triggers multiple actions/scheduling)
--
RGRDZ Harut
, like, if a persisted block is lost or otherwise
unavailable later.
On Sun, Mar 29, 2015 at 9:07 AM, Harut Martirosyan
harut.martiros...@gmail.com wrote:
Hi.
rdd.persist()
rdd.count()
rdd.transform()...
is there a chance transform() runs before persist() is complete?
--
RGRDZ
Hi.
rdd.persist()
rdd.count()
rdd.transform()...
is there a chance transform() runs before persist() is complete?
--
RGRDZ Harut
to trigger actions?
-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org
--
RGRDZ Harut
--driver-memory 10g --num-executors 8
There are no evident logs, just messages that executors are being lost, and
connection refused errors, (apparently due to executor failures)
The cluster is the same, 8 nodes, 64Gb RAM each.
Format is parquet.
--
RGRDZ Harut
Hi guys.
Basically, we had to define a UDF that does that, is there a built in
function that we can use for it?
--
RGRDZ Harut
I'm trying to build a dashboard to visualize stream of events coming from
mobile devices.
For example, I have event called add_photo, from which I want to calculate
trending tags for added photos for last x minutes. Then I'd like to
aggregate that by country, etc. I've built the streaming part,
of
filter, then view results. This is easy when you have 1 user, but if you
have team of analysts all specifying their own criteria, it becomes hard to
manage them all.
On 20 March 2015 at 12:02, Jeffrey Jedele jeffrey.jed...@gmail.com wrote:
Hey Harut,
I don't think there'll by any general
Management Award
NetworkWorld 10 Startups to Watch
EMA Most Notable Vendor
On Fri, Mar 20, 2015 at 1:06 AM, Harut Martirosyan
harut.martiros...@gmail.com wrote:
Hey Jeffrey.
Thanks for reply.
I already have something similar, I use Grafana and Graphite, and for
simple metric streaming
15 matches
Mail list logo