In what you describe it doesn’t sound like a state change as much as just a dag
that has as its first operator an http operator, and another at the end that
communicates final state?
Sent from a device with less than stellar autocorrect
> On Nov 7, 2018, at 11:12 AM, Aleks Shulman
> wrote:
How do you trigger it externally?
We have several custom operators that trigger other jobs and we had to be
really careful or we’d get duplicate keys for the dag run and it would fail to
kick off.
One scheduler, but we saw it repeatedly and have it noted as a thing to watch
out for.
Brian
My first thought was this, but my understanding is That if you had a large
number of dags “waiting” the sensor would consume all the concurrency.
And what if the user doesn’t approve?
How about the dag you have as it’s last step writes to an api/db the status.
Then 2 other dags (or one with
The operator is a thin wrapper around their SDK. The qubole SDK makes heavy
use of **kwargs, and the operator just passes them through.
Short of writing our own operator with named Params to keep the base operator
happy and then delegate to the qubole sdk, there’s no other way to silence the
and start from:
>>
>> A lot of new users get confused by how "execution_date" works.
>>
>> I recognize that some of this is a learning curve issue and some of this is
>> a mindset issue but it begs the question: do enough users benefit from the
>> cu
It took a minute to grok, but in the larger context of how af works it makes
perfect sense the way it is. Changing something so fundamentally breaking to
every dag in existence should bring a comparable benefit. Beyond the avoiding
teaching a concept you disagree with, what benefits does the
Prior to using airflow for much, on first inspection, I think I may have agreed
with you.
After a bit of use I’d agree with Fokko and others - this isn’t really a
problem, and separating them seems to do more harm than good related to
deployment.
I was gonna stop there, but why?
You can
I suggest it’ll work for your needs.
Sent from a device with less than stellar autocorrect
> On May 21, 2018, at 10:16 AM, purna pradeep wrote:
>
> Hi ,
>
> I’m trying to evaluate airflow to see if it suits my needs.
>
> Basically i can have below steps in a DAG
>
>
Okay I’ll bite... WT* does that mean?
one of the best things about airflow is how easy it is to connect disparate
systems... some would even say that’s much of the reason it exists..
It records reasonable metadata into an rdms (I suppose you could argue for
other designs, but it’s pretty
+1
Sent from a device with less than stellar autocorrect
> On Apr 25, 2018, at 12:04 PM, James Meickle wrote:
>
> Another reason you would want separated infrastructure is that there are a
> lot of ways to exhaust Airflow resources or otherwise cause contention -
>
un in parallel.
> I think I have setup the tasks correctly to use pool but may be missing the
> priority_weight setting correctly. Appreciate if you could run by your
> configs just to see if I am not missing any simple point.
>
> thanks much,
> Manish
>
>
To be clear, you’re hoping that setting the slots to 1 will cause the tasks
across district dags to run in order based on the assumption that they’ll queue
up and then execute off the pool?
I don’t think it will quite work that way - there’s no guarantee the scheduler
will execute your tasks
I’d think we’d have privilege ‘can_view’ etc, and then a join table (priv) <->
(dagid) <-> (user/group). Then it’s a simple query to get the perms for a
given dag (as you list In option 2 below).
It also makes a “secure by default” easy - a lack of entries in that table for
a dag can mean
My $.02 - posted to SO as well.
I fought this use case for a long time. In short, a dag that’s built based on
the state of a changing resource, especially a db table, doesn’t fly so well in
airflow.
My solution was to write a small custom operator that’s a subclass if
truggerdagoperator, it
Good evening,
Here's the setup -
I'm following a "trigger dag -> called(processing) dag" kind of structure.
Then I set timespans, catchups, etc on the triggering dag. This allows me
to separate the scheduling from the execution, and so far is working great.
The triggered dag uses parameters
15 matches
Mail list logo