I didn't expect to participate at any debate on that software, as I'm a
complete newcomer. But I'm almost forced as I am the target audience, too.

To answer your initial question, after reading a lot of documentation I
find the term execution_date really counterintuitive, so yes maybe
period_start and period_end might be a better naming to help to understand
how all the initial scheduling works. Because even after reading the
scheduling section of the doc and the FAQ, it was still not clear in my
mind. Btw, I find some ideas exposed by James Meickle in the [DISCUSS]
AIRFLOW-4192 very interesting and I share his opinion that there's still
room for improvement.
But a mode to change from "run at end of period, I need all the data
available for this period" (the current) to "run at _this_ time on the
schedule_interval would be awesome.

Regards,
Flo

On Tue, Apr 9, 2019 at 4:41 PM Ash Berlin-Taylor <a...@apache.org> wrote:

> Yeah, that's the other thing that has been talked about from time-to-time,
> which is a mode to change from "run at end of period, I need all the data
> available for this period" (the current) to "run at _this_ time on the
> schedule_interval, don't wait for the period to end".
>
> (No such flag exists right now, before you go looking.)
>
> > On 9 Apr 2019, at 15:31, Shaw, Damian P. <
> damian.sha...@credit-suisse.com> wrote:
> >
> > Hi all,
> >
> > I'm new to this Airflow Dev mailing list so I wasn't expecting to reply
> to anything but I feel I am the target audience for this question. I am
> quite new to airflow and have been setting up an airflow environment for my
> business this last month.
> >
> > I find the current "execution_date" a small technical burden and a large
> cognitive burden. Our workflow is based on DAGs running at a specified time
> in a specified timezone using the same date as the current calendar date.
> >
> > I have worked around this by creating my own macro and context
> variables, with the logic looking like this:
> >        airflow_execution_date = context['execution_date']
> >        dag_timezone = context['dag'].timezone
> >        local_execution_date =
> dag_timezone.convert(airflow_execution_date)
> >        local_cal_date = local_execution_date + datetime.timedelta(days=1)
> >
> > As you can see this isn't a lot of technical effort, but having a date
> that 1) is in the timezone the business users are working in, and 2) Is the
> same calendar date the business users are working in it significantly
> reduces the cognitive effort required to set-up tasks. Of course this
> doesn't help with cron format scheduling which I just let the business give
> me the requirements for and I set it up myself as the date logic there is
> still confusing as it doesn't work like real cron scheduling which everyone
> is familiar with.
> >
> > Maybe "period_start" and "period_end" might help people on Day 0 of
> understanding Airflow get that the dates you are dealing with are not what
> you expect, but Day 1+ there's still a lot of cognitive overhead if you
> don't have the exact same model as AirBnb for running DAGs and tasks.
> >
> > My 2 cents anyway,
> > Damian Shaw
> >
> >
> > -----Original Message-----
> > From: Ash Berlin-Taylor [mailto:a...@apache.org]
> > Sent: Tuesday, April 09, 2019 10:08 AM
> > To: dev@airflow.apache.org
> > Subject: [DISCUSS] period_start/period_end instead of
> execution_date/next_execution_date
> >
> > (trying to break this out in to another thread)
> >
> > The ML doesn't allow  images, but I can guess that it is the deps
> section of a task instance details screen?
> >
> > I'm not saying it's not clear once you know to look there, but I'm
> trying remove/reduce the confusion in the first place. And I think we as
> committers aren't best placed to know what makes sense as we have
> internalised how Airflow works :)
> >
> > So I guess this is a question to the newest people on the list: Would
> `period_start` and `period_end` be more or less confusing for you when you
> were first getting started with Airflow?
> >
> > -ash
> >
> >> On 9 Apr 2019, at 14:47, Driesprong, Fokko <fo...@driesprong.frl>
> wrote:
> >>
> >> Ash,
> >>
> >> Personally, I think this is quite clear, there is a list of reasons why
> the job isn't being scheduled:
> >>
> >>
> >> Coming back to the question of Bas, I believe that yesterday_ds does
> not make sense since we cannot assume that the schedule is daily. I don't
> see any usage of this variable. Personally, I do use next_execution_date
> quite extensively. When you have a job that runs daily, but you want to
> change this to an hourly job. In such a case you don't want to change {{
> (execution_date + macros.timedelta(days=1)) }} to {{ (execution_date +
> macros.timedelta(hours=1)) }} everywhere.
> >>
> >> I'm just not sure if the aggressive deprecation of is really worth it.
> I don't see too much harm in letting them stay.
> >>
> >> Cheers, Fokko
> >>
> >> Op di 9 apr. 2019 om 12:17 schreef Ash Berlin-Taylor <a...@apache.org
> <mailto:a...@apache.org>>:
> >> To (slightly) hijack this thread:
> >>
> >> On the subject of execuction_date: as I'm sure we're all aware the
> concept of execution_date is confusing to new-commers to Airflow (there are
> many questions about "why hasn't my DAG run yet"? "Why is my dag a day
> behind?" etc.) and although we mention this in the docs it's a confusing
> concept.
> >>
> >> What to people think about adding two new parameters: `period_start`
> and `period_end` and making these the preferred terms in place of
> execution_date and next_execution_date?
> >>
> >> This hopefully avoids any ambitious terms like "execution" or "run"
> which is understandably easy to conflate with the time the task is being
> run (i.e. `now()`)
> >>
> >> If people think this naming is better and less confusing I would
> suggest we update all the docs and examples to use these terms (but still
> mention the old names somewhere, probably in the macros docs)
> >>
> >> Thoughts?
> >>
> >> -ash
> >>
> >>
> >>> On 8 Apr 2019, at 16:20, Arthur Wiedmer <arthur.wied...@gmail.com
> <mailto:arthur.wied...@gmail.com>> wrote:
> >>>
> >>> Hi Bas,
> >>>
> >>> 1) I am aware of a few places where those parameters are used in
> production
> >>> in a few hundred jobs. I highly recommend we don't deprecate them
> unless we
> >>> do it in a major version.
> >>>
> >>> 2) As James mentioned, inlets and outlets are a lineage annotation
> feature
> >>> which is still under development. Let's leave them in, but we can guard
> >>> them behind a feature flag if you prefer.
> >>>
> >>> 3) the yesterday*/tomorrow* params are convenience ones if you use a
> daily
> >>> ETL. I agree with you that they are simple to compute, but not everyone
> >>> using Apache Airflow is amazing with Python. Some users are only
> trying to
> >>> get a query scheduled and rely on a couple of niceties like these to
> get by.
> >>>
> >>> 4) latest_date, end_date (I feel like there used to be start_date, but
> >>> maybe it got lost) were a blend of things which were used by a backfill
> >>> framework used internally at Airbnb. Latest date was used if you
> needed to
> >>> join to a dimension for which you only wanted the latest version of the
> >>> attributes in you backfill. end_date was used for time ranges where
> several
> >>> days were processed together in a range to save on compute. I don't
> see an
> >>> issue with removing them.
> >>>
> >>> Best regards,
> >>> Arthur
> >>>
> >>>
> >>>
> >>> On Mon, Apr 8, 2019 at 5:37 AM Bas Harenslak <
> basharens...@godatadriven.com <mailto:basharens...@godatadriven.com>>
> >>> wrote:
> >>>
> >>>> Hi all,
> >>>>
> >>>> Following Tao Feng’s question to discuss this PR<
> >>>> https://github.com/apache/airflow/pull/5010 <
> https://github.com/apache/airflow/pull/5010>> (AIRFLOW-4192<
> >>>> https://issues.apache.org/jira/browse/AIRFLOW-4192 <
> https://issues.apache.org/jira/browse/AIRFLOW-4192>>), please discuss here
> >>>> if you agree/disagree/would change.
> >>>>
> >>>> -----------
> >>>>
> >>>> The summary of the PR:
> >>>>
> >>>> I was confused by the task context values and suggest to clean up and
> >>>> clarify these variables. Some are derivations from other variables,
> some
> >>>> are undocumented and unused, some are wrong (name doesn’t match the
> value).
> >>>> Please discuss what you think of the removal of these variables:
> >>>>
> >>>>
> >>>> *   Removed yesterday_ds, yesterday_ds_nodash, tomorrow_ds,
> >>>> tomorrow_ds_nodash. IMO the next_* and previous_* variables are useful
> >>>> since these require complex logic to compute the next execution date,
> >>>> however would leave computing the yesterday* and tomorrow* variables
> up to
> >>>> the user since they are simple one-liners and don't relate to the DAG
> >>>> interval.
> >>>> *   Removed tables. This is a field in params, and is thus also
> >>>> accessible by the user ({{ params.tables }}). Also, it was
> undocumented.
> >>>> *   Removed latest_date. It's the same as ds and was also
> undocumented.
> >>>> *   Removed inlets and outlets. Also undocumented, and have the
> >>>> inlets/outlets ever worked/ever been used by anybody?
> >>>> *   Removed end_date and END_DATE. Both have the same value, so it
> >>>> doesn't make sense to have both variables. Also, the value is ds which
> >>>> contains the start date of the interval, so the naming didn't make
> sense to
> >>>> me. However, if anybody argues in favour of adding "start_date" and
> >>>> "end_date" to provide the start and end datetime of task instance
> >>>> intervals, I'd be happy to add them.
> >>>>
> >>>> Cheers,
> >>>> Bas
> >>>>
> >>
> >
> >
> >
> >
> ===============================================================================
>
> > Please access the attached hyperlink for an important electronic
> communications disclaimer:
> > http://www.credit-suisse.com/legal/en/disclaimer_email_ib.html
> >
> ===============================================================================
>
>
>

Reply via email to