[jira] [Commented] (AIRFLOW-695) Retries do not execute because dagrun is in FAILED state

2018-11-15 Thread pranav agrawal (JIRA)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-695?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16687829#comment-16687829
 ] 

pranav agrawal commented on AIRFLOW-695:


I am still seeing this issue in 1.9.0, please suggest workaround.

 

> Retries do not execute because dagrun is in FAILED state
> 
>
> Key: AIRFLOW-695
> URL: https://issues.apache.org/jira/browse/AIRFLOW-695
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: DagRun
>Reporter: Harvey Xia
>Priority: Blocker
>  Labels: executor, scheduler
>
> Currently on the latest master commit 
> (15ff540ecd5e60e7ce080177ea3ea227582a4672), running on the LocalExecutor, 
> retries on tasks do not execute because the state of the corresponding dagrun 
> changes to FAILED. The task instance then gets blocked because "Task 
> instance's dagrun was not in the 'running' state but in the state 'failed'," 
> the error message produced by the following lines: 
> https://github.com/apache/incubator-airflow/blob/master/airflow/ti_deps/deps/dagrun_exists_dep.py#L48-L50
> This error can be reproduced with the following simple DAG:
> {code:title=DAG.py|borderStyle=solid}
> dag = models.DAG(dag_id='test_retry_handling')
> task = BashOperator(
> task_id='test_retry_handling_op',
> bash_command='exit 1',
> retries=1,
> retry_delay=datetime.timedelta(minutes=1),
> dag=dag,
> owner='airflow',
> start_date=datetime.datetime(2016, 2, 1, 0, 0, 0))
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (AIRFLOW-695) Retries do not execute because dagrun is in FAILED state

2018-07-29 Thread Brian Kleszyk (JIRA)


[ 
https://issues.apache.org/jira/browse/AIRFLOW-695?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16561356#comment-16561356
 ] 

Brian Kleszyk commented on AIRFLOW-695:
---

Also appear still have this issue in 1.8.1. This was supposedly fixed by 1.8.1.

> Retries do not execute because dagrun is in FAILED state
> 
>
> Key: AIRFLOW-695
> URL: https://issues.apache.org/jira/browse/AIRFLOW-695
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: DagRun
>Reporter: Harvey Xia
>Priority: Blocker
>  Labels: executor, scheduler
>
> Currently on the latest master commit 
> (15ff540ecd5e60e7ce080177ea3ea227582a4672), running on the LocalExecutor, 
> retries on tasks do not execute because the state of the corresponding dagrun 
> changes to FAILED. The task instance then gets blocked because "Task 
> instance's dagrun was not in the 'running' state but in the state 'failed'," 
> the error message produced by the following lines: 
> https://github.com/apache/incubator-airflow/blob/master/airflow/ti_deps/deps/dagrun_exists_dep.py#L48-L50
> This error can be reproduced with the following simple DAG:
> {code:title=DAG.py|borderStyle=solid}
> dag = models.DAG(dag_id='test_retry_handling')
> task = BashOperator(
> task_id='test_retry_handling_op',
> bash_command='exit 1',
> retries=1,
> retry_delay=datetime.timedelta(minutes=1),
> dag=dag,
> owner='airflow',
> start_date=datetime.datetime(2016, 2, 1, 0, 0, 0))
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (AIRFLOW-695) Retries do not execute because dagrun is in FAILED state

2018-05-25 Thread Tylar Murray (JIRA)

[ 
https://issues.apache.org/jira/browse/AIRFLOW-695?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16491080#comment-16491080
 ] 

Tylar Murray commented on AIRFLOW-695:
--

I believe I am seeing this issue using airflow 1.9.0. What is the fix version 
for this bug? Given that this was committed over a year ago I wouldn't expect 
to see this in 1.9.0

> Retries do not execute because dagrun is in FAILED state
> 
>
> Key: AIRFLOW-695
> URL: https://issues.apache.org/jira/browse/AIRFLOW-695
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: DagRun
>Reporter: Harvey Xia
>Priority: Blocker
>  Labels: executor, scheduler
>
> Currently on the latest master commit 
> (15ff540ecd5e60e7ce080177ea3ea227582a4672), running on the LocalExecutor, 
> retries on tasks do not execute because the state of the corresponding dagrun 
> changes to FAILED. The task instance then gets blocked because "Task 
> instance's dagrun was not in the 'running' state but in the state 'failed'," 
> the error message produced by the following lines: 
> https://github.com/apache/incubator-airflow/blob/master/airflow/ti_deps/deps/dagrun_exists_dep.py#L48-L50
> This error can be reproduced with the following simple DAG:
> {code:title=DAG.py|borderStyle=solid}
> dag = models.DAG(dag_id='test_retry_handling')
> task = BashOperator(
> task_id='test_retry_handling_op',
> bash_command='exit 1',
> retries=1,
> retry_delay=datetime.timedelta(minutes=1),
> dag=dag,
> owner='airflow',
> start_date=datetime.datetime(2016, 2, 1, 0, 0, 0))
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (AIRFLOW-695) Retries do not execute because dagrun is in FAILED state

2017-01-03 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/AIRFLOW-695?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15795052#comment-15795052
 ] 

ASF subversion and git services commented on AIRFLOW-695:
-

Commit 2e166b7928c5f66735c687a830f82ff9e1a733b6 in incubator-airflow's branch 
refs/heads/master from root
[ https://git-wip-us.apache.org/repos/asf?p=incubator-airflow.git;h=2e166b7 ]

[AIRFLOW-695] Retries do not execute because dagrun is in FAILED state

The scheduler checks the tasks instances without taking into account
if the executor already reported back. In this case the executor
reports back several iterations later, but the task is queued nevertheless.
Due to the fact tasks will not enter the queue when the task is considered
running, the task state will be "queued” indefinitely and in limbo
between the scheduler and the executor.


> Retries do not execute because dagrun is in FAILED state
> 
>
> Key: AIRFLOW-695
> URL: https://issues.apache.org/jira/browse/AIRFLOW-695
> Project: Apache Airflow
>  Issue Type: Bug
>  Components: DagRun
>Reporter: Harvey Xia
>Priority: Blocker
>  Labels: executor, scheduler
>
> Currently on the latest master commit 
> (15ff540ecd5e60e7ce080177ea3ea227582a4672), running on the LocalExecutor, 
> retries on tasks do not execute because the state of the corresponding dagrun 
> changes to FAILED. The task instance then gets blocked because "Task 
> instance's dagrun was not in the 'running' state but in the state 'failed'," 
> the error message produced by the following lines: 
> https://github.com/apache/incubator-airflow/blob/master/airflow/ti_deps/deps/dagrun_exists_dep.py#L48-L50
> This error can be reproduced with the following simple DAG:
> {code:title=DAG.py|borderStyle=solid}
> dag = models.DAG(dag_id='test_retry_handling')
> task = BashOperator(
> task_id='test_retry_handling_op',
> bash_command='exit 1',
> retries=1,
> retry_delay=datetime.timedelta(minutes=1),
> dag=dag,
> owner='airflow',
> start_date=datetime.datetime(2016, 2, 1, 0, 0, 0))
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)