[ 
https://issues.apache.org/jira/browse/AIRFLOW-168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15301461#comment-15301461
 ] 

Chris Riccomini edited comment on AIRFLOW-168 at 5/26/16 3:24 AM:
------------------------------------------------------------------

I noticed that the scheduler log shows (stacktrace at bottom):

{noformat}
[2016-05-25 20:22:37,925] {jobs.py:580} INFO - Prioritizing 0 queued jobs
[2016-05-25 20:22:37,933] {jobs.py:732} INFO - Starting 0 scheduler jobs
[2016-05-25 20:22:37,933] {jobs.py:747} INFO - Done queuing tasks, calling the 
executor's heartbeat
[2016-05-25 20:22:37,933] {jobs.py:750} INFO - Loop took: 0.011795 seconds
[2016-05-25 20:22:37,936] {models.py:308} INFO - Finding 'running' jobs without 
a recent heartbeat
[2016-05-25 20:22:37,937] {models.py:314} INFO - Failing jobs without heartbeat 
after 2016-05-25 20:20:22.937222
[2016-05-25 20:22:42,925] {jobs.py:580} INFO - Prioritizing 0 queued jobs
[2016-05-25 20:22:42,934] {jobs.py:732} INFO - Starting 1 scheduler jobs
[2016-05-25 20:22:42,977] {models.py:2703} INFO - Checking state for <DagRun 
example_xcom @ 2016-05-25 20:22:42.953808: 
scheduled__2016-05-25T20:22:42.953808, externally triggered: False>
[2016-05-25 20:22:42,983] {jobs.py:504} INFO - Getting list of tasks to skip 
for active runs.
[2016-05-25 20:22:42,986] {jobs.py:520} INFO - Checking dependencies on 3 tasks 
instances, minus 0 skippable ones
[2016-05-25 20:22:42,991] {base_executor.py:36} INFO - Adding to queue: airflow 
run example_xcom push 2016-05-25T20:22:42.953808 --local -sd 
DAGS_FOLDER/example_dags/example_xcom.py 
[2016-05-25 20:22:42,993] {base_executor.py:36} INFO - Adding to queue: airflow 
run example_xcom push_by_returning 2016-05-25T20:22:42.953808 --local -sd 
DAGS_FOLDER/example_dags/example_xcom.py 
[2016-05-25 20:22:43,011] {jobs.py:747} INFO - Done queuing tasks, calling the 
executor's heartbeat
[2016-05-25 20:22:43,012] {jobs.py:750} INFO - Loop took: 0.089461 seconds
[2016-05-25 20:22:43,018] {models.py:308} INFO - Finding 'running' jobs without 
a recent heartbeat
[2016-05-25 20:22:43,019] {models.py:314} INFO - Failing jobs without heartbeat 
after 2016-05-25 20:20:28.019143
[2016-05-25 20:22:43,028] {sequential_executor.py:26} INFO - Executing command: 
airflow run example_xcom push 2016-05-25T20:22:42.953808 --local -sd 
DAGS_FOLDER/example_dags/example_xcom.py 
[2016-05-25 20:22:43,453] {__init__.py:36} INFO - Using executor 
SequentialExecutor
Logging into: 
/Users/chrisr/airflow/logs/example_xcom/push/2016-05-25T20:22:42.953808
[2016-05-25 20:22:44,300] {__init__.py:36} INFO - Using executor 
SequentialExecutor
[2016-05-25 20:22:48,937] {sequential_executor.py:26} INFO - Executing command: 
airflow run example_xcom push_by_returning 2016-05-25T20:22:42.953808 --local 
-sd DAGS_FOLDER/example_dags/example_xcom.py 
[2016-05-25 20:22:49,366] {__init__.py:36} INFO - Using executor 
SequentialExecutor
Logging into: 
/Users/chrisr/airflow/logs/example_xcom/push_by_returning/2016-05-25T20:22:42.953808
[2016-05-25 20:22:50,210] {__init__.py:36} INFO - Using executor 
SequentialExecutor
[2016-05-25 20:22:54,844] {jobs.py:580} INFO - Prioritizing 0 queued jobs
[2016-05-25 20:22:54,853] {jobs.py:732} INFO - Starting 1 scheduler jobs
[2016-05-25 20:22:54,903] {models.py:2703} INFO - Checking state for <DagRun 
example_xcom @ 2015-01-01 00:00:00: scheduled__2015-01-01T00:00:00, externally 
triggered: False>
[2016-05-25 20:22:54,907] {models.py:2703} INFO - Checking state for <DagRun 
example_xcom @ 2016-05-25 20:22:42.953808: 
scheduled__2016-05-25T20:22:42.953808, externally triggered: False>
[2016-05-25 20:22:54,911] {jobs.py:504} INFO - Getting list of tasks to skip 
for active runs.
[2016-05-25 20:22:54,913] {jobs.py:520} INFO - Checking dependencies on 6 tasks 
instances, minus 2 skippable ones
[2016-05-25 20:22:54,920] {base_executor.py:36} INFO - Adding to queue: airflow 
run example_xcom push 2015-01-01T00:00:00 --local -sd 
DAGS_FOLDER/example_dags/example_xcom.py 
[2016-05-25 20:22:54,921] {base_executor.py:36} INFO - Adding to queue: airflow 
run example_xcom push_by_returning 2015-01-01T00:00:00 --local -sd 
DAGS_FOLDER/example_dags/example_xcom.py 
[2016-05-25 20:22:54,935] {base_executor.py:36} INFO - Adding to queue: airflow 
run example_xcom puller 2016-05-25T20:22:42.953808 --local -sd 
DAGS_FOLDER/example_dags/example_xcom.py 
[2016-05-25 20:22:54,954] {jobs.py:747} INFO - Done queuing tasks, calling the 
executor's heartbeat
[2016-05-25 20:22:54,954] {jobs.py:750} INFO - Loop took: 0.113319 seconds
[2016-05-25 20:22:54,960] {models.py:308} INFO - Finding 'running' jobs without 
a recent heartbeat
[2016-05-25 20:22:54,960] {models.py:314} INFO - Failing jobs without heartbeat 
after 2016-05-25 20:20:39.960629
[2016-05-25 20:22:54,978] {sequential_executor.py:26} INFO - Executing command: 
airflow run example_xcom push_by_returning 2015-01-01T00:00:00 --local -sd 
DAGS_FOLDER/example_dags/example_xcom.py 
[2016-05-25 20:22:55,410] {__init__.py:36} INFO - Using executor 
SequentialExecutor
Logging into: 
/Users/chrisr/airflow/logs/example_xcom/push_by_returning/2015-01-01T00:00:00
[2016-05-25 20:22:56,239] {__init__.py:36} INFO - Using executor 
SequentialExecutor
[2016-05-25 20:23:00,873] {sequential_executor.py:26} INFO - Executing command: 
airflow run example_xcom push 2015-01-01T00:00:00 --local -sd 
DAGS_FOLDER/example_dags/example_xcom.py 
[2016-05-25 20:23:01,477] {__init__.py:36} INFO - Using executor 
SequentialExecutor
Logging into: /Users/chrisr/airflow/logs/example_xcom/push/2015-01-01T00:00:00
[2016-05-25 20:23:02,639] {__init__.py:36} INFO - Using executor 
SequentialExecutor
[2016-05-25 20:23:07,150] {sequential_executor.py:26} INFO - Executing command: 
airflow run example_xcom puller 2016-05-25T20:22:42.953808 --local -sd 
DAGS_FOLDER/example_dags/example_xcom.py 
[2016-05-25 20:23:07,620] {__init__.py:36} INFO - Using executor 
SequentialExecutor
Logging into: 
/Users/chrisr/airflow/logs/example_xcom/puller/2016-05-25T20:22:42.953808
[2016-05-25 20:23:08,498] {__init__.py:36} INFO - Using executor 
SequentialExecutor
[2016-05-25 20:23:13,128] {jobs.py:580} INFO - Prioritizing 0 queued jobs
[2016-05-25 20:23:13,129] {models.py:157} INFO - Filling up the DagBag from 
/Users/chrisr/airflow/dags
[2016-05-25 20:23:13,230] {jobs.py:732} INFO - Starting 1 scheduler jobs
[2016-05-25 20:23:13,257] {jobs.py:678} ERROR - (sqlite3.IntegrityError) UNIQUE 
constraint failed: dag_run.dag_id, dag_run.run_id [SQL: u'INSERT INTO dag_run 
(dag_id, execution_date, start_date, end_date, state, run_id, external_trigger, 
conf) VALUES (?, ?, ?, ?, ?, ?, ?, ?)'] [parameters: ('example_xcom', 
'2015-01-01 00:00:00.000000', '2016-05-25 20:23:13.254863', None, u'running', 
u'scheduled__2015-01-01T00:00:00', 0, None)]
Traceback (most recent call last):
  File "/Users/chrisr/Code/airflow/airflow/jobs.py", line 674, in _do_dags
    self.schedule_dag(dag)
  File "/Users/chrisr/Code/airflow/airflow/jobs.py", line 464, in schedule_dag
    external_trigger=False
  File "/Users/chrisr/Code/airflow/airflow/utils/db.py", line 53, in wrapper
    result = func(*args, **kwargs)
  File "/Users/chrisr/Code/airflow/airflow/models.py", line 3122, in 
create_dagrun
    session.commit()
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/session.py", line 
801, in commit
    self.transaction.commit()
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/session.py", line 
392, in commit
    self._prepare_impl()
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/session.py", line 
372, in _prepare_impl
    self.session.flush()
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/session.py", line 
2019, in flush
    self._flush(objects)
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/session.py", line 
2137, in _flush
    transaction.rollback(_capture_exception=True)
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/util/langhelpers.py", 
line 60, in __exit__
    compat.reraise(exc_type, exc_value, exc_tb)
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/session.py", line 
2101, in _flush
    flush_context.execute()
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/unitofwork.py", line 
373, in execute
    rec.execute(self)
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/unitofwork.py", line 
532, in execute
    uow
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/persistence.py", line 
174, in save_obj
    mapper, table, insert)
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/persistence.py", line 
800, in _emit_insert_statements
    execute(statement, params)
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/engine/base.py", line 
914, in execute
    return meth(self, multiparams, params)
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/sql/elements.py", line 
323, in _execute_on_connection
    return connection._execute_clauseelement(self, multiparams, params)
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/engine/base.py", line 
1010, in _execute_clauseelement
    compiled_sql, distilled_params
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/engine/base.py", line 
1146, in _execute_context
    context)
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/engine/base.py", line 
1341, in _handle_dbapi_exception
    exc_info
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/util/compat.py", line 
202, in raise_from_cause
    reraise(type(exception), exception, tb=exc_tb, cause=cause)
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/engine/base.py", line 
1139, in _execute_context
    context)
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/engine/default.py", line 
450, in do_execute
    cursor.execute(statement, parameters)
IntegrityError: (sqlite3.IntegrityError) UNIQUE constraint failed: 
dag_run.dag_id, dag_run.run_id [SQL: u'INSERT INTO dag_run (dag_id, 
execution_date, start_date, end_date, state, run_id, external_trigger, conf) 
VALUES (?, ?, ?, ?, ?, ?, ?, ?)'] [parameters: ('example_xcom', '2015-01-01 
00:00:00.000000', '2016-05-25 20:23:13.254863', None, u'running', 
u'scheduled__2015-01-01T00:00:00', 0, None)]
[2016-05-25 20:23:13,268] {jobs.py:747} INFO - Done queuing tasks, calling the 
executor's heartbeat
[2016-05-25 20:23:13,269] {jobs.py:750} INFO - Loop took: 0.14371 seconds
[2016-05-25 20:23:13,273] {models.py:308} INFO - Finding 'running' jobs without 
a recent heartbeat
[2016-05-25 20:23:13,273] {models.py:314} INFO - Failing jobs without heartbeat 
after 2016-05-25 20:20:58.273414
{noformat}

This is when a clean Airflow install from master, with JUST example_xcom 
enabled.


was (Author: criccomini):
I noticed that the scheduler log shows:

{noformat}
[2016-05-25 20:22:37,925] {jobs.py:580} INFO - Prioritizing 0 queued jobs
[2016-05-25 20:22:37,933] {jobs.py:732} INFO - Starting 0 scheduler jobs
[2016-05-25 20:22:37,933] {jobs.py:747} INFO - Done queuing tasks, calling the 
executor's heartbeat
[2016-05-25 20:22:37,933] {jobs.py:750} INFO - Loop took: 0.011795 seconds
[2016-05-25 20:22:37,936] {models.py:308} INFO - Finding 'running' jobs without 
a recent heartbeat
[2016-05-25 20:22:37,937] {models.py:314} INFO - Failing jobs without heartbeat 
after 2016-05-25 20:20:22.937222
[2016-05-25 20:22:42,925] {jobs.py:580} INFO - Prioritizing 0 queued jobs
[2016-05-25 20:22:42,934] {jobs.py:732} INFO - Starting 1 scheduler jobs
[2016-05-25 20:22:42,977] {models.py:2703} INFO - Checking state for <DagRun 
example_xcom @ 2016-05-25 20:22:42.953808: 
scheduled__2016-05-25T20:22:42.953808, externally triggered: False>
[2016-05-25 20:22:42,983] {jobs.py:504} INFO - Getting list of tasks to skip 
for active runs.
[2016-05-25 20:22:42,986] {jobs.py:520} INFO - Checking dependencies on 3 tasks 
instances, minus 0 skippable ones
[2016-05-25 20:22:42,991] {base_executor.py:36} INFO - Adding to queue: airflow 
run example_xcom push 2016-05-25T20:22:42.953808 --local -sd 
DAGS_FOLDER/example_dags/example_xcom.py 
[2016-05-25 20:22:42,993] {base_executor.py:36} INFO - Adding to queue: airflow 
run example_xcom push_by_returning 2016-05-25T20:22:42.953808 --local -sd 
DAGS_FOLDER/example_dags/example_xcom.py 
[2016-05-25 20:22:43,011] {jobs.py:747} INFO - Done queuing tasks, calling the 
executor's heartbeat
[2016-05-25 20:22:43,012] {jobs.py:750} INFO - Loop took: 0.089461 seconds
[2016-05-25 20:22:43,018] {models.py:308} INFO - Finding 'running' jobs without 
a recent heartbeat
[2016-05-25 20:22:43,019] {models.py:314} INFO - Failing jobs without heartbeat 
after 2016-05-25 20:20:28.019143
[2016-05-25 20:22:43,028] {sequential_executor.py:26} INFO - Executing command: 
airflow run example_xcom push 2016-05-25T20:22:42.953808 --local -sd 
DAGS_FOLDER/example_dags/example_xcom.py 
[2016-05-25 20:22:43,453] {__init__.py:36} INFO - Using executor 
SequentialExecutor
Logging into: 
/Users/chrisr/airflow/logs/example_xcom/push/2016-05-25T20:22:42.953808
[2016-05-25 20:22:44,300] {__init__.py:36} INFO - Using executor 
SequentialExecutor
[2016-05-25 20:22:48,937] {sequential_executor.py:26} INFO - Executing command: 
airflow run example_xcom push_by_returning 2016-05-25T20:22:42.953808 --local 
-sd DAGS_FOLDER/example_dags/example_xcom.py 
[2016-05-25 20:22:49,366] {__init__.py:36} INFO - Using executor 
SequentialExecutor
Logging into: 
/Users/chrisr/airflow/logs/example_xcom/push_by_returning/2016-05-25T20:22:42.953808
[2016-05-25 20:22:50,210] {__init__.py:36} INFO - Using executor 
SequentialExecutor
[2016-05-25 20:22:54,844] {jobs.py:580} INFO - Prioritizing 0 queued jobs
[2016-05-25 20:22:54,853] {jobs.py:732} INFO - Starting 1 scheduler jobs
[2016-05-25 20:22:54,903] {models.py:2703} INFO - Checking state for <DagRun 
example_xcom @ 2015-01-01 00:00:00: scheduled__2015-01-01T00:00:00, externally 
triggered: False>
[2016-05-25 20:22:54,907] {models.py:2703} INFO - Checking state for <DagRun 
example_xcom @ 2016-05-25 20:22:42.953808: 
scheduled__2016-05-25T20:22:42.953808, externally triggered: False>
[2016-05-25 20:22:54,911] {jobs.py:504} INFO - Getting list of tasks to skip 
for active runs.
[2016-05-25 20:22:54,913] {jobs.py:520} INFO - Checking dependencies on 6 tasks 
instances, minus 2 skippable ones
[2016-05-25 20:22:54,920] {base_executor.py:36} INFO - Adding to queue: airflow 
run example_xcom push 2015-01-01T00:00:00 --local -sd 
DAGS_FOLDER/example_dags/example_xcom.py 
[2016-05-25 20:22:54,921] {base_executor.py:36} INFO - Adding to queue: airflow 
run example_xcom push_by_returning 2015-01-01T00:00:00 --local -sd 
DAGS_FOLDER/example_dags/example_xcom.py 
[2016-05-25 20:22:54,935] {base_executor.py:36} INFO - Adding to queue: airflow 
run example_xcom puller 2016-05-25T20:22:42.953808 --local -sd 
DAGS_FOLDER/example_dags/example_xcom.py 
[2016-05-25 20:22:54,954] {jobs.py:747} INFO - Done queuing tasks, calling the 
executor's heartbeat
[2016-05-25 20:22:54,954] {jobs.py:750} INFO - Loop took: 0.113319 seconds
[2016-05-25 20:22:54,960] {models.py:308} INFO - Finding 'running' jobs without 
a recent heartbeat
[2016-05-25 20:22:54,960] {models.py:314} INFO - Failing jobs without heartbeat 
after 2016-05-25 20:20:39.960629
[2016-05-25 20:22:54,978] {sequential_executor.py:26} INFO - Executing command: 
airflow run example_xcom push_by_returning 2015-01-01T00:00:00 --local -sd 
DAGS_FOLDER/example_dags/example_xcom.py 
[2016-05-25 20:22:55,410] {__init__.py:36} INFO - Using executor 
SequentialExecutor
Logging into: 
/Users/chrisr/airflow/logs/example_xcom/push_by_returning/2015-01-01T00:00:00
[2016-05-25 20:22:56,239] {__init__.py:36} INFO - Using executor 
SequentialExecutor
[2016-05-25 20:23:00,873] {sequential_executor.py:26} INFO - Executing command: 
airflow run example_xcom push 2015-01-01T00:00:00 --local -sd 
DAGS_FOLDER/example_dags/example_xcom.py 
[2016-05-25 20:23:01,477] {__init__.py:36} INFO - Using executor 
SequentialExecutor
Logging into: /Users/chrisr/airflow/logs/example_xcom/push/2015-01-01T00:00:00
[2016-05-25 20:23:02,639] {__init__.py:36} INFO - Using executor 
SequentialExecutor
[2016-05-25 20:23:07,150] {sequential_executor.py:26} INFO - Executing command: 
airflow run example_xcom puller 2016-05-25T20:22:42.953808 --local -sd 
DAGS_FOLDER/example_dags/example_xcom.py 
[2016-05-25 20:23:07,620] {__init__.py:36} INFO - Using executor 
SequentialExecutor
Logging into: 
/Users/chrisr/airflow/logs/example_xcom/puller/2016-05-25T20:22:42.953808
[2016-05-25 20:23:08,498] {__init__.py:36} INFO - Using executor 
SequentialExecutor
[2016-05-25 20:23:13,128] {jobs.py:580} INFO - Prioritizing 0 queued jobs
[2016-05-25 20:23:13,129] {models.py:157} INFO - Filling up the DagBag from 
/Users/chrisr/airflow/dags
[2016-05-25 20:23:13,230] {jobs.py:732} INFO - Starting 1 scheduler jobs
[2016-05-25 20:23:13,257] {jobs.py:678} ERROR - (sqlite3.IntegrityError) UNIQUE 
constraint failed: dag_run.dag_id, dag_run.run_id [SQL: u'INSERT INTO dag_run 
(dag_id, execution_date, start_date, end_date, state, run_id, external_trigger, 
conf) VALUES (?, ?, ?, ?, ?, ?, ?, ?)'] [parameters: ('example_xcom', 
'2015-01-01 00:00:00.000000', '2016-05-25 20:23:13.254863', None, u'running', 
u'scheduled__2015-01-01T00:00:00', 0, None)]
Traceback (most recent call last):
  File "/Users/chrisr/Code/airflow/airflow/jobs.py", line 674, in _do_dags
    self.schedule_dag(dag)
  File "/Users/chrisr/Code/airflow/airflow/jobs.py", line 464, in schedule_dag
    external_trigger=False
  File "/Users/chrisr/Code/airflow/airflow/utils/db.py", line 53, in wrapper
    result = func(*args, **kwargs)
  File "/Users/chrisr/Code/airflow/airflow/models.py", line 3122, in 
create_dagrun
    session.commit()
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/session.py", line 
801, in commit
    self.transaction.commit()
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/session.py", line 
392, in commit
    self._prepare_impl()
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/session.py", line 
372, in _prepare_impl
    self.session.flush()
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/session.py", line 
2019, in flush
    self._flush(objects)
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/session.py", line 
2137, in _flush
    transaction.rollback(_capture_exception=True)
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/util/langhelpers.py", 
line 60, in __exit__
    compat.reraise(exc_type, exc_value, exc_tb)
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/session.py", line 
2101, in _flush
    flush_context.execute()
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/unitofwork.py", line 
373, in execute
    rec.execute(self)
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/unitofwork.py", line 
532, in execute
    uow
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/persistence.py", line 
174, in save_obj
    mapper, table, insert)
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/persistence.py", line 
800, in _emit_insert_statements
    execute(statement, params)
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/engine/base.py", line 
914, in execute
    return meth(self, multiparams, params)
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/sql/elements.py", line 
323, in _execute_on_connection
    return connection._execute_clauseelement(self, multiparams, params)
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/engine/base.py", line 
1010, in _execute_clauseelement
    compiled_sql, distilled_params
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/engine/base.py", line 
1146, in _execute_context
    context)
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/engine/base.py", line 
1341, in _handle_dbapi_exception
    exc_info
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/util/compat.py", line 
202, in raise_from_cause
    reraise(type(exception), exception, tb=exc_tb, cause=cause)
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/engine/base.py", line 
1139, in _execute_context
    context)
  File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/engine/default.py", line 
450, in do_execute
    cursor.execute(statement, parameters)
IntegrityError: (sqlite3.IntegrityError) UNIQUE constraint failed: 
dag_run.dag_id, dag_run.run_id [SQL: u'INSERT INTO dag_run (dag_id, 
execution_date, start_date, end_date, state, run_id, external_trigger, conf) 
VALUES (?, ?, ?, ?, ?, ?, ?, ?)'] [parameters: ('example_xcom', '2015-01-01 
00:00:00.000000', '2016-05-25 20:23:13.254863', None, u'running', 
u'scheduled__2015-01-01T00:00:00', 0, None)]
[2016-05-25 20:23:13,268] {jobs.py:747} INFO - Done queuing tasks, calling the 
executor's heartbeat
[2016-05-25 20:23:13,269] {jobs.py:750} INFO - Loop took: 0.14371 seconds
[2016-05-25 20:23:13,273] {models.py:308} INFO - Finding 'running' jobs without 
a recent heartbeat
[2016-05-25 20:23:13,273] {models.py:314} INFO - Failing jobs without heartbeat 
after 2016-05-25 20:20:58.273414
{noformat}

This is when a clean Airflow install from master, with JUST example_xcom 
enabled.

> schedule_interval @once scheduling dag atleast twice
> ----------------------------------------------------
>
>                 Key: AIRFLOW-168
>                 URL: https://issues.apache.org/jira/browse/AIRFLOW-168
>             Project: Apache Airflow
>          Issue Type: Bug
>          Components: scheduler
>    Affects Versions: Airflow 1.7.1.2
>            Reporter: Sumit Maheshwari
>         Attachments: Screen Shot 2016-05-24 at 9.51.50 PM.png, 
> screenshot-1.png
>
>
> I was looking at example_xcom example and found that it got scheduled twice. 
> Ones at the start_time and ones at the current time. To be correct I tried 
> multiple times (by reloading db) and its same. 
> I am on airflow master, using sequential executor with sqlite3. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to