[ https://issues.apache.org/jira/browse/AIRFLOW-168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15301461#comment-15301461 ]
Chris Riccomini edited comment on AIRFLOW-168 at 5/26/16 3:24 AM: ------------------------------------------------------------------ I noticed that the scheduler log shows (stacktrace at bottom): {noformat} [2016-05-25 20:22:37,925] {jobs.py:580} INFO - Prioritizing 0 queued jobs [2016-05-25 20:22:37,933] {jobs.py:732} INFO - Starting 0 scheduler jobs [2016-05-25 20:22:37,933] {jobs.py:747} INFO - Done queuing tasks, calling the executor's heartbeat [2016-05-25 20:22:37,933] {jobs.py:750} INFO - Loop took: 0.011795 seconds [2016-05-25 20:22:37,936] {models.py:308} INFO - Finding 'running' jobs without a recent heartbeat [2016-05-25 20:22:37,937] {models.py:314} INFO - Failing jobs without heartbeat after 2016-05-25 20:20:22.937222 [2016-05-25 20:22:42,925] {jobs.py:580} INFO - Prioritizing 0 queued jobs [2016-05-25 20:22:42,934] {jobs.py:732} INFO - Starting 1 scheduler jobs [2016-05-25 20:22:42,977] {models.py:2703} INFO - Checking state for <DagRun example_xcom @ 2016-05-25 20:22:42.953808: scheduled__2016-05-25T20:22:42.953808, externally triggered: False> [2016-05-25 20:22:42,983] {jobs.py:504} INFO - Getting list of tasks to skip for active runs. [2016-05-25 20:22:42,986] {jobs.py:520} INFO - Checking dependencies on 3 tasks instances, minus 0 skippable ones [2016-05-25 20:22:42,991] {base_executor.py:36} INFO - Adding to queue: airflow run example_xcom push 2016-05-25T20:22:42.953808 --local -sd DAGS_FOLDER/example_dags/example_xcom.py [2016-05-25 20:22:42,993] {base_executor.py:36} INFO - Adding to queue: airflow run example_xcom push_by_returning 2016-05-25T20:22:42.953808 --local -sd DAGS_FOLDER/example_dags/example_xcom.py [2016-05-25 20:22:43,011] {jobs.py:747} INFO - Done queuing tasks, calling the executor's heartbeat [2016-05-25 20:22:43,012] {jobs.py:750} INFO - Loop took: 0.089461 seconds [2016-05-25 20:22:43,018] {models.py:308} INFO - Finding 'running' jobs without a recent heartbeat [2016-05-25 20:22:43,019] {models.py:314} INFO - Failing jobs without heartbeat after 2016-05-25 20:20:28.019143 [2016-05-25 20:22:43,028] {sequential_executor.py:26} INFO - Executing command: airflow run example_xcom push 2016-05-25T20:22:42.953808 --local -sd DAGS_FOLDER/example_dags/example_xcom.py [2016-05-25 20:22:43,453] {__init__.py:36} INFO - Using executor SequentialExecutor Logging into: /Users/chrisr/airflow/logs/example_xcom/push/2016-05-25T20:22:42.953808 [2016-05-25 20:22:44,300] {__init__.py:36} INFO - Using executor SequentialExecutor [2016-05-25 20:22:48,937] {sequential_executor.py:26} INFO - Executing command: airflow run example_xcom push_by_returning 2016-05-25T20:22:42.953808 --local -sd DAGS_FOLDER/example_dags/example_xcom.py [2016-05-25 20:22:49,366] {__init__.py:36} INFO - Using executor SequentialExecutor Logging into: /Users/chrisr/airflow/logs/example_xcom/push_by_returning/2016-05-25T20:22:42.953808 [2016-05-25 20:22:50,210] {__init__.py:36} INFO - Using executor SequentialExecutor [2016-05-25 20:22:54,844] {jobs.py:580} INFO - Prioritizing 0 queued jobs [2016-05-25 20:22:54,853] {jobs.py:732} INFO - Starting 1 scheduler jobs [2016-05-25 20:22:54,903] {models.py:2703} INFO - Checking state for <DagRun example_xcom @ 2015-01-01 00:00:00: scheduled__2015-01-01T00:00:00, externally triggered: False> [2016-05-25 20:22:54,907] {models.py:2703} INFO - Checking state for <DagRun example_xcom @ 2016-05-25 20:22:42.953808: scheduled__2016-05-25T20:22:42.953808, externally triggered: False> [2016-05-25 20:22:54,911] {jobs.py:504} INFO - Getting list of tasks to skip for active runs. [2016-05-25 20:22:54,913] {jobs.py:520} INFO - Checking dependencies on 6 tasks instances, minus 2 skippable ones [2016-05-25 20:22:54,920] {base_executor.py:36} INFO - Adding to queue: airflow run example_xcom push 2015-01-01T00:00:00 --local -sd DAGS_FOLDER/example_dags/example_xcom.py [2016-05-25 20:22:54,921] {base_executor.py:36} INFO - Adding to queue: airflow run example_xcom push_by_returning 2015-01-01T00:00:00 --local -sd DAGS_FOLDER/example_dags/example_xcom.py [2016-05-25 20:22:54,935] {base_executor.py:36} INFO - Adding to queue: airflow run example_xcom puller 2016-05-25T20:22:42.953808 --local -sd DAGS_FOLDER/example_dags/example_xcom.py [2016-05-25 20:22:54,954] {jobs.py:747} INFO - Done queuing tasks, calling the executor's heartbeat [2016-05-25 20:22:54,954] {jobs.py:750} INFO - Loop took: 0.113319 seconds [2016-05-25 20:22:54,960] {models.py:308} INFO - Finding 'running' jobs without a recent heartbeat [2016-05-25 20:22:54,960] {models.py:314} INFO - Failing jobs without heartbeat after 2016-05-25 20:20:39.960629 [2016-05-25 20:22:54,978] {sequential_executor.py:26} INFO - Executing command: airflow run example_xcom push_by_returning 2015-01-01T00:00:00 --local -sd DAGS_FOLDER/example_dags/example_xcom.py [2016-05-25 20:22:55,410] {__init__.py:36} INFO - Using executor SequentialExecutor Logging into: /Users/chrisr/airflow/logs/example_xcom/push_by_returning/2015-01-01T00:00:00 [2016-05-25 20:22:56,239] {__init__.py:36} INFO - Using executor SequentialExecutor [2016-05-25 20:23:00,873] {sequential_executor.py:26} INFO - Executing command: airflow run example_xcom push 2015-01-01T00:00:00 --local -sd DAGS_FOLDER/example_dags/example_xcom.py [2016-05-25 20:23:01,477] {__init__.py:36} INFO - Using executor SequentialExecutor Logging into: /Users/chrisr/airflow/logs/example_xcom/push/2015-01-01T00:00:00 [2016-05-25 20:23:02,639] {__init__.py:36} INFO - Using executor SequentialExecutor [2016-05-25 20:23:07,150] {sequential_executor.py:26} INFO - Executing command: airflow run example_xcom puller 2016-05-25T20:22:42.953808 --local -sd DAGS_FOLDER/example_dags/example_xcom.py [2016-05-25 20:23:07,620] {__init__.py:36} INFO - Using executor SequentialExecutor Logging into: /Users/chrisr/airflow/logs/example_xcom/puller/2016-05-25T20:22:42.953808 [2016-05-25 20:23:08,498] {__init__.py:36} INFO - Using executor SequentialExecutor [2016-05-25 20:23:13,128] {jobs.py:580} INFO - Prioritizing 0 queued jobs [2016-05-25 20:23:13,129] {models.py:157} INFO - Filling up the DagBag from /Users/chrisr/airflow/dags [2016-05-25 20:23:13,230] {jobs.py:732} INFO - Starting 1 scheduler jobs [2016-05-25 20:23:13,257] {jobs.py:678} ERROR - (sqlite3.IntegrityError) UNIQUE constraint failed: dag_run.dag_id, dag_run.run_id [SQL: u'INSERT INTO dag_run (dag_id, execution_date, start_date, end_date, state, run_id, external_trigger, conf) VALUES (?, ?, ?, ?, ?, ?, ?, ?)'] [parameters: ('example_xcom', '2015-01-01 00:00:00.000000', '2016-05-25 20:23:13.254863', None, u'running', u'scheduled__2015-01-01T00:00:00', 0, None)] Traceback (most recent call last): File "/Users/chrisr/Code/airflow/airflow/jobs.py", line 674, in _do_dags self.schedule_dag(dag) File "/Users/chrisr/Code/airflow/airflow/jobs.py", line 464, in schedule_dag external_trigger=False File "/Users/chrisr/Code/airflow/airflow/utils/db.py", line 53, in wrapper result = func(*args, **kwargs) File "/Users/chrisr/Code/airflow/airflow/models.py", line 3122, in create_dagrun session.commit() File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/session.py", line 801, in commit self.transaction.commit() File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/session.py", line 392, in commit self._prepare_impl() File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/session.py", line 372, in _prepare_impl self.session.flush() File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/session.py", line 2019, in flush self._flush(objects) File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/session.py", line 2137, in _flush transaction.rollback(_capture_exception=True) File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/util/langhelpers.py", line 60, in __exit__ compat.reraise(exc_type, exc_value, exc_tb) File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/session.py", line 2101, in _flush flush_context.execute() File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/unitofwork.py", line 373, in execute rec.execute(self) File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/unitofwork.py", line 532, in execute uow File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/persistence.py", line 174, in save_obj mapper, table, insert) File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/persistence.py", line 800, in _emit_insert_statements execute(statement, params) File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/engine/base.py", line 914, in execute return meth(self, multiparams, params) File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/sql/elements.py", line 323, in _execute_on_connection return connection._execute_clauseelement(self, multiparams, params) File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/engine/base.py", line 1010, in _execute_clauseelement compiled_sql, distilled_params File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/engine/base.py", line 1146, in _execute_context context) File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/engine/base.py", line 1341, in _handle_dbapi_exception exc_info File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/util/compat.py", line 202, in raise_from_cause reraise(type(exception), exception, tb=exc_tb, cause=cause) File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/engine/base.py", line 1139, in _execute_context context) File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/engine/default.py", line 450, in do_execute cursor.execute(statement, parameters) IntegrityError: (sqlite3.IntegrityError) UNIQUE constraint failed: dag_run.dag_id, dag_run.run_id [SQL: u'INSERT INTO dag_run (dag_id, execution_date, start_date, end_date, state, run_id, external_trigger, conf) VALUES (?, ?, ?, ?, ?, ?, ?, ?)'] [parameters: ('example_xcom', '2015-01-01 00:00:00.000000', '2016-05-25 20:23:13.254863', None, u'running', u'scheduled__2015-01-01T00:00:00', 0, None)] [2016-05-25 20:23:13,268] {jobs.py:747} INFO - Done queuing tasks, calling the executor's heartbeat [2016-05-25 20:23:13,269] {jobs.py:750} INFO - Loop took: 0.14371 seconds [2016-05-25 20:23:13,273] {models.py:308} INFO - Finding 'running' jobs without a recent heartbeat [2016-05-25 20:23:13,273] {models.py:314} INFO - Failing jobs without heartbeat after 2016-05-25 20:20:58.273414 {noformat} This is when a clean Airflow install from master, with JUST example_xcom enabled. was (Author: criccomini): I noticed that the scheduler log shows: {noformat} [2016-05-25 20:22:37,925] {jobs.py:580} INFO - Prioritizing 0 queued jobs [2016-05-25 20:22:37,933] {jobs.py:732} INFO - Starting 0 scheduler jobs [2016-05-25 20:22:37,933] {jobs.py:747} INFO - Done queuing tasks, calling the executor's heartbeat [2016-05-25 20:22:37,933] {jobs.py:750} INFO - Loop took: 0.011795 seconds [2016-05-25 20:22:37,936] {models.py:308} INFO - Finding 'running' jobs without a recent heartbeat [2016-05-25 20:22:37,937] {models.py:314} INFO - Failing jobs without heartbeat after 2016-05-25 20:20:22.937222 [2016-05-25 20:22:42,925] {jobs.py:580} INFO - Prioritizing 0 queued jobs [2016-05-25 20:22:42,934] {jobs.py:732} INFO - Starting 1 scheduler jobs [2016-05-25 20:22:42,977] {models.py:2703} INFO - Checking state for <DagRun example_xcom @ 2016-05-25 20:22:42.953808: scheduled__2016-05-25T20:22:42.953808, externally triggered: False> [2016-05-25 20:22:42,983] {jobs.py:504} INFO - Getting list of tasks to skip for active runs. [2016-05-25 20:22:42,986] {jobs.py:520} INFO - Checking dependencies on 3 tasks instances, minus 0 skippable ones [2016-05-25 20:22:42,991] {base_executor.py:36} INFO - Adding to queue: airflow run example_xcom push 2016-05-25T20:22:42.953808 --local -sd DAGS_FOLDER/example_dags/example_xcom.py [2016-05-25 20:22:42,993] {base_executor.py:36} INFO - Adding to queue: airflow run example_xcom push_by_returning 2016-05-25T20:22:42.953808 --local -sd DAGS_FOLDER/example_dags/example_xcom.py [2016-05-25 20:22:43,011] {jobs.py:747} INFO - Done queuing tasks, calling the executor's heartbeat [2016-05-25 20:22:43,012] {jobs.py:750} INFO - Loop took: 0.089461 seconds [2016-05-25 20:22:43,018] {models.py:308} INFO - Finding 'running' jobs without a recent heartbeat [2016-05-25 20:22:43,019] {models.py:314} INFO - Failing jobs without heartbeat after 2016-05-25 20:20:28.019143 [2016-05-25 20:22:43,028] {sequential_executor.py:26} INFO - Executing command: airflow run example_xcom push 2016-05-25T20:22:42.953808 --local -sd DAGS_FOLDER/example_dags/example_xcom.py [2016-05-25 20:22:43,453] {__init__.py:36} INFO - Using executor SequentialExecutor Logging into: /Users/chrisr/airflow/logs/example_xcom/push/2016-05-25T20:22:42.953808 [2016-05-25 20:22:44,300] {__init__.py:36} INFO - Using executor SequentialExecutor [2016-05-25 20:22:48,937] {sequential_executor.py:26} INFO - Executing command: airflow run example_xcom push_by_returning 2016-05-25T20:22:42.953808 --local -sd DAGS_FOLDER/example_dags/example_xcom.py [2016-05-25 20:22:49,366] {__init__.py:36} INFO - Using executor SequentialExecutor Logging into: /Users/chrisr/airflow/logs/example_xcom/push_by_returning/2016-05-25T20:22:42.953808 [2016-05-25 20:22:50,210] {__init__.py:36} INFO - Using executor SequentialExecutor [2016-05-25 20:22:54,844] {jobs.py:580} INFO - Prioritizing 0 queued jobs [2016-05-25 20:22:54,853] {jobs.py:732} INFO - Starting 1 scheduler jobs [2016-05-25 20:22:54,903] {models.py:2703} INFO - Checking state for <DagRun example_xcom @ 2015-01-01 00:00:00: scheduled__2015-01-01T00:00:00, externally triggered: False> [2016-05-25 20:22:54,907] {models.py:2703} INFO - Checking state for <DagRun example_xcom @ 2016-05-25 20:22:42.953808: scheduled__2016-05-25T20:22:42.953808, externally triggered: False> [2016-05-25 20:22:54,911] {jobs.py:504} INFO - Getting list of tasks to skip for active runs. [2016-05-25 20:22:54,913] {jobs.py:520} INFO - Checking dependencies on 6 tasks instances, minus 2 skippable ones [2016-05-25 20:22:54,920] {base_executor.py:36} INFO - Adding to queue: airflow run example_xcom push 2015-01-01T00:00:00 --local -sd DAGS_FOLDER/example_dags/example_xcom.py [2016-05-25 20:22:54,921] {base_executor.py:36} INFO - Adding to queue: airflow run example_xcom push_by_returning 2015-01-01T00:00:00 --local -sd DAGS_FOLDER/example_dags/example_xcom.py [2016-05-25 20:22:54,935] {base_executor.py:36} INFO - Adding to queue: airflow run example_xcom puller 2016-05-25T20:22:42.953808 --local -sd DAGS_FOLDER/example_dags/example_xcom.py [2016-05-25 20:22:54,954] {jobs.py:747} INFO - Done queuing tasks, calling the executor's heartbeat [2016-05-25 20:22:54,954] {jobs.py:750} INFO - Loop took: 0.113319 seconds [2016-05-25 20:22:54,960] {models.py:308} INFO - Finding 'running' jobs without a recent heartbeat [2016-05-25 20:22:54,960] {models.py:314} INFO - Failing jobs without heartbeat after 2016-05-25 20:20:39.960629 [2016-05-25 20:22:54,978] {sequential_executor.py:26} INFO - Executing command: airflow run example_xcom push_by_returning 2015-01-01T00:00:00 --local -sd DAGS_FOLDER/example_dags/example_xcom.py [2016-05-25 20:22:55,410] {__init__.py:36} INFO - Using executor SequentialExecutor Logging into: /Users/chrisr/airflow/logs/example_xcom/push_by_returning/2015-01-01T00:00:00 [2016-05-25 20:22:56,239] {__init__.py:36} INFO - Using executor SequentialExecutor [2016-05-25 20:23:00,873] {sequential_executor.py:26} INFO - Executing command: airflow run example_xcom push 2015-01-01T00:00:00 --local -sd DAGS_FOLDER/example_dags/example_xcom.py [2016-05-25 20:23:01,477] {__init__.py:36} INFO - Using executor SequentialExecutor Logging into: /Users/chrisr/airflow/logs/example_xcom/push/2015-01-01T00:00:00 [2016-05-25 20:23:02,639] {__init__.py:36} INFO - Using executor SequentialExecutor [2016-05-25 20:23:07,150] {sequential_executor.py:26} INFO - Executing command: airflow run example_xcom puller 2016-05-25T20:22:42.953808 --local -sd DAGS_FOLDER/example_dags/example_xcom.py [2016-05-25 20:23:07,620] {__init__.py:36} INFO - Using executor SequentialExecutor Logging into: /Users/chrisr/airflow/logs/example_xcom/puller/2016-05-25T20:22:42.953808 [2016-05-25 20:23:08,498] {__init__.py:36} INFO - Using executor SequentialExecutor [2016-05-25 20:23:13,128] {jobs.py:580} INFO - Prioritizing 0 queued jobs [2016-05-25 20:23:13,129] {models.py:157} INFO - Filling up the DagBag from /Users/chrisr/airflow/dags [2016-05-25 20:23:13,230] {jobs.py:732} INFO - Starting 1 scheduler jobs [2016-05-25 20:23:13,257] {jobs.py:678} ERROR - (sqlite3.IntegrityError) UNIQUE constraint failed: dag_run.dag_id, dag_run.run_id [SQL: u'INSERT INTO dag_run (dag_id, execution_date, start_date, end_date, state, run_id, external_trigger, conf) VALUES (?, ?, ?, ?, ?, ?, ?, ?)'] [parameters: ('example_xcom', '2015-01-01 00:00:00.000000', '2016-05-25 20:23:13.254863', None, u'running', u'scheduled__2015-01-01T00:00:00', 0, None)] Traceback (most recent call last): File "/Users/chrisr/Code/airflow/airflow/jobs.py", line 674, in _do_dags self.schedule_dag(dag) File "/Users/chrisr/Code/airflow/airflow/jobs.py", line 464, in schedule_dag external_trigger=False File "/Users/chrisr/Code/airflow/airflow/utils/db.py", line 53, in wrapper result = func(*args, **kwargs) File "/Users/chrisr/Code/airflow/airflow/models.py", line 3122, in create_dagrun session.commit() File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/session.py", line 801, in commit self.transaction.commit() File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/session.py", line 392, in commit self._prepare_impl() File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/session.py", line 372, in _prepare_impl self.session.flush() File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/session.py", line 2019, in flush self._flush(objects) File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/session.py", line 2137, in _flush transaction.rollback(_capture_exception=True) File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/util/langhelpers.py", line 60, in __exit__ compat.reraise(exc_type, exc_value, exc_tb) File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/session.py", line 2101, in _flush flush_context.execute() File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/unitofwork.py", line 373, in execute rec.execute(self) File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/unitofwork.py", line 532, in execute uow File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/persistence.py", line 174, in save_obj mapper, table, insert) File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/orm/persistence.py", line 800, in _emit_insert_statements execute(statement, params) File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/engine/base.py", line 914, in execute return meth(self, multiparams, params) File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/sql/elements.py", line 323, in _execute_on_connection return connection._execute_clauseelement(self, multiparams, params) File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/engine/base.py", line 1010, in _execute_clauseelement compiled_sql, distilled_params File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/engine/base.py", line 1146, in _execute_context context) File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/engine/base.py", line 1341, in _handle_dbapi_exception exc_info File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/util/compat.py", line 202, in raise_from_cause reraise(type(exception), exception, tb=exc_tb, cause=cause) File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/engine/base.py", line 1139, in _execute_context context) File "build/bdist.macosx-10.10-intel/egg/sqlalchemy/engine/default.py", line 450, in do_execute cursor.execute(statement, parameters) IntegrityError: (sqlite3.IntegrityError) UNIQUE constraint failed: dag_run.dag_id, dag_run.run_id [SQL: u'INSERT INTO dag_run (dag_id, execution_date, start_date, end_date, state, run_id, external_trigger, conf) VALUES (?, ?, ?, ?, ?, ?, ?, ?)'] [parameters: ('example_xcom', '2015-01-01 00:00:00.000000', '2016-05-25 20:23:13.254863', None, u'running', u'scheduled__2015-01-01T00:00:00', 0, None)] [2016-05-25 20:23:13,268] {jobs.py:747} INFO - Done queuing tasks, calling the executor's heartbeat [2016-05-25 20:23:13,269] {jobs.py:750} INFO - Loop took: 0.14371 seconds [2016-05-25 20:23:13,273] {models.py:308} INFO - Finding 'running' jobs without a recent heartbeat [2016-05-25 20:23:13,273] {models.py:314} INFO - Failing jobs without heartbeat after 2016-05-25 20:20:58.273414 {noformat} This is when a clean Airflow install from master, with JUST example_xcom enabled. > schedule_interval @once scheduling dag atleast twice > ---------------------------------------------------- > > Key: AIRFLOW-168 > URL: https://issues.apache.org/jira/browse/AIRFLOW-168 > Project: Apache Airflow > Issue Type: Bug > Components: scheduler > Affects Versions: Airflow 1.7.1.2 > Reporter: Sumit Maheshwari > Attachments: Screen Shot 2016-05-24 at 9.51.50 PM.png, > screenshot-1.png > > > I was looking at example_xcom example and found that it got scheduled twice. > Ones at the start_time and ones at the current time. To be correct I tried > multiple times (by reloading db) and its same. > I am on airflow master, using sequential executor with sqlite3. -- This message was sent by Atlassian JIRA (v6.3.4#6332)