And, I guess, I found another issue.
With `store_dag_code` enabled scheduler throws the following error:
```
Traceback (most recent call last):
File "/usr/local/lib/python3.7/multiprocessing/process.py", line
297, in _bootstrap
self.run()
File "/usr/local/lib/python3.7/multiprocessing/process.py", line 99, in run
self._target(*self._args, **self._kwargs)
File "/usr/local/lib/python3.7/site-packages/airflow/jobs/scheduler_job.py",
line 158, in _run_file_processor
pickle_dags)
File "/usr/local/lib/python3.7/site-packages/airflow/utils/db.py",
line 74, in wrapper
return func(*args, **kwargs)
File "/usr/local/lib/python3.7/site-packages/airflow/jobs/scheduler_job.py",
line 1582, in process_file
dag.sync_to_db()
File "/usr/local/lib/python3.7/site-packages/airflow/utils/db.py",
line 74, in wrapper
return func(*args, **kwargs)
File "/usr/local/lib/python3.7/site-packages/airflow/models/dag.py",
line 1519, in sync_to_db
DagCode.bulk_sync_to_db([dag.fileloc for dag in orm_dag])
TypeError: 'DagModel' object is not iterable
```
Looks like this is caused by a mistake during backporting of
https://github.com/apache/airflow/commit/e146518#diff-e5cbc8f771ec50ccb79ad8505f6f5697R1533
as
https://github.com/apache/airflow/commit/eb308e9#diff-e5cbc8f771ec50ccb79ad8505f6f5697R1519
The `orm_dag` variable in the 1.10 branch is produced from `DagModel`
with `.first()` -- which is a single model instance; while in master
`orm_dags` is
produced with `.all()` -- hence the "object is not iterable" error.
Best,
Kostya