@Kostya Thanks for testing the RC, finding the bugs and providing enough details for me to validate.
I am canceling the vote to fix these bugs. Regards, Kaxil On Sun, Apr 5, 2020 at 1:31 AM Kostya Esmukov <kostya.sh...@gmail.com> wrote: > And, I guess, I found another issue. > > With `store_dag_code` enabled scheduler throws the following error: > > ``` > Traceback (most recent call last): > File "/usr/local/lib/python3.7/multiprocessing/process.py", line > 297, in _bootstrap > self.run() > File "/usr/local/lib/python3.7/multiprocessing/process.py", line 99, in > run > self._target(*self._args, **self._kwargs) > File > "/usr/local/lib/python3.7/site-packages/airflow/jobs/scheduler_job.py", > line 158, in _run_file_processor > pickle_dags) > File "/usr/local/lib/python3.7/site-packages/airflow/utils/db.py", > line 74, in wrapper > return func(*args, **kwargs) > File > "/usr/local/lib/python3.7/site-packages/airflow/jobs/scheduler_job.py", > line 1582, in process_file > dag.sync_to_db() > File "/usr/local/lib/python3.7/site-packages/airflow/utils/db.py", > line 74, in wrapper > return func(*args, **kwargs) > File "/usr/local/lib/python3.7/site-packages/airflow/models/dag.py", > line 1519, in sync_to_db > DagCode.bulk_sync_to_db([dag.fileloc for dag in orm_dag]) > TypeError: 'DagModel' object is not iterable > ``` > > Looks like this is caused by a mistake during backporting of > > https://github.com/apache/airflow/commit/e146518#diff-e5cbc8f771ec50ccb79ad8505f6f5697R1533 > as > > https://github.com/apache/airflow/commit/eb308e9#diff-e5cbc8f771ec50ccb79ad8505f6f5697R1519 > > The `orm_dag` variable in the 1.10 branch is produced from `DagModel` > with `.first()` -- which is a single model instance; while in master > `orm_dags` is > produced with `.all()` -- hence the "object is not iterable" error. > > Best, > Kostya >