Hi guys, So I have airflow 1.8 running at my company now. Overall, the performance have improved and scheduling has been faster. The jobs are running and the pipeline do progress but I am running into few issues. Please help if you have seen this before. Any help will be appreciated.
1. Jobs getting scheduled -> queued but not Running. Read an email from Bolke where the suggestion was to increase the size of Pools. But this hasn't worked. I manually cleared the tasks and saw airflow running them after clearing. 2. For the same above issue, I saw that there were Timeout errors seen: I still havent able to understand why this happens. This is the entire trace: [2017-03-22 19:35:16,332] {models.py:167} INFO - Filling up the DagBag from /usr/local/airflow/pipeline/pipeline.py [2017-03-22 19:35:22,451] {airflow_configuration.py:40} INFO - loading setup.cfg file [2017-03-22 19:35:51,041] {timeout.py:37} ERROR - Process timed out [2017-03-22 19:35:51,041] {models.py:266} ERROR - Failed to import: /usr/local/airflow/pipeline/pipeline.py Traceback (most recent call last): File "/usr/local/lib/python2.7/dist-packages/airflow/models.py", line 263, in process_file m = imp.load_source(mod_name, filepath) File "/usr/local/airflow/pipeline/pipeline.py", line 167, in <module> create_tasks(dbguid, version, dag, override_start_date) File "/usr/local/airflow/pipeline/pipeline.py", line 104, in create_tasks t = create_task(dbguid, dag, taskInfo, version, override_date) File "/usr/local/airflow/pipeline/pipeline.py", line 85, in create_task retries, 1, depends_on_past, version, override_dag_date) File "/usr/local/airflow/pipeline/dags/base_pipeline.py", line 90, in create_python_operator depends_on_past=depends_on_past) File "/usr/local/lib/python2.7/dist-packages/airflow/utils/decorators.py", line 86, in wrapper result = func(*args, **kwargs) File "/usr/local/lib/python2.7/dist-packages/airflow/operators/python_operator.py", line 65, in __init__ super(PythonOperator, self).__init__(*args, **kwargs) File "/usr/local/lib/python2.7/dist-packages/airflow/utils/decorators.py", line 70, in wrapper sig = signature(func) File "/usr/local/lib/python2.7/dist-packages/funcsigs/__init__.py", line 105, in signature return Signature.from_function(obj) File "/usr/local/lib/python2.7/dist-packages/funcsigs/__init__.py", line 594, in from_function __validate_parameters__=False) File "/usr/local/lib/python2.7/dist-packages/funcsigs/__init__.py", line 518, in __init__ for param in parameters)) File "/usr/lib/python2.7/collections.py", line 52, in __init__ self.__update(*args, **kwds) File "/usr/lib/python2.7/_abcoll.py", line 548, in update self[key] = value File "/usr/lib/python2.7/collections.py", line 61, in __setitem__ last[1] = root[0] = self.__map[key] = [last, root, key] File "/usr/local/lib/python2.7/dist-packages/airflow/utils/timeout.py", line 38, in handle_timeout raise AirflowTaskTimeout(self.error_message) AirflowTaskTimeout: Timeout 3. "_cmd" doesnt work anymore for fetching sqlalchemy_connection. Even when I am using mysql (connection url doesnt include 'sqlite' anywhere): "error: cannot use sqlite with the LocalExecutor"