FYI, 

I've had this issue for a few different reasons, the first was that my worker 
failed to read the file (this for me was a permission issue). Now I'm having it 
a second time and it appears
 Airflow timeouts connecting to the DB while setting up the task, for me I can 
see this in the stdout of the Airflow worker (not in stderr though).
 
I am looking to increase the connect timeout using this config item: 
https://airflow.readthedocs.io/en/latest/configurations-ref.html#sql-alchemy-connect-args
 and this sqlalchemy option: https://stackoverflow.com/a/35640876/2958068 

(side note: concerningly this option doesn't seem to be documented in the new 
documentation https://airflow.apache.org/docs/stable/configurations-ref.html . 
But I found the exact code in question: 
https://github.com/apache/airflow/blob/6ec9664a6be151e8fa94ef207cd98b459d392af2/airflow/settings.py#L187
 )

This is a snippet of the exception I am getting, it seems to be caused during 
session.merge(self) in _check_and_change_state_before_execution. Check if you 
are logging your Airflow worker stdout somewhere and see if you have a similar 
exception:

Traceback (most recent call last):
  File ".../lib/python3.7/site-packages/pymysql/connections.py", line 583, in 
connect
    **kwargs)
  File ".../lib/python3.7/socket.py", line 727, in create_connection
    raise err
  File ".../lib/python3.7/socket.py", line 716, in create_connection
    sock.connect(sa)
socket.timeout: timed out

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File ".../lib/python3.7/site-packages/sqlalchemy/engine/base.py", line 2276, 
in _wrap_pool_connect
    return fn()
  ...
  File ".../lib/python3.7/site-packages/pymysql/connections.py", line 630, in 
connect
    raise exc
pymysql.err.OperationalError: (2003, "Can't connect to MySQL server on 
'my_db.my_domain.net' (timed out)")

The above exception was the direct cause of the following exception:

Traceback (most recent call last):
  File ".../bin/airflow", line 37, in <module>
    args.func(args)
  ...
  File ".../lib/python3.7/site-packages/airflow/models/taskinstance.py", line 
852, in _check_and_change_state_before_execution
    session.merge(self)
  ...
  File ".../lib/python3.7/site-packages/pymysql/connections.py", line 630, in 
connect
    raise exc
sqlalchemy.exc.OperationalError: (pymysql.err.OperationalError) (2003, "Can't 
connect to MySQL server on 'my_db.my_domain.net' (timed out)")

Regards
Damian 


-----Original Message-----
From: Mehmet Ersoy <mehmet.ersoy1...@gmail.com> 
Sent: Wednesday, February 26, 2020 03:18
To: dev@airflow.apache.org
Subject: Re: task failed without running

Hi Heng,
I have same problem. Failed jobs do not dump log files. Do you have this 
problem too?
Have you been using Redis as message queue? Also can you send configurations 
related parallelism and concurrent in airflow.cfg file?

Best regards,
Mehmet.

heng gu <heng...@yahoo.com.invalid>, 24 Şub 2020 Pzt, 17:51 tarihinde şunu
yazdı:

> I have this dag with a branchpythonoperator task kicking off many of 
> 24 tasks, in this case, 4 tasks. 2 of the tasks were successful, the 
> other two (register_YZ, register_ZY) were failed without running (see 
> the attached UI screen shots). There is no log for tasks register_YZ 
> and register_ZY. I am using Celery Executor and running 12 workers 
> executing register_XX tasks. I am using airflow version 1.10.6. Any idea how 
> to fix it?
>


--
Mehmet ERSOY



=============================================================================== 
Please access the attached hyperlink for an important electronic communications 
disclaimer: 
http://www.credit-suisse.com/legal/en/disclaimer_email_ib.html 
=============================================================================== 

Reply via email to