I just found out that airflow create some tmp config files in /tmp dir on
servers that run Celery workers. All the files have name tmpxxxxxxx, and the
same content like:
{"core": {"dags_folder": "/data/data366/airflow/dags", "base_log_folder":
"/data/data366/airflow/logs", "remote_logging": "False", "remote_log_conn_id":
"", "remote_base_log_folder": "", "encrypt_s3_logs": "False", "logging_level":
"INFO", "fab_logging_level": "WARN", "logging_config_class": "",
"colored_console_log": "True", "colored_log_format":
"[%%(blue)s%%(asctime)s%%(reset)s]
{%%(blue)s%%(filename)s:%%(reset)s%%(lineno)d}
%%(log_color)s%%(levelname)s%%(reset)s - %%(log_color)s%%(message)s%%(reset)s",
"colored_formatter_class":
"airflow.utils.log.colored_log.CustomTTYColoredFormatter", "log_format":
"[%%(asctime)s] {%%(filename)s:%%(lineno)d} %%(levelname)s - %%(message)s",
"simple_log_format": "%%(asctime)s %%(levelname)s - %%(message)s",
"log_filename_template": "{{ ti.dag_id }}/{{ ti.task_id }}/{{ ts }}/{{
try_number }}.log", "log_processor_filename_template": "{{ filename }}.log",
"dag_processor_manager_log_location":
"/data/data366/airflow/logs/dag_processor_manager/dag_processor_manager.log",
"hostname_callable": "socket:getfqdn", "default_timezone": "utc", "executor":
"CeleryExecutor", "sql_alchemy_conn":
"postgresql+psycopg2://airflow:Itpwpc12@rhw9133/airflow",
"sql_engine_encoding": "utf-8", "sql_alchemy_pool_enabled": "True",
"sql_alchemy_pool_size": "5", "sql_alchemy_max_overflow": "10",
"sql_alchemy_pool_recycle": "1800", "sql_alchemy_pool_pre_ping": "True",
"sql_alchemy_schema": "", "parallelism": "32", "dag_concurrency": "16",
"dags_are_paused_at_creation": "True", "max_active_runs_per_dag": "16",
"load_examples": "False", "plugins_folder": "/data/data366/airflow/plugins",
"fernet_key": "rTXn28iYM2m2KO1qgMygpd6Em9hXNCS700K4JtCU32k=", "donot_pickle":
"False", "dagbag_import_timeout": "30", "dag_file_processor_timeout": "50",
"task_runner": "StandardTaskRunner", "default_impersonation": "", "security":
"", "secure_mode": "False", "unit_test_mode": "False", "task_log_reader":
"task", "enable_xcom_pickling": "True", "killed_task_cleanup_time": "60",
"dag_run_conf_overrides_params": "False", "worker_precheck": "False",
"dag_discovery_safe_mode": "True", "default_task_retries": "0",
"sql_alchemy_reconnect_timeout": "300", "non_pooled_task_slot_count": "128"},
"cli": {"api_client": "airflow.api.client.local_client", "endpoint_url":
"http://localhost:8080"}, "api": {"auth_backend":
"airflow.api.auth.backend.default"}, "lineage": {"backend": ""}, "atlas":
{"sasl_enabled": "False", "host": "", "port": "21000", "username": "",
"password": ""}, "operators": {"default_owner": "Airflow", "default_cpus": "1",
"default_ram": "512", "default_disk": "512", "default_gpus": "0"}, "hive":
{"default_hive_mapred_queue": ""}, "webserver": {"base_url":
"http://localhost:8080", "web_server_host": "0.0.0.0", "web_server_port":
"8080", "web_server_ssl_cert": "", "web_server_ssl_key": "", "w
eb_server_master_timeout": "120", "web_server_worker_timeout": "120",
"worker_refresh_batch_size": "1", "worker_
Which airflow process creates the tmp file? Why cannot it be just one tmp file
or permanent file since all the tmp files have the same content?
Heng