gpt4 book ai didi

elasticsearch - Airflow 弹性日志配置

转载 作者:行者123 更新时间:2023-12-02 22:28:15 46 4
gpt4 key购买 nike

我在Apache Airflow中设置弹性日志记录时遇到一些问题。
从版本1.10开始,弹性日志记录已添加到配置中。

查看airflow.cfg文件时,我们有两个与Elastic相关的部分:

# Airflow can store logs remotely in AWS S3, Google Cloud Storage or Elastic Search.
# Users must supply an Airflow connection id that provides access to the storage
# location. If remote_logging is set to true, see UPDATING.md for additional
# configuration requirements.
remote_logging = True
remote_log_conn_id =
remote_base_log_folder =
encrypt_s3_logs = False

[elasticsearch]
elasticsearch_host = xxx.xxx.xxx.xxx
elasticsearch_log_id_template = {dag_id}-{task_id}-{execution_date}-{try_number}
elasticsearch_end_of_log_mark = end_of_log



现在我不确定如何设置。查看airflow_local_settings.py文件时,我们可以看到以下代码:
if REMOTE_LOGGING and REMOTE_BASE_LOG_FOLDER.startswith('s3://'):
DEFAULT_LOGGING_CONFIG['handlers'].update(REMOTE_HANDLERS['s3'])
elif REMOTE_LOGGING and REMOTE_BASE_LOG_FOLDER.startswith('gs://'):
DEFAULT_LOGGING_CONFIG['handlers'].update(REMOTE_HANDLERS['gcs'])
elif REMOTE_LOGGING and REMOTE_BASE_LOG_FOLDER.startswith('wasb'):
DEFAULT_LOGGING_CONFIG['handlers'].update(REMOTE_HANDLERS['wasb'])
elif REMOTE_LOGGING and ELASTICSEARCH_HOST:
DEFAULT_LOGGING_CONFIG['handlers'].update(REMOTE_HANDLERS['elasticsearch'])

因此,从逻辑上讲,如果我将远程日志记录设置为True,然后将elastic的host/ip放在elastic部分中,则它应该可以工作。
目前, Airflow 实例未生成任何日志。

最佳答案

根据Airflow ElasticsearchTaskHandler doc

    ElasticsearchTaskHandler is a python log handler that
reads logs from Elasticsearch. Note logs are not directly
indexed into Elasticsearch. Instead, it flushes logs
into local files. Additional software setup is required
to index the log into Elasticsearch, such as using
Filebeat and Logstash.

不幸的是,该日志处理程序不会将日志直接刷新到您的ES。

关于elasticsearch - Airflow 弹性日志配置,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/54770033/

46 4 0
Copyright 2021 - 2024 cfsdn All Rights Reserved 蜀ICP备2022000587号
广告合作:1813099741@qq.com 6ren.com