gpt4 book ai didi

concurrency - gunicorn uvicorn worker.py 如何遵守 limit_concurrency 设置

转载 作者:行者123 更新时间:2023-12-04 01:15:35 44 4
gpt4 key购买 nike

FastAPI 使用 gunicorn 来启动 uvicorn worker,如 https://www.uvicorn.org/settings/ 中所述。
然而,gunicorn 不允许使用自定义设置启动 uvicorn,如 https://github.com/encode/uvicorn/issues/343 中所述。
该问题建议覆盖源文件中的 config_kwargs,如 https://github.com/encode/uvicorn/blob/master/uvicorn/workers.py
我们尝试过,但 uvicorn 不尊重设置 limit_concurrency在源中的多个 uvicorn 文件中:
https://github.com/encode/uvicorn/blob/master/uvicorn/workers.py

# fail

config_kwargs = {
"app": None,
"log_config": None,
"timeout_keep_alive": self.cfg.keepalive,
"timeout_notify": self.timeout,
"callback_notify": self.callback_notify,
"limit_max_requests": self.max_requests, "limit_concurrency": 10000,
"forwarded_allow_ips": self.cfg.forwarded_allow_ips,
}

https://github.com/encode/uvicorn/blob/master/uvicorn/main.py
# fail

kwargs = {
"app": app,
"host": host,
"port": port,
"uds": uds,
"fd": fd,
"loop": loop,
"http": http,
"ws": ws,
"lifespan": lifespan,
"env_file": env_file,
"log_config": LOGGING_CONFIG if log_config is None else log_config,
"log_level": log_level,
"access_log": access_log,
"interface": interface,
"debug": debug,
"reload": reload,
"reload_dirs": reload_dirs if reload_dirs else None,
"workers": workers,
"proxy_headers": proxy_headers,
"forwarded_allow_ips": forwarded_allow_ips,
"root_path": root_path,
"limit_concurrency": 10000,
"backlog": backlog,
"limit_max_requests": limit_max_requests,
"timeout_keep_alive": timeout_keep_alive,
"ssl_keyfile": ssl_keyfile,
"ssl_certfile": ssl_certfile,
"ssl_version": ssl_version,
"ssl_cert_reqs": ssl_cert_reqs,
"ssl_ca_certs": ssl_ca_certs,
"ssl_ciphers": ssl_ciphers,
"headers": list([header.split(":") for header in headers]),
"use_colors": use_colors,
}

uvicorn 怎么能被迫遵守这个设定呢?我们仍然收到来自 FastAPI 的 503 错误
- - - -更新 - - - - - -
枪炮设定 --worker-connections 1000当发出 100 个分发给许多工作人员的并行请求时,仍然会导致 503。
但是,我认为这是一个更复杂的问题:我们的 API 端点做了很多繁重的工作,通常需要 5 秒才能完成。
2 个内核、2 个 worker 的压力测试:
  • A. 100+ 并发请求,端点重负载 --worker-connections 1
  • B. 100+ 并发请求,端点重负载 --worker-connections 1000
  • C. 100+ 并发请求,端点低负载 --worker-connections 1
  • D. 100+ 并发请求,端点低负载 --worker-connections 1000

  • 两个实验 A 和 B 都产生了 503 个响应,因此假设 worker-connections 设置确实有效,太多的模拟连接似乎不会导致我们的 503 错误。
    我们对这种行为感到困惑,因为我们希望 gunicorn/uvicorn 将工作排队,而不是抛出 503 错误。

    最佳答案

    来自 gunicorn doc

    worker-connections

    The maximum number of simultaneous clients.


    来自 uvicorn doc

    limit-concurrency

    Maximum number of concurrent connections or tasks to allow, before issuing HTTP 503 responses.


    根据此信息,两个设置变量都在做同样的事情。所以
    uvicorn --limit-concurrency 100 application:demo_app
    几乎一样
    gunicorn --worker-connections 100 -k uvicorn.workers.UvicornWorker application:demo_app
    注:我还没有对此进行任何实际测试,如果我错了,请纠正我。

    此外,您可以设置 limit-concurrency (或 limit_concurrency )通过子类化 uvicorn.workers.UvicornWorker 类(class)
    from uvicorn.workers import UvicornWorker


    class CustomUvicornWorker(UvicornWorker):
    CONFIG_KWARGS = {
    "loop": "uvloop",
    "http": "httptools",
    "limit_concurrency": 100
    }
    现在使用这个 CustomUvicornWorkergunicorn命令为,
    gunicorn -k path.to.custom_worker.CustomUvicornWorker application:demo_app
    注:您可以检查 self.config.limit_concurrencyCustomUvicornWorker类以确保该值已正确设置。

    关于concurrency - gunicorn uvicorn worker.py 如何遵守 limit_concurrency 设置,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/63471960/

    44 4 0
    Copyright 2021 - 2024 cfsdn All Rights Reserved 蜀ICP备2022000587号
    广告合作:1813099741@qq.com 6ren.com