support for vLLM 0.19.0 OpenAI-compatible chat endpoints and fixes the Qwen reasoning toggle so flash mode can actually disable thinking. Co-authored-by: NmanQAQ <normangyao@qq.com> Co-authored-by: Willem Jiang <willem.jiang@gmail.com> |
||
|---|---|---|
| .. | ||
| agents | ||
| community | ||
| config | ||
| guardrails | ||
| mcp | ||
| models | ||
| reflection | ||
| runtime | ||
| sandbox | ||
| skills | ||
| subagents | ||
| tools | ||
| tracing | ||
| uploads | ||
| utils | ||
| __init__.py | ||
| client.py | ||