[2026-03-10 13:58:28] Starting inference alias server on 127.0.0.1:8803 [2026-03-10 13:58:28] Backend: http://127.0.0.1:8800 [inference-alias] listening on 127.0.0.1:8803, backend=http://127.0.0.1:8800, aliases=glm,glm-4.7,qwen-122b,qwen-35b,qwen-397b [inference-alias] 127.0.0.1 - "GET /health HTTP/1.1" 200 - [inference-alias] 127.0.0.1 - "GET /v1/models HTTP/1.1" 200 - [inference-alias] 127.0.0.1 - "GET /v1/models HTTP/1.1" 200 - [inference-alias] 127.0.0.1 - "POST /v1/chat/completions HTTP/1.1" 200 - [inference-alias] 127.0.0.1 - "POST /v1/chat/completions HTTP/1.1" 200 - [2026-03-10 13:59:30] Starting inference alias server on 127.0.0.1:8803 [2026-03-10 13:59:30] Backend: http://127.0.0.1:8800 [inference-alias] listening on 127.0.0.1:8803, backend=http://127.0.0.1:8800, aliases=glm,glm-4.7,qwen-122b,qwen-35b,qwen-397b [inference-alias] 127.0.0.1 - "POST /v1/chat/completions HTTP/1.1" 200 - [inference-alias] 127.0.0.1 - "POST /v1/chat/completions HTTP/1.1" 200 - [inference-alias] 127.0.0.1 - "GET /v1/models HTTP/1.1" 200 - [inference-alias] 127.0.0.1 - "POST /v1/chat/completions HTTP/1.1" 200 - [inference-alias] 127.0.0.1 - "GET /v1/models HTTP/1.1" 200 - [inference-alias] 127.0.0.1 - "POST /v1/chat/completions HTTP/1.1" 200 - [inference-alias] 127.0.0.1 - "POST /v1/chat/completions HTTP/1.1" 200 - [inference-alias] 127.0.0.1 - "GET /v1/models HTTP/1.1" 200 - [inference-alias] 127.0.0.1 - "GET /v1/models HTTP/1.1" 200 - [inference-alias] 127.0.0.1 - "GET /v1/models HTTP/1.1" 200 - [inference-alias] 127.0.0.1 - "GET /v1/models HTTP/1.1" 200 -