Skip to content

Conversation

@Ceng23333
Copy link
Contributor

@Ceng23333 Ceng23333 commented Jan 26, 2026

#193

  1. /models接口返回正确modelid
  2. 非流式接口返回openai格式
  3. 响应请求中sampling params和chat_template_kwargs
  4. tokenizer decode调用参考vllm
  5. request调度检查

@Ceng23333 Ceng23333 requested a review from a team January 26, 2026 06:14
@Ceng23333 Ceng23333 changed the title Issue/193 Issue/193: inference_server适配部署需求 Jan 26, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants