-
Notifications
You must be signed in to change notification settings - Fork 16
Issues: OpenCSGs/llm-inference
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Api server was blocked when LLM deployment scaling config beyond the cluster resouces
#134
by SeanHH86
was closed May 8, 2024
Error happen when do inference for wukong dtype=bfloat16 of use default transformer pipeline load model
#128
by SeanHH86
was closed Apr 30, 2024
avoid to ping huggingface when start serving to speed up the deployement
#117
by depenglee1707
was closed Apr 24, 2024
vllm, gguf, llamacpp, these integration cannot address local path of model
#116
by depenglee1707
was closed Apr 23, 2024
enable reset generate config on fly
enhancement
New feature or request
#104
by depenglee1707
was closed Apr 23, 2024
The usage introduction of Good for newcomers
llm-serve
is not correct in quick_start.md
good first issue
#100
by depenglee1707
was closed Apr 16, 2024
Generate incorrect text format when use pipeline defaulttransformers
#53
by SeanHH86
was closed Mar 27, 2024
Enhance inference API to support OpenAI style
enhancement
New feature or request
#52
by SeanHH86
was closed May 7, 2024
[BUG] Get error when try "translation" downstream model
#16
by depenglee1707
was closed Mar 11, 2024
ProTip!
What’s not been updated in a month: updated:<2025-02-11.