Add stream output support
1
#21 opened 2 days ago
by
JamePeng2023
Add full tool calling support to chat template
5
#20 opened 2 days ago
by
CISCai
安装vllm出现bug
#15 opened 9 days ago
by
sanwuge
How to use infinity long context with LLMxMapReduce?
1
#14 opened 9 days ago
by
lixiangtian
Run on CPU
#13 opened 9 days ago
by
J22
Speed slower than Qwen7B
#4 opened 14 days ago
by
MonolithFoundation
V100上无法直接推理
1
#3 opened 14 days ago
by
MonolithFoundation
这个有官方支持在ollama上部署的吗?
2
#2 opened 14 days ago
by
zhaoyang0618
Adding `safetensors` variant of this model
#1 opened 14 days ago
by
SFconvertbot