请问下 如果想微调 能给个案列吗?
#30 opened 6 days ago
by
wa595566
Exceptional Release, This is The Most Powerful OSS Model for 24GB Cards
👍
2
#29 opened 20 days ago
by
AaronFeng753
super bad german
#28 opened 28 days ago
by
abiteddie
Prevent tokenizer calls from returning `token_type_ids`
❤️
1
2
#26 opened about 1 month ago
by
joaogante
Hiding Thinking Process
2
#25 opened about 2 months ago
by
Xzendor7
Draft model for speculative decoding
#24 opened about 2 months ago
by
oobabooga
这模型数学能力很强
2
#23 opened about 2 months ago
by
pypry
Suppose I have 96g vram is it possible to do fine tuning?
1
#22 opened about 2 months ago
by
win10
Would love to see coding version :)
👍
2
2
#21 opened about 2 months ago
by
AekDevDev
chat_template:it keeps all thinking content?
1
#20 opened about 2 months ago
by
TAnGx1411
Getting error when deploying on HF Inference with 2 A100 GPU's of AWS on region us-east
#19 opened about 2 months ago
by
streebo
FP8 Quants Please
➕
1
#18 opened 2 months ago
by
rjmehta
online demo please
👍
1
#15 opened 2 months ago
by
pypry
vram Requirements for full size
2
#14 opened 2 months ago
by
tazomatalax
demo Inference
➕
2
#13 opened 2 months ago
by
devops724
bug in chat template?
1
#11 opened 2 months ago
by
J22
Why is the chat_template mixed with Chinese and English?
👍
2
5
#8 opened 2 months ago
by
Daucloud
is docker image available?
#7 opened 2 months ago
by
ZhifengKong
Deployment support for sglang
👍
3
#5 opened 2 months ago
by
XiChen0415
vllm error:operator _C::marlin_qqq_gemm does not exist
3
#4 opened 2 months ago
by
HourseCircle
Category Error
👍
1
1
#3 opened 2 months ago
by
CO-IR
Sorry for Askin here
👀
5
#2 opened 2 months ago
by
ryg81
Official vllm support
👀
2
1
#1 opened 2 months ago
by
shash42