ElliotGao
tclf90
AI & ML interests
None yet
Recent Activity
liked a model about 1 hour ago
Soul-AILab/SoulX-Duplug-0.6B new activity 2 days ago
QuantTrio/Qwen3.5-27B-AWQ:AWQ 4-bit version of this Opus-Distilled-v2 model? updated a collection 2 days ago
Qwen3.5-AWQOrganizations
AWQ 4-bit version of this Opus-Distilled-v2 model?
8
#5 opened 3 days ago
by
0xburakcelik
--max-model-len 32768 seems a bit too small for agent use cases ?
3
#3 opened 21 days ago
by
edwarddukewu
Install & run QuantTrio/MiniMax-M2-AWQ easily using llmpm
👍 1
1
#8 opened 21 days ago
by
sarthak-saxena
My personal vLLM launch cmd on my old personal 2x3090 workstation
4
#1 opened about 1 month ago
by
tclf90
Can't get vLLM running on 1xRTX 4090
3
#1 opened about 1 month ago
by
slyfox1186
Easy to fall into infinite loop
👍 1
7
#2 opened about 1 month ago
by
dwaynedu
GLM-5-AWQ vLLM 部署指南
👍 1
2
#2 opened about 1 month ago
by
CharlesChen2023
Great work
5
#1 opened about 1 month ago
by
JoeyHwong
How run this model on Sglang?
1
#2 opened about 1 month ago
by
Salvadori
Anyone else getting only exclamation marks?
15
#3 opened about 1 month ago
by
Halbin
QuantTrio/Qwen3.5-397B-A17B-AWQ reponse is !
8
#5 opened about 1 month ago
by
duyuting
GPTQ int4-int8 mixed
2
#4 opened about 1 month ago
by
darkstar3537
--max-model-len 32768 ?
1
#1 opened 2 months ago
by
pathosethoslogos
The model startup using vllm failed.
10
#5 opened 3 months ago
by
beausoft
Accessing LLM, response without<think>start tag
5
#2 opened 3 months ago
by
sudage
Minimax-M2.1 AWQ Please
2
#6 opened 3 months ago
by
mtcl
How do you run this?
3
#1 opened 3 months ago
by
mtcl
Once again Thanks, here is my review for 8 x RTX 5090 setup
17
#2 opened 3 months ago
by
crystech
can this model be Quantized?
3
#4 opened 4 months ago
by
tinging
endless response
3
#5 opened 4 months ago
by
ramidahbash