Man Cub
mancub
ยท
AI & ML interests
None yet
Recent Activity
new activity 6 days ago
QuantTrio/Qwen3.5-27B-AWQ:My personal vLLM launch cmd on my old personal 2x3090 workstation new activity 6 days ago
mudler/gemma-4-26B-A4B-it-APEX-GGUF:What was just updated and why? new activity 17 days ago
adamjen/Devstral-Small-2-24B-Opus-Reasoning:How to use it with llama-server ?Organizations
None yet
My personal vLLM launch cmd on my old personal 2x3090 workstation
7
#1 opened about 1 month ago
by
tclf90
What was just updated and why?
๐ 1
2
#1 opened 6 days ago
by
mancub
How to use it with llama-server ?
๐ 1
3
#1 opened 17 days ago
by
mancub
How to split this model between 2 (3) GPUs and CPU/RAM ?
18
#12 opened 24 days ago
by
mancub
Poor performance and pretty lobotomized
2
#1 opened 23 days ago
by
mancub
Love the license, confused by some of the decisions.
๐ค๐ 15
15
#15 opened 24 days ago
by
CyborgPaloma
It's really good.
๐ 1
26
#3 opened about 1 month ago
by
Shuasimodo
Increasing the precision of some of the weights when quantizing
๐ 4
57
#2 opened about 2 months ago
by
Shuasimodo
A draft model with less parameters, for speculative thinking?
8
#5 opened 2 months ago
by
mancub
Jan 21: All GLM-4.7-Flash quants reuploaded - much better outputs!
๐ฅโค๏ธ 7
29
#10 opened 3 months ago
by
danielhanchen
Fast loras
2
#8 opened 3 months ago
by
melmass
Wan-Lighting : 4steps per model or 4steps total?
4
#59 opened 8 months ago
by
NielsGx
Can we have a Llama-3.1-8B-Lexi-Uncensored-V2_fp8_scaled.safetensors
๐ฅ 1
12
#10 opened 12 months ago
by
drguolai
Within Seconds ?
7
#8 opened about 1 year ago
by
Daemontatox
Is it censored output?
12
#2 opened about 1 year ago
by
KurtcPhotoED
Please work with llama.cpp before releasing new models.
2
#10 opened 12 months ago
by
bradhutchings
Lack of 33B models?
๐ 1
7
#1 opened over 2 years ago
by
mancub
No config.json ?
3
#1 opened almost 3 years ago
by
0x12d3
is this working properly?
23
#1 opened almost 3 years ago
by
Boffy
Uh oh, the "q's"...
27
#2 opened almost 3 years ago
by
mancub