Timon
KeyboardMasher
AI & ML interests
None yet
Recent Activity
new activity
1 day ago
qwp4w3hyb/gemma-3-27b-it-iMat-GGUF:Thanks!
new activity
about 1 month ago
bartowski/QVQ-72B-Preview-GGUF:llama.cpp inference too slow?
Organizations
None yet
KeyboardMasher's activity
Thanks!
#1 opened 1 day ago
by
KeyboardMasher
Ranked twice. This got to be a mistake.
#1 opened 3 days ago
by
KeyboardMasher
llama.cpp inference too slow?
3
#6 opened 2 months ago
by
ygsun

Over 2 tok/sec agg backed by NVMe SSD on 96GB RAM + 24GB VRAM AM5 rig with llama.cpp
9
#13 opened about 1 month ago
by
ubergarm
Issue with --n-gpu-layers 5 Parameter: Model Only Running on CPU
12
#10 opened 2 months ago
by
vuk123
Advice on running llama-server with Q2_K_L quant
3
#6 opened 2 months ago
by
vmajor

I loaded DeepSeek-V3-Q5_K_M up on my 10yrs old old Tesla M40 (Dell C4130)
3
#8 opened 2 months ago
by
gng2info
Model will need to be requantized, rope issues for long context
3
#2 opened 2 months ago
by
treehugg3
Instruct version?
3
#1 opened 4 months ago
by
KeyboardMasher
Feedback
1
#2 opened 4 months ago
by
KeyboardMasher
we need llama athene 3.1 70b
5
#5 opened 7 months ago
by
gopi87
Change the 'Original model' link to tree/9092a8a, which contains the updated weights.
1
#2 opened 4 months ago
by
AaronFeng753
Remove this model from Recent highlights collection
1
#9 opened 4 months ago
by
KeyboardMasher
Continuous output
8
#1 opened 5 months ago
by
kth8
Q8_0 file is damaged.
5
#1 opened 12 months ago
by
KeyboardMasher
Q8_0 file is damaged.
5
#1 opened 12 months ago
by
KeyboardMasher