Skip to main content
Scour
Browse
Getting Started
Login
Sign Up
You are offline. Trying to reconnect...
Copied to clipboard
Unable to share or copy to clipboard
LocalLlama
reddit.com
Unsloth
Qwen3.6
unsloth.ai
·
2w
·
r/LocalLLaMA
Vektor-Memory/Vex
:
Vex
- Vector Exchange - By
Vektor
Memory - A Cross-standard Vector DB migration tool - Open Source
github.com
·
3d
·
r/LocalLLaMA
Implemented
TurboQuant
and results don’t fully match paper
github.com
·
3d
·
r/LocalLLaMA
Eamon2009/Quadtrix.cpp
: AI engine built in C++ and python to run Language Models directly on your own computer. It skips the need for expensive hardware by optimizing
github.com
·
3d
·
r/LocalLLaMA
Unsloth
solved bug in
Mistral
Medium 3.5 implementation
unsloth.ai
·
3d
·
r/LocalLLaMA
A Dark-Money Campaign Is Paying
Influencers
to
Frame
Chinese AI as a Threat
wired.com
·
4d
·
Hacker News
,
r/China
,
r/LocalLLaMA
,
r/OpenAI
,
r/artificial
,
r/baba
A tool that REMOVES
censorship
from ANY open-weight LLM with a single
click
github.com
·
8w
·
Hacker News
,
r/LocalLLaMA
KV Cache Quantization —
WikiText-2
PPL sweep (llama3.2:3b, llama3.1:8b, qwen2.5:7b, qwen3.5:9b,
gemma4
:27b) on Tesla P40
gist.github.com
·
4d
·
r/LocalLLaMA
How people ask Claude for personal
guidance
anthropic.com
·
5d
·
Hacker News
,
r/ChatGPT
,
r/LocalLLaMA
Advanced
Quantization
Algorithm for
LLMs/VLMs
github.com
·
54w
·
Hacker News
,
r/LocalLLaMA
Closed-loop
SVG
Images
github.com
·
4d
·
r/LocalLLaMA
Blog: AI
evals
are becoming the new compute
bottleneck
evalevalai.com
·
1w
·
r/LocalLLaMA
angrygiraffe/claude-opus-4.6-4.7-reasoning-8.7k
huggingface.co
·
4d
·
r/LocalLLaMA
Running llama.cpp on Snapdragon
Hexagon
NPU
seems promising
github.com
·
4d
·
r/LocalLLaMA
nvidia/Gemma-4-26B-A4B-NVFP4
huggingface.co
·
5d
·
r/LocalLLaMA
Got
hipfire
running in Docker on my RX 7900 XTX alongside
llamacpp
github.com
·
5d
·
r/LocalLLaMA
froggeric/Mistral-Medium-3.5-128B-MLX-4bit
huggingface.co
·
5d
·
r/LocalLLaMA
My
Calculator
Is a
Transformer
sinclairs.gitlab.io
·
5d
·
Hacker News
,
r/LocalLLaMA
noonghunna/club-3090
: Community recipes for serving LLMs on RTX 3090. Multi-engine (vLLM, llama.cpp,
SGLang
) and model-agnostic. Currently shipping Qwen3.6-27B configs for 1× and 2× cards.
github.com
·
5d
·
r/LocalLLaMA
Can't replicate Reddit numbers with Qwen
27B
on a
3090TI
.
huggingface.co
·
5d
·
r/LocalLLaMA
« Page 1
·
Page 3 »
Log in to enable infinite scrolling
Keyboard Shortcuts
Navigation
Next / previous item
j
/
k
Open post
o
or
Enter
Preview post
v
Post Actions
Love post
a
Like post
l
Dislike post
d
Undo reaction
u
Save / unsave
s
Recommendations
Add interest / feed
Enter
Not interested
x
Go to
Home
g
h
Interests
g
i
Feeds
g
f
Likes
g
l
History
g
y
Changelog
g
c
Settings
g
s
Browse
g
b
Search
/
Pagination
Next page
n
Previous page
p
General
Show this help
?
Submit feedback
!
Close modal / unfocus
Esc
Press
?
anytime to show this help