Skip to main content
Scour
Browse
Getting Started
Login
Sign Up
You are offline. Trying to reconnect...
Copied to clipboard
Unable to share or copy to clipboard
LocalLlama
reddit.com
Intel Arc Pro
B70
32GB 256-bit
GDDR6
PCI Express 5.0 x16, 32 Ray Tracing Units, 32 Xe Cores, AI and Workstation Graphics Card
newegg.com
·
4w
·
r/LocalLLaMA
llama.cpp
Gemma4
Tokenizer
Fix Was Merged Into Main Branch
github.com
·
4w
·
r/LocalLLaMA
ace-step/ACE-Step-1.5
: The most powerful local music generation model that
outperforms
most commercial alternatives
github.com
·
13w
·
Hacker News
,
Hacker News
,
r/LocalLLaMA
,
r/LocalLLaMA
Google strongly
implies
the
existence
of large Gemma 4 models
huggingface.co
·
4w
·
r/LocalLLaMA
paperscarecrow/Gemma-4-31B-it-abliterated-gguf
huggingface.co
·
4w
·
r/LocalLLaMA
philtomson/llama.cpp
: LLM inference in C/C++ (fork of
PrismML
fork that enables CPU (incl AVX2 and AVX512) and ROCm for AMD GPUs
github.com
·
4w
·
r/LocalLLaMA
google/gemma-4-31B-it
huggingface.co
·
4w
·
r/LocalLLaMA
MCP
Servers
store is released
mcp.so
·
73w
·
Hacker News
,
r/LocalLLaMA
A bug in
Bun
may have been the
root
cause of the Claude Code source code leak.
github.com
·
5w
·
Hacker News
,
r/LocalLLaMA
,
r/programming
TinyGPU
– Use AMD and Nvidia
GPUs
on macOS with Tinygrad
docs.tinygrad.org
·
5w
·
Hacker News
,
r/LocalLLaMA
Jackrong/Qwen3.5-27B-Claude-4.6-Opus-Reasoning-Distilled
huggingface.co
·
9w
·
Hacker News
,
r/LocalLLaMA
TRL
v1.0: Post-Training Library That Holds When the Field
Invalidates
Its Own Assumptions
huggingface.co
·
5w
·
Hacker News
,
r/LocalLLaMA
Aider-AI/aider
:
aider
is AI
pair
programming in your terminal
github.com
·
57w
·
Hacker News
,
r/LocalLLaMA
Qwen 3.5 9B LLM
GGUF
quantized
for local structured extraction
huggingface.co
·
5w
·
r/LocalLLaMA
"The Child That Surpassed Both Parents"
Darwin-35B-A3B-Opus
(
35B/3B
MoE) with Model MRI Technique
huggingface.co
·
5w
·
r/LocalLLaMA
Yo-GPT
- a model you can run locally to
replicate
the iconic app
neurometric.ai
·
5w
·
Hacker News
,
r/LocalLLaMA
Pure C implementation of the
TurboQuant
paper (
ICLR
2026) for KV cache compression in LLM inference.
github.com
·
5w
·
r/LocalLLaMA
attn-rot
(
ggerganov
's "TurboQuant lite") is on the cusp of getting merged into llama.cpp
github.com
·
5w
·
r/LocalLLaMA
You guys seen this? 1-bit model with an
MMLU-R
of 65.7, 8B
params
huggingface.co
·
5w
·
r/LocalLLaMA
Training
mRNA
Language Models Across 25
Species
for $165
huggingface.co
·
5w
·
Hacker News
,
r/LocalLLaMA
« Page 12
·
Page 14 »
Log in to enable infinite scrolling
Keyboard Shortcuts
Navigation
Next / previous item
j
/
k
Open post
o
or
Enter
Preview post
v
Post Actions
Love post
a
Like post
l
Dislike post
d
Undo reaction
u
Save / unsave
s
Recommendations
Add interest / feed
Enter
Not interested
x
Go to
Home
g
h
Interests
g
i
Feeds
g
f
Likes
g
l
History
g
y
Changelog
g
c
Settings
g
s
Browse
g
b
Search
/
Pagination
Next page
n
Previous page
p
General
Show this help
?
Submit feedback
!
Close modal / unfocus
Esc
Press
?
anytime to show this help