Skip to main content
Scour
Browse
Getting Started
Login
Sign Up
You are offline. Trying to reconnect...
Copied to clipboard
Unable to share or copy to clipboard
LocalLlama
reddit.com
Structured
CoT
: Shorter Reasoning with a
Grammar
File
andthattoo.dev
·
1w
·
r/LocalLLaMA
The 2026 AI
Index
Report
hai.stanford.edu
·
3w
·
Hacker News
,
r/LocalLLaMA
[Models]
Cohere
MoE by
Terrencezzj
· Pull Request #40817
github.com
·
1w
·
r/LocalLLaMA
Gemma 4 and Qwen 3.6 with
q8
_0 and q4_0 KV cache: KL
divergence
results
localbench.substack.com
·
1w
·
r/LocalLLaMA
An update on
recent
Claude Code quality reports
anthropic.com
·
1w
·
DEV
,
Hacker News
,
r/ClaudeAI
,
r/GithubCopilot
,
r/LocalLLaMA
mHC
:
Manifold-Constrained
Hyper-Connections
arxiv.org
·
17w
·
Hacker News
,
Hacker News
,
r/LocalLLaMA
Max-and-Omnis/Nemotron-3-Super-64B-A12B-Math-REAP-GGUF
huggingface.co
·
1w
·
r/LocalLLaMA
DeepSeek-V4
released
huggingface.co
·
1w
·
r/LocalLLaMA
,
r/singularity
Should I get macbook Air
M2
or
M4
apple.com
·
43w
·
r/LocalLLaMA
,
r/programming
To run deepseek
v4
flash how much max
vram
we need? 175 gb or 320gb?
lushbinary.com
·
1w
·
r/LocalLLaMA
Do you really want the US to “win” AI?
geohot.github.io
·
1w
·
Hacker News
,
r/LocalLLaMA
An Overnight Stack for Qwen3.6–27B: 85 TPS, 125K Context, Vision — on One RTX 3090 | by
Wasif
Basharat
medium.com
·
1w
·
r/LocalLLaMA
tencent/Hy3-preview
huggingface.co
·
1w
·
r/LocalLLaMA
China's BOE Is Drowning In Its Own Success, And Memory Players
CXMT
And
YMTC
Are Next
wccftech.com
·
1w
·
r/LocalLLaMA
OpenAI Privacy
Filter
huggingface.co
·
1w
·
Hacker News
,
r/LocalLLaMA
unsloth
Qwen3.6-27B-GGUF
huggingface.co
·
1w
·
r/LocalLLaMA
Qwen3.6-27B-GGUF
is here!
huggingface.co
·
1w
·
r/LocalLLaMA
Qwen
3.6
27B
is out
huggingface.co
·
1w
·
Hacker News
,
r/LocalLLaMA
Youtuber tries Qwen 3.5
35B
, Qwen 3.6
35B
, and Gemma 4
27b
to reverse engineer some large JS, with good results for Qwen 3.6
youtube.com
·
2w
·
r/LocalLLaMA
eb1386/1386.ai
: A lightweight transformer language model built from scratch in PyTorch, trained on a single consumer GPU with a full pipeline for data processing,
pretraining
, and instruction tuning.
github.com
·
2w
·
r/LocalLLaMA
« Page 4
·
Page 6 »
Log in to enable infinite scrolling
Keyboard Shortcuts
Navigation
Next / previous item
j
/
k
Open post
o
or
Enter
Preview post
v
Post Actions
Love post
a
Like post
l
Dislike post
d
Undo reaction
u
Save / unsave
s
Recommendations
Add interest / feed
Enter
Not interested
x
Go to
Home
g
h
Interests
g
i
Feeds
g
f
Likes
g
l
History
g
y
Changelog
g
c
Settings
g
s
Browse
g
b
Search
/
Pagination
Next page
n
Previous page
p
General
Show this help
?
Submit feedback
!
Close modal / unfocus
Esc
Press
?
anytime to show this help