Skip to main content
Scour
Browse
Getting Started
Login
Sign Up
You are offline. Trying to reconnect...
Close
Copied to clipboard
Close
Unable to share or copy to clipboard
Close
🧠 Local AI
local models, LLM inference, Ollama, self-hosted AI
Filter Results
Timeframe
Fresh
Past Hour
Today
This Week
This Month
Feeds to Scour
Subscribed
All
Scoured
148739
posts in
11.4
ms
Inference
Arena
– new
benchmark
of local inference and training
⚗️
BEAM Ecosystem
kvark.github.io
·
4d
·
Hacker News
Blink: CPU-Free LLM Inference by
Delegating
the Serving Stack to GPU and
SmartNIC
⚙️
Systems Programming
arxiv.org
·
1h
RetroInfer
: A
Vector
Storage Engine for Scalable Long-Context LLM Inference
🐫
OCaml
vldb.org
·
1d
The Engine Behind Modern LLM Inference, Part 1: Continuous
Batching
,
PagedAttention
, and the End of…
🐫
OCaml
medium.com
·
12h
Ollama
is still the
easiest
way to start local LLMs, but it's the worst way to keep running them
🐫
OCaml
xda-developers.com
·
1d
Best Open Source
Offline
AI Agent
🤖
AI Engineering
news.ycombinator.com
·
15h
·
Hacker News
LLM
inference
engine from
scratch
in C++
🐫
OCaml
anirudhsathiya.com
·
4d
·
Hacker News
(123)
Hermes
Agent Local Ai Setup Guide with Qwen3.5 +
OpenWebUI
💧
Elixir
youtube.com
·
19h
n8n
+
Ollama
: self-hosted AI automation that actually works
🏠
Self-Hosting
lemmy.world
·
2d
We
burned
$200 learning that local AI
wasn
't about the money
🤖
AI Engineering
write.as
·
14h
Decentralized
AI in 50
Lines
of Python
🐫
OCaml
iamtrask.github.io
·
2d
The Rise of Local AI (and Why It Won’t
Replace
Cloud AI)
🤖
AI Engineering
jasoneckert.github.io
·
6d
My AI Learning
Journey
– Part 4
🤖
AI Engineering
blog.wirelessmoves.com
·
1d
TurboQuant
Explained: Extreme AI
Compression
for Faster, Cheaper LLM Inference and Vector Search
🐫
OCaml
medium.com
·
4d
EU's
Exposed
AI Infrastructure
⚗️
BEAM Ecosystem
insecurestack.substack.com
·
2d
·
Substack
Run
Qwen3.5
on an Old Laptop: A
Lightweight
Local Agentic AI Setup Guide
🤖
AI Engineering
kdnuggets.com
·
1d
Local AI vs Cloud AI: When Does Each Make
Sense
🤖
AI Engineering
jakubjirak.medium.com
·
3d
Running
Gemma
4 Locally with
Ollama
on Your PC
💧
Elixir
analyticsvidhya.com
·
1d
Inside LLM Inference: KV Cache,
Prefill
, and the
Decode
Bottleneck
⚙️
Compiler Design
pub.towardsai.net
·
1d
How I Test Local Ai LLMs
🤖
AI Engineering
digitalspaceport.com
·
6d
Loading...
Loading more...
Page 2 »
Keyboard Shortcuts
Navigation
Next / previous item
j
/
k
Open post
o
or
Enter
Preview post
v
Post Actions
Love post
a
Like post
l
Dislike post
d
Undo reaction
u
Save / unsave
s
Recommendations
Add interest / feed
Enter
Not interested
x
Go to
Home
g
h
Interests
g
i
Feeds
g
f
Likes
g
l
History
g
y
Changelog
g
c
Settings
g
s
Browse
g
b
Search
/
Pagination
Next page
n
Previous page
p
General
Show this help
?
Submit feedback
!
Close modal / unfocus
Esc
Press
?
anytime to show this help