Skip to main content
Scour
Browse
Getting Started
Login
Sign Up
You are offline. Trying to reconnect...
Close
Copied to clipboard
Close
Unable to share or copy to clipboard
Close
🧠 Local AI
local models, LLM inference, Ollama, self-hosted AI
Filter Results
Timeframe
Fresh
Past Hour
Today
This Week
This Month
Feeds to Scour
Subscribed
All
Scoured
187425
posts in
12.0
ms
alibekashirali/deepowl
: Local AI tutor for your private documents. Runs entirely on your machine.
🗃️
PKM
github.com
·
15h
·
r/SideProject
Using a local LLM in
OpenCode
with
llama.cpp
🐫
OCaml
aayushgarg.dev
·
5d
Show HN: A privacy first
companion
AI
🤖
AI Engineering
friendai.pro
·
5d
·
Hacker News
Ollama
Server
Component
Recommendations
☎️
OTP
lemmy.world
·
5d
When Hidden States Drift: Can
KV
Caches
Rescue Long-Range Speculative Decoding?
🔁
Spaced Repetition
arxiv.org
·
23h
Home Assistant's local LLM support
outperforms
Gemini for Home, and Google
knows
it
🏠
Self-Hosting
xda-developers.com
·
2d
ShotSieve
: Free Local AI Photo
Culling
Software
💧
Elixir
carlosfelic.io
·
4d
Linux kernel's ‘second-in-command’ uses local AI bot to hunt bugs, powered by '
clanker
' system with AMD's Ryzen AI Max+ — Framework Desktop has
resulted
in close to two dozen patches
⚙️
Systems Programming
tomshardware.com
·
3d
Linux Weekly 6:
9950X3D2
, Local AI on Intel GPUs, and Forbidden
Routers
⚙️
Systems Programming
youtube.com
·
6d
hi!! if you are willing, could you be our private ai eye on @
5ner
? i am pretty positive it’s ai, but your
discerning
appraisal would be much appreciated!!
💧
Elixir
8pxl.tumblr.com
·
6d
Doza
Assist
🗃️
PKM
producthunt.com
·
3d
Show HN-style: Blue
Arrow
– modular
orchestration
system with state-driven execution, local LLaMA integration and post-execution verification
☎️
OTP
github.com
·
2h
·
r/node
These feature make the
ThinkPad
P16
Gen 3 a local AI powerhouse
🗃️
PKM
notebookcheck.net
·
3d
I
replaced
NotebookLM
with a local LLM, and the difference is night and day
🗃️
PKM
xda-developers.com
·
2d
Network Edge Inference for Large Language Models:
Principles
,
Techniques
, and Opportunities
🐫
OCaml
arxiv.org
·
2d
noonghunna/club-3090
: Community recipes for serving LLMs on RTX 3090. Multi-engine (vLLM, llama.cpp,
SGLang
) and model-agnostic. Currently shipping Qwen3.6-27B configs for 1× and 2× cards.
💧
Elixir
github.com
·
15h
·
r/LocalLLaMA
Less Is More: Engineering Challenges of On-Device Small Language Model
Integration
in a Mobile
Application
⚙️
Compiler Design
arxiv.org
·
2d
DepthKV
: Layer-Dependent
KV
Cache Pruning for Long-Context LLM Inference
🔁
Spaced Repetition
arxiv.org
·
2d
heymrun/heym
: Self-hosted AI workflow automation visual canvas, agents, RAG,
HITL
, MCP, and observability in one runtime.
🤖
AI Engineering
github.com
·
2d
·
DEV
PolyKV
: A Shared
Asymmetrically-Compressed
KV Cache Pool for Multi-Agent LLM Inference
📡
Erlang
arxiv.org
·
1d
Sign up or log in to see more results
Sign Up
Login
« Page 2
Log in to enable infinite scrolling
Keyboard Shortcuts
Navigation
Next / previous item
j
/
k
Open post
o
or
Enter
Preview post
v
Post Actions
Love post
a
Like post
l
Dislike post
d
Undo reaction
u
Save / unsave
s
Recommendations
Add interest / feed
Enter
Not interested
x
Go to
Home
g
h
Interests
g
i
Feeds
g
f
Likes
g
l
History
g
y
Changelog
g
c
Settings
g
s
Browse
g
b
Search
/
Pagination
Next page
n
Previous page
p
General
Show this help
?
Submit feedback
!
Close modal / unfocus
Esc
Press
?
anytime to show this help