Skip to main content
Scour
Browse
Getting Started
Login
Sign Up
You are offline. Trying to reconnect...
Close
Copied to clipboard
Close
Unable to share or copy to clipboard
Close
🤖 AI
Broad
Llama, qwen, OpenAI, Claude, Anthropic, GPUs, Ollama, Local LLMs
Filter Results
Timeframe
Fresh
Past Hour
Today
This Week
This Month
Feeds to Scour
Subscribed
All
Scoured
4889
posts in
9.5
ms
janit/viiwork
: LLM inference load balancer optimized for AMD Radeon VII GPUs
⚙️
MLOps
github.com
·
4d
·
Hacker News
smart
routing
across Google, OpenAI,
Anthropic
, and more
🕷️
Web Crawling
api.lxg2it.com
·
12h
·
Hacker News
EU's
Exposed
AI Infrastructure
📊
Self-Hosted Analytics
insecurestack.substack.com
·
2d
·
Substack
Use a gaming PC as a Google
Gemma
4 LLM server
💉
Prompt Injection
sgbarker.com
·
12h
·
Hacker News
UCCL-EP
: Portable
Expert-Parallel
Communication
🪝
eBPF
uccl-project.github.io
·
2d
·
Hacker News
Testing Open Source and
Commercial
LLMs – Can Anyone Beat Claude
Opus
?
💉
Prompt Injection
akitaonrails.com
·
4d
·
Hacker News
The BEST Local LLM for opencode ! Gemma 4
26B
A4B
. No GPU required
🪄
Prompt Engineering
grigio.org
·
5d
·
Hacker News
[
AINews
] Good Friday
✨
Gemini
latent.space
·
6d
Running Google Gemma 4 Locally with
LM
Studio's New
Headless
CLI and Claude Code
⚙️
MLOps
ai.georgeliu.com
·
4d
·
Hacker News
Llamafile
: Run AI Models
Locally
on Your PC with Just One File
💉
Prompt Injection
firethering.com
·
3d
·
Hacker News
benchmarking
inference
of popular models on consumer hardware
⚙️
MLOps
inferena.tech
·
5d
·
Hacker News
[
discussion
] the
BSDs
in the AI Age
🇨🇳
Chinese AI
lists.nycbug.org
·
3d
·
Lobsters
,
Hacker News
Luce-Org/luce-megakernel
:
Megakernel
to match Apple Silicon Efficiency at 2x the Throughput on a RTX 3090
🔐
Hardware Security
github.com
·
1d
·
Hacker News
mahadillahm4di-cyber/mh-gdpr-ai.eu-s-plus
: Switch GPT ↔ Claude ↔ Llama, never lose context. Your AI brain as 3D stars. 100% local — your memory belongs to you, not to OpenAI.
🛡️
AI Security
github.com
·
1d
·
Hacker News
PacifAIst/Quansloth
: Based on the implementation of Google's TurboQuant (ICLR 2026) —
Quansloth
brings elite KV cache compression to local LLM inference.
Quansloth
is a fully private, air-gapped AI server that runs massive context models natively on consumer hardware with ease
🇨🇳
Chinese AI
github.com
·
3d
·
Hacker News
ypollak2/llm-router
: Smart LLM router for Claude Code — auto-picks cheapest model per task,
routes
within Claude subscription first. 70-85% cost savings.
🤖
Agent Payments
github.com
·
4d
·
Hacker News
vLLM
introduces memory
optimizations
for long-context inference
🤖
LLM
github.com
·
5d
·
Hacker News
AmrDab/clawdcursor
: AI desktop agent — sees your screen, controls your cursor, completes tasks autonomously.
🎯
Cursor IDE
github.com
·
5d
·
Hacker News
milla-jovovich/mempalace
: The highest-scoring AI memory system ever benchmarked. And it's free.
⚡
Edge AI
github.com
·
3d
·
Hacker News
,
Hacker News
,
r/singularity
youngbryan97/aura
: Autonomous Cognitive Architecture — IIT 4.0, Global Workspace Theory, closed-loop
affective
steering via residual stream modulation
🔍
AI Interpretability
github.com
·
4d
·
Hacker News
,
r/SideProject
,
r/artificial
,
r/vibecoding
Loading...
Loading more...
Page 2 »
Keyboard Shortcuts
Navigation
Next / previous item
j
/
k
Open post
o
or
Enter
Preview post
v
Post Actions
Love post
a
Like post
l
Dislike post
d
Undo reaction
u
Save / unsave
s
Recommendations
Add interest / feed
Enter
Not interested
x
Go to
Home
g
h
Interests
g
i
Feeds
g
f
Likes
g
l
History
g
y
Changelog
g
c
Settings
g
s
Browse
g
b
Search
/
Pagination
Next page
n
Previous page
p
General
Show this help
?
Submit feedback
!
Close modal / unfocus
Esc
Press
?
anytime to show this help