Skip to main content
Scour
Browse
Getting Started
Login
Sign Up
You are offline. Trying to reconnect...
Close
You're currently offline. Some features may not work.
Close
Copied to clipboard
Close
Unable to share or copy to clipboard
Close
⚡ Glommio
io_uring Runtime, Thread-per-Core, Async IO, High Performance
Filter Results
Timeframe
Fresh
Past Hour
Today
This Week
This Month
Feeds to Scour
Subscribed
All
Scoured
25946
posts in
221.6
ms
Breaking the
Tractability
Barrier: A Generic Low-Level Solver for
NP-Hard
Instances (N=63) on Commodity 64-Bit Silicon
zenodo.org
·
52m
·
Discuss:
Hacker News
🧮
SMT Solvers
Linux 7.0
MM
Changes Bring Some Very Nice Performance
Optimizations
phoronix.com
·
9h
📋
Zero-Copy
Scaling llama.cpp On
Neoverse
N2: Solving
Cross-NUMA
Performance Issues
semiengineering.com
·
1d
🔮
Prefetching
UMEM
: Unified Memory Extraction and Management Framework for
Generalizable
Memory
arxiv.org
·
1d
🧠
Memory Management
An
async
HTTP server in ~80 lines of modern C++ (
coroutines
)
vixcpp.com
·
1d
·
Discuss:
Hacker News
💫
IO_uring
harishsg993010/tiny-NPU
: opensource NPU for LLM inference (this run
gpt2
)
github.com
·
15h
·
Discuss:
r/LocalLLaMA
🏗️
LLM Infrastructure
Functional
Optics
for Modern Java
blog.scottlogic.com
·
1d
🌊
Async Patterns
Go 1.26
unleashes
performance-boosting Green Tea
GC
infoworld.com
·
1d
🏹
Apache Arrow
Uncached
buffered
IO [LWN.net]
lwn.net
·
2d
💾
Disk I/O
AndPuQing/gflow
: A lightweight, single-node GPU job scheduler implemented in Rust.
github.com
·
1d
·
Discuss:
Hacker News
🔬
Rust Profiling
A high-performance Go framework (
250k
+
req/sec
) with built-in batteries
go-kvolt.github.io
·
2d
🚀
Async Optimization
Execution-Centric Characterization of
FP8
Matrix Cores, Asynchronous Execution, and Structured Sparsity on AMD
MI300A
arxiv.org
·
1d
🔮
Prefetching
Introducing GPT‑5.3‑
Codex
‑
Spark
simonwillison.net
·
13h
🖥
GPUs
How to build a distributed
queue
in a single
JSON
file on object storage
turbopuffer.com
·
1d
·
Discuss:
Lobsters
,
Hacker News
⚡
Zero-Copy Serialization
Supercharging
Inference for AI Factories: KV Cache
Offload
as a Memory-Hierarchy Problem
blog.min.io
·
19h
🏗️
LLM Infrastructure
[
AINews
] Qwen Image 2 and
Seedance
2
latent.space
·
2d
🏗️
LLM Infrastructure
Bitsum
. Real-time
CPU
Optimization and Automation
bitsum.com
·
1d
🔮
Prefetching
[Development] 4MB 32-bit
SRAM
for the
MicroMac
Performer
68kmla.org
·
11h
⚙️
Mechanical Sympathy
Introducing
Dedicated
Container Inference:
Delivering
2.6x faster inference for custom AI models
together.ai
·
1d
🏗️
LLM Infrastructure
ShareChat
hit a billion features per second, then it had to make it 10x
cheaper
thenewstack.io
·
20h
🏗️
Infrastructure Economics
Loading...
Loading more...
Page 2 »
Keyboard Shortcuts
Navigation
Next / previous item
j
/
k
Open post
o
or
Enter
Preview post
v
Post Actions
Love post
a
Like post
l
Dislike post
d
Undo reaction
u
Recommendations
Add interest / feed
Enter
Not interested
x
Go to
Home
g
h
Interests
g
i
Feeds
g
f
Likes
g
l
History
g
y
Changelog
g
c
Settings
g
s
Browse
g
b
Search
/
Pagination
Next page
n
Previous page
p
General
Show this help
?
Submit feedback
!
Close modal / unfocus
Esc
Press
?
anytime to show this help