Skip to main content
Scour
Browse
Getting Started
Login
Sign Up
You are offline. Trying to reconnect...
Close
Copied to clipboard
Close
Unable to share or copy to clipboard
Close
🔄 MLOps
ML pipelines, model deployment, MLflow, model monitoring
Filter Results
Timeframe
Fresh
Past Hour
Today
This Week
This Month
Feeds to Scour
Subscribed
All
Scoured
186577
posts in
15.0
ms
MCAP
: Deployment-Time Layer
Profiling
for Memory-Constrained LLM Inference
⚙️
AI Engineering
arxiv.org
·
6d
Flow
generation through natural language: An agentic
modeling
approach (11 minute read)
🔄
AI Workflows
shopify.engineering
·
1d
GoogleCloudPlatform/activation-model-scanner
:
Verify
language model safety before deployment by analyzing activation patterns
🗺️
Codebase Navigation
github.com
·
21h
·
Hacker News
How AI-Driven Kubernetes Optimization
Reclaimed
Millions from 47%
Idle
Capacity
⚙️
AI Engineering
engineering.salesforce.com
·
5h
Introducing
DigitalOcean
AI-Native Cloud for Production AI
Workloads
🔄
AI Workflows
digitalocean.com
·
2d
AI Infrastructure
Architect
·
Builder
· Author
⚙️
AI Engineering
markferraz.com
·
6h
·
Hacker News
Build Strands Agents with
SageMaker
AI models and
MLflow
🔄
AI Workflows
aws.amazon.com
·
3d
The Data
Layer
Tax for Robot Learning
🖼
Stable Diffusion
rerun.io
·
12h
·
Hacker News
Agentic Data Engineering with
Genie
Code and
Lakeflow
🔄
AI Workflows
databricks.com
·
2d
[
AINews
] The Inference
Inflection
🤖
Machine learning
latent.space
·
23h
Three
Cobblers
, One
Zhuge
Liang: Making Cheaper Models Work Together
🤖
AI Coding Tools
markhuang.ai
·
1d
·
Hacker News
Monitoring LLM behavior: Drift,
retries
, and
refusal
patterns
🧠
Context Engineering
venturebeat.com
·
5d
·
Hacker News
The
Inference
Economy:
Token
Use
🧠
Context Engineering
frontierai.substack.com
·
7h
·
Substack
Caltech
’s
PrismML
shrinks AI models to fit your phone without losing their mind
⚙️
AI Engineering
startupfortune.com
·
2d
An Empirical Study of Methods for
SFTing
Opaque
Reasoning Models
🧠
Context Engineering
lesswrong.com
·
6d
Adaptive and Fine-grained Module-wise Expert Pruning for Efficient
LoRA-MoE
Fine-Tuning
⚙️
AI Engineering
arxiv.org
·
21h
not much
happened
today
🔄
AI Workflows
news.smol.ai
·
2d
Load
balancer
for
vLLM
server instances?
🛠️
LLM Tooling
docs.vllm.ai
·
2d
·
r/LocalLLaMA
AmSach/kvquant
: Drop-in KV cache compressor for local LLM inference - Run 70B models on 8GB RAM
🤖
Machine learning
github.com
·
13h
·
DEV
OpenShift
AI observability
summarizer
: Transform metrics into meaning
🔄
AI Workflows
developers.redhat.com
·
3d
Page 2 »
Log in to enable infinite scrolling
Keyboard Shortcuts
Navigation
Next / previous item
j
/
k
Open post
o
or
Enter
Preview post
v
Post Actions
Love post
a
Like post
l
Dislike post
d
Undo reaction
u
Save / unsave
s
Recommendations
Add interest / feed
Enter
Not interested
x
Go to
Home
g
h
Interests
g
i
Feeds
g
f
Likes
g
l
History
g
y
Changelog
g
c
Settings
g
s
Browse
g
b
Search
/
Pagination
Next page
n
Previous page
p
General
Show this help
?
Submit feedback
!
Close modal / unfocus
Esc
Press
?
anytime to show this help