Skip to main content
Scour
Browse
Getting Started
Login
Sign Up
You are offline. Trying to reconnect...
Close
You're currently offline. Some features may not work.
Close
Copied to clipboard
Close
Unable to share or copy to clipboard
Close
📉 Model Quantization
INT8, Post-Training, QAT, Pruning, Model Compression
Filter Results
Timeframe
Fresh
Past Hour
Today
This Week
This Month
Feeds to Scour
Subscribed
All
Scoured
82786
posts in
408.5
ms
D$^2$
Quant
:
Accurate
Low-bit Post-Training Weight Quantization for LLMs
arxiv.org
·
11h
🎯
Tensor Cores
Quantization-Aware
Regularizers
for Deep Neural Networks Compression
arxiv.org
·
11h
📊
Gradient Accumulation
7 Advanced Feature Engineering
Tricks
Using LLM
Embeddings
machinelearningmastery.com
·
1d
🎓
Model Distillation
Training Design for Text-to-Image Models:
Lessons
from
Ablations
huggingface.co
·
1d
📊
Gradient Accumulation
What is
Overfitting
? -
Overfitting
in Machine Learning
Explained
aws.amazon.com
·
12h
·
Discuss:
Hacker News
📊
Gradient Accumulation
qcc4cp/qcc
: Source code for the book "Quantum Computing for
Programmers
", Cambridge University Press
github.com
·
8h
·
Discuss:
Hacker News
🔄
ONNX
The 4-Step
Magic
: How Neural Networks Actually Learn (With Real
Examples
)
pub.towardsai.net
·
3h
📊
Gradient Accumulation
## Deep Learning-Driven Predictive Modeling of Shell Degradation in Juvenile *
Crassostrea
virginica
* under Simulated Ocean Acidification Conditions: A Multi-Modal Approach
freederia.com
·
3h
🔄
ONNX
Convert
&
Compress
frontendmasters.com
·
1d
🐕
Ruff
Automatic
RGBA
Decomposition
image-layered.app
·
2d
·
Discuss:
Hacker News
🧮
cuDNN
Training a Small Language Model
elijahpotter.dev
·
1d
🏎️
TensorRT
Future
leakage
in
block-quantized
attention
matx.com
·
1d
·
Discuss:
Hacker News
⚡
Flash Attention
I built a free
ML
practice
platform - would love your feedback [P]
reddit.com
·
7h
·
Discuss:
r/MachineLearning
🚀
MLOps
Optimized
LLM Inference
Engines
rishirajacharya.com
·
1h
⚡
ONNX Runtime
Information
Retrieval
Part 2: How To Get Into Model Training Data
searchenginejournal.com
·
2h
📊
Gradient Accumulation
Physics 1 – Attention can’t exactly
simulate
uniform
linear motion
kindxiaoming.github.io
·
1d
👁️
Attention Optimization
A spatial domain variable block size
luma
dependent
chroma
compression algorithm
bitsnbites.eu
·
5h
·
Discuss:
Hacker News
⚡
Flash Attention
Writing an LLM from scratch, part
32a
-- Interventions: training a
baseline
model
gilesthomas.com
·
14h
·
Discuss:
Hacker News
📊
Gradient Accumulation
A Black Box Made Less
Opaque
(part 2)
lesswrong.com
·
11h
🛠
Ml-eng
Zephyr
: Direct Distillation of
LM
Alignment
dev.to
·
5h
·
Discuss:
DEV
🛠
Ml-eng
Loading...
Loading more...
Page 2 »
Keyboard Shortcuts
Navigation
Next / previous item
j
/
k
Open post
o
or
Enter
Preview post
v
Post Actions
Love post
a
Like post
l
Dislike post
d
Undo reaction
u
Recommendations
Add interest / feed
Enter
Not interested
x
Go to
Home
g
h
Interests
g
i
Feeds
g
f
Likes
g
l
History
g
y
Changelog
g
c
Settings
g
s
Browse
g
b
Search
/
Pagination
Next page
n
Previous page
p
General
Show this help
?
Submit feedback
!
Close modal / unfocus
Esc
Press
?
anytime to show this help