Kimi Linear: An Expressive, Efficient Attention Architecture
arxiviq.substack.com·1d·
Discuss: Substack
👁️Attention Optimization
Flag this post
Attention Illuminates LLM Reasoning: The Preplan-and-Anchor Rhythm EnablesFine-Grained Policy Optimization
paperium.net·12h·
Discuss: DEV
👁️Attention Optimization
Flag this post
Everything About Transformers
krupadave.com·3d
👁️Attention Optimization
Flag this post
ViSurf: Visual Supervised-and-Reinforcement Fine-Tuning for LargeVision-and-Language Models
dev.to·21h·
Discuss: DEV
👁️Attention Optimization
Flag this post
Can-t stop till you get enough
cant.bearblog.dev·4h·
Discuss: Hacker News
📜TorchScript
Flag this post
A Minimal Route to Transformer Attention
neelsomaniblog.com·3d·
Discuss: Hacker News
👁️Attention Optimization
Flag this post
[D] Best (free) courses on neural networks
reddit.com·1d·
👁️Attention Optimization
Flag this post
MobileNetV3 Paper Walkthrough: The Tiny Giant Getting Even Smarter
towardsdatascience.com·10h
🎯Tensor Cores
Flag this post
Vision = Language: I Decoded VLM Tokens to See What AI 'Sees' 🔬
reddit.com·4h·
Discuss: r/LocalLLaMA
🛠Ml-eng
Flag this post
Dual-format attentional template during preparation in human visual cortex
elifesciences.org·4d
Flash Attention
Flag this post
The middle brother in classifier development: What is RandAugment?
openaccess.thecvf.com·11h·
Discuss: DEV
📊Gradient Accumulation
Flag this post
Semantic search with embeddings in PHP: a hands-on guide using Neuron AI and Ollama
ollama.com·1d·
Discuss: DEV
🛠Ml-eng
Flag this post
An underqualified reading list about the transformer architecture
fvictorio.github.io·3d·
Discuss: Hacker News
Flash Attention
Flag this post
Scalable In-Memory Associative Processing for Graph Neural Network Inference
dev.to·11h·
Discuss: DEV
Flash Attention
Flag this post
ViSurf: Visual Supervised-and-Reinforcement Fine-Tuning for LargeVision-and-Language Models
paperium.net·21h·
Discuss: DEV
🏎️TensorRT
Flag this post
University of Surrey researchers mimic brain wiring to improve AI - BBC
news.google.com·10h
Flash Attention
Flag this post
Toward a Compressed Core of Human Knowledge: The High-Dimensional Vector Network for AI
reddit.com·5h·
Discuss: r/AfterClass
📊Gradient Accumulation
Flag this post
Platform generated AI slop at scale
markjgsmith.com·1h
🤖AI Coding Tools
Flag this post
Specialized structure of neural population codes in parietal cortex outputs
nature.com·2d
Flash Attention
Flag this post