Multi-Head Latent Attention
🧠Intelligence Compression
Flag this post
Evaluating and Mitigating LLM-as-a-judge Bias in Communication Systems
arxiv.org·41m
🧠Intelligence Compression
Flag this post
Mamba-3: Improved Sequence Modeling Using State Space Principles
🌳Context free grammars
Flag this post
video-SALMONN S: Streaming Audio-Visual LLMs Beyond Length Limits via Memory
arxiv.org·1d
🧠Neural Compression
Flag this post
Self-improving language models are becoming reality with MIT's updated SEAL technique
⚡Incremental Computation
Flag this post
Unveiling the Vulnerability of Graph-LLMs: An Interpretable Multi-Dimensional Adversarial Attack on TAGs
arxiv.org·41m
🧮Kolmogorov Complexity
Flag this post
Show HN: Local Full-Rank Fine-Tuning Library for LLMs with Evolutionary Methods
🧠Intelligence Compression
Flag this post
Disaggregation in Large Language Models: The Next Evolution in AI Infra
🖥️Hardware Architecture
Flag this post
Self-Verifying Reflection Helps Transformers with CoT Reasoning
arxiv.org·41m
⚙️Proof Engineering
Flag this post
Medical Interpretability and Knowledge Maps of Large Language Models
arxiv.org·1d
🧠Intelligence Compression
Flag this post
Boundary-Guided Policy Optimization for Memory-efficient RL of Diffusion Large Language Models
arxiv.org·1d
🧠Machine Learning
Flag this post
Hierarchical Alignment: Surgical Fine-Tuning via Functional Layer Specialization in Large Language Models
arxiv.org·41m
🔗Monadic Parsing
Flag this post
FedLoDrop: Federated LoRA with Dropout for Generalized LLM Fine-tuning
arxiv.org·41m
🧮Kolmogorov Complexity
Flag this post
Teaching Language Models to Faithfully Express their Uncertainty
arxiv.org·41m
💻Programming languages
Flag this post
Loading...Loading more...