Learned Compression, Deep Learning, Rate-Distortion, Entropy Models
SlimMoE: Structured Compression of Large MoE Models via Expert Slimming and Distillation
arxiv.org·1d
Why Your Next LLM Might Not Have A Tokenizer
towardsdatascience.com·1d
Brain Mapping with Dense Features: Grounding Cortical Semantic Selectivity in Natural Images With Vision Transformers
arxiv.org·19h
Loading...Loading more...