Neural Codecs, Deep Learning, Entropy Models, Rate-Distortion
SlimMoE: Structured Compression of Large MoE Models via Expert Slimming and Distillation
arxiv.org·1d
Why Your Next LLM Might Not Have A Tokenizer
towardsdatascience.com·22h
Loading...Loading more...