Beyond Manually Designed Pruning Policies with Second-Level Performance Prediction: A Pruning Framework for LLMs
arxiv.org·18h
TRACEALIGN -- Tracing the Drift: Attributing Alignment Failures to Training-Time Belief Sources in LLMs
arxiv.org·18h
UrBLiMP: A Benchmark for Evaluating the Linguistic Competence of Large Language Models in Urdu
arxiv.org·18h
Hearing More with Less: Multi-Modal Retrieval-and-Selection Augmented Conversational LLM-Based ASR
arxiv.org·18h
LinkQA: Synthesizing Diverse QA from Multiple Seeds Strongly Linked by Knowledge Points
arxiv.org·18h
Understanding and Mitigating Political Stance Cross-topic Generalization in Large Language Models
arxiv.org·18h
Loading...Loading more...