trailofbits/buttercup
github.com·13h
Can LLMs Detect Their Confabulations? Estimating Reliability in Uncertainty-Aware Language Models
arxiv.org·11h
What Makes "Good" Distractors for Object Hallucination Evaluation in Large Vision-Language Models?
arxiv.org·11h
BigTokDetect: A Clinically-Informed Vision-Language Model Framework for Detecting Pro-Bigorexia Videos on TikTok
arxiv.org·11h
Loading...Loading more...