trailofbits/buttercup
github.comยท9h
What Makes "Good" Distractors for Object Hallucination Evaluation in Large Vision-Language Models?
arxiv.orgยท7h
Can LLMs Detect Their Confabulations? Estimating Reliability in Uncertainty-Aware Language Models
arxiv.orgยท7h
BigTokDetect: A Clinically-Informed Vision-Language Model Framework for Detecting Pro-Bigorexia Videos on TikTok
arxiv.orgยท7h
Loading...Loading more...