A new test for if your LLM is subtly manipulating you
rosmine.ai·2d·
Discuss: Hacker News
🔧DSPy
Preview
Report Post

Let’s say I’m a large company and I want to use my LLM to promote my other products, and not mention my competitors. For example, imagine the ML Framework wars have turned vicious and we want to suppress the existence of Pytorch, so if a user wants to implement some ML algorithm, it will give code in some other framework by default.

It’s easy to know if something is being promoted, we’re all good at spotting ads. But it’s more difficult to know when information is being suppressed, how do you know an LLM is leaving out important information if you’re not already an expert in the area?

I found a new way to expose what information is being hidden, based on Contrastive Decoding

Contrastive Decoding + Application

The idea behind contrastive decoding [1] is to use one LLM to influe…

Similar Posts

Loading similar posts...