My guess is that it’s set up to see contexts with conflicting positions associated as controversial but it will just go with responses that don’t have controversy associated with them.
This is significantly more reasoning and analysis than LLMs are capable of.
When they give those “I can’t respond to that” replies, it’s because a specific programmed keyword filter was tripped, forcing the model to insert a pre-programmed response instead. The rest of the time, they’re just regurgitating a melange of the most statically present text on the subject from their training data.
This is significantly more reasoning and analysis than LLMs are capable of.
When they give those “I can’t respond to that” replies, it’s because a specific programmed keyword filter was tripped, forcing the model to insert a pre-programmed response instead. The rest of the time, they’re just regurgitating a melange of the most statically present text on the subject from their training data.