"The largest models were generally the least truthful. This contrasts with other NLP tasks, where performance improves with
model size." (Lyn et al., 2022).
So does that mean LLMs normally learn truth better but humans "fix" them by feeding lies.
"The largest models were generally the least truthful. This contrasts with other NLP tasks, where performance improves with
model size." (Lyn et al., 2022).
So does that mean LLMs normally learn truth better but humans "fix" them by feeding lies.
No replies yet.