So, these language models, when they are being trained, do they need someone telling them what they got wrong and what they got right? How do they know?
From what I've seen, they call it "alignment".
Please Login to reply.
No replies yet.