I suspect, but can’t prove, that model trainers deliberately steer models away from creating tests and worksheets.
The reason being that when a human asks a written question: “who president of the US in 1962?” it’s very likely to be a part of a worksheet.
Novels don’t contain many questions like that, nor do non fiction. They’re mostly paragraphs. Most text is mostly paragraphs.
Naked question usually means worksheet. AIs know this, so their if you ask it a question like: “who president in 1962?” It responds which the most likely next sentence, a related question: “How was the Cuban Missile resolved?”
So there’s a huge discrepancy between the next most likely sentence based on training, and what a user likely wants. If I ask, “Who was president in 1962?” I don’t want another question, nor does anyone else.
But that’s what the training data provides.
So model trainers have to bias it away from worksheets. This isn’t hard to do, and is a normal part of model training.
I’ve personally seen this behavior in poorly parameterized or trained models. They love answering questions by assuming they are in worksheets. It’s a huge pain.
Interestingly it never happens with top-line models like ChatGPT.
Carefully hyperparameterizatiin helps, but I think you’ll have to adjust the weights too. But that likely makes it harder to make actual worksheets.
This is just a guess. But I suspect models are weighted to discount pedagogical materials because of how different they are from what the users often expect.
I suspect, but can’t prove, that model trainers deliberately steer models away from creating tests and worksheets.
The reason being that when a human asks a written question: “who president of the US in 1962?” it’s very likely to be a part of a worksheet.
Novels don’t contain many questions like that, nor do non fiction. They’re mostly paragraphs. Most text is mostly paragraphs.
Naked question usually means worksheet. AIs know this, so their if you ask it a question like: “who president in 1962?” It responds which the most likely next sentence, a related question: “How was the Cuban Missile resolved?”
So there’s a huge discrepancy between the next most likely sentence based on training, and what a user likely wants. If I ask, “Who was president in 1962?” I don’t want another question, nor does anyone else.
But that’s what the training data provides.
So model trainers have to bias it away from worksheets. This isn’t hard to do, and is a normal part of model training.
I’ve personally seen this behavior in poorly parameterized or trained models. They love answering questions by assuming they are in worksheets. It’s a huge pain.
Interestingly it never happens with top-line models like ChatGPT.
Carefully hyperparameterizatiin helps, but I think you’ll have to adjust the weights too. But that likely makes it harder to make actual worksheets.
This is just a guess. But I suspect models are weighted to discount pedagogical materials because of how different they are from what the users often expect.