Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

>The human model isn’t just trained in language though, it’s trained on the world.

It's not trained on "the world". It's trained on a small slice of it, a few senses that are themselves slimmed down and then fabricated at parts.

To the bird that can feel and sense electromagnetic waves intuitively to guide travels, you're not trained on "the world"

>This is why they are so prone to hallucination.

No it's not. LLMs can distinguish between truth and hallucination. They just don't care to communicate that.

GPT-4 logits calibration pre RLHF - https://imgur.com/a/3gYel9r

Just Ask for Calibration: Strategies for Eliciting Calibrated Confidence Scores from Language Models Fine-Tuned with Human Feedback - https://arxiv.org/abs/2305.14975

Teaching Models to Express Their Uncertainty in Words - https://arxiv.org/abs/2205.14334

Language Models (Mostly) Know What They Know - https://arxiv.org/abs/2207.05221

The Geometry of Truth: Emergent Linear Structure in Large Language Model Representations of True/False Datasets - https://arxiv.org/abs/2310.06824



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: