Discussion about this post

User's avatar
Jurgen Gravestein's avatar

To me the whole paper feels like a giant cope. "Look, it's not the technology that's broken, it's the evaluations."

Expand full comment
Mike X Cohen's avatar

Thanks for the clear write-up, Alberto. Hallucinations are tricky, because the stochasticity of an LLM's output should itself be context-dependent: If you ask a model to generate an imaginative children's story, you want it to be more "creative" by making shit up; whereas if you ask the model for historical facts, you don't want anything resembling hallucinations.

Expand full comment
6 more comments...

No posts