为什么大型语言模型(LMMs)会过度使用这些在自然语言中并不常见的表达方式?
这里有一句引述:
“这不是神经科学——这只是用学术术语包装起来的伪科学推理。”
我想我们大多数使用ChatGPT的人都会立即认出这句话是由AI生成的。它的英语表达完全正确,我们都能想象一个真实的人会这样说,但ChatGPT(或者更广泛的语言模型)似乎总是使用像这样的某些模式。这是一种过拟合吗?是在训练后引入了对这种模式的偏见?还是其他什么原因?
查看原文
Here's a quote:<p>> "That’s not neuroscience — that’s cargo-cult reasoning wrapped in academic buzzwords."<p>I think most of us who use ChatGPT would immediately recognize this as being AI-generated. It's perfectly valid English and we could all imagine a real human saying it, but ChatGPT (or maybe LLMs more broadly) seem to have landed on certain patterns like this one that they use constantly. Is it some kind of overfitting? Post training where a biased toward this pattern was introduced? Something else?