Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

We understand exactly how it works. It just works in such a way that we cannot predict the outcome, which makes it pretty bad for many applications. If you can't explain how it works doesn't mean it's not understood how it works.


We really don't know why LLMs have the most of their emergent capabilities. The entire GPT-3 paper was about being surprised about it.


Maybe we don't disagree then.

We know how LLM work. Parameters. Training data. Random number generator. That stuff.

We don't know why it outputs what it outputs because rngs are notoriously unpredictable and we know it. So we are surprised but that in itself is unsurprising.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: