Hacker News new | past | comments | ask | show | jobs | submit login

I interpreted that statement as saying the current practice is to make LLMs larger and larger (so they effectively memorize more and more data) to make them more powerful, but from the perspective of information theory, if models were powerful and "understanding", then models could stay the same size and become more and more powerful as they get increasingly better at compressing the available information. I am not sure if this interpretation was what was meant though.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: