Whenever someone says how LLMs have improved, all I can think is that the changes during all this time have felt very minor to me.
-
wrote 25 days ago last edited by
Whenever someone says how LLMs have improved, all I can think is that the changes during all this time have felt very minor to me. The one thing I would consider a real breakthrough is a new architecture that doesn't make hallucinating the core mechanism for how it works. Otherwise I just don't see them ever become reliable.
-
Whenever someone says how LLMs have improved, all I can think is that the changes during all this time have felt very minor to me. The one thing I would consider a real breakthrough is a new architecture that doesn't make hallucinating the core mechanism for how it works. Otherwise I just don't see them ever become reliable.
wrote 25 days ago last edited by@volpeon
According to most researchers, hallucinations will never go away.
-
Whenever someone says how LLMs have improved, all I can think is that the changes during all this time have felt very minor to me. The one thing I would consider a real breakthrough is a new architecture that doesn't make hallucinating the core mechanism for how it works. Otherwise I just don't see them ever become reliable.
wrote 25 days ago last edited by@volpeon some of the MOE models that don’t need to run all the parameters in the model seem interesting in that they will run on more modest hardware.
I had one of the quantised quen 3 30b models running on a 24core first gen epyc and getting okish performance.
Getting good (relatively anyway) models that don’t need globs of hardware would be good progression although I suspect it will still burn a shitton of resources training them.
-
Whenever someone says how LLMs have improved, all I can think is that the changes during all this time have felt very minor to me. The one thing I would consider a real breakthrough is a new architecture that doesn't make hallucinating the core mechanism for how it works. Otherwise I just don't see them ever become reliable.
wrote 25 days ago last edited by@volpeon "Look, NetBeans doesn't needs 5s for autocompletion, now it only takes 2s!" -
Whenever someone says how LLMs have improved, all I can think is that the changes during all this time have felt very minor to me. The one thing I would consider a real breakthrough is a new architecture that doesn't make hallucinating the core mechanism for how it works. Otherwise I just don't see them ever become reliable.
wrote 25 days ago last edited by@volpeon not hallucinating is fundamentally impossible for what an LLM is, which is a probabilistic model that takes in a string of tokens and predicts what should come next. we're a long long way away from something which actually reasons in a meaningful way -
@volpeon some of the MOE models that don’t need to run all the parameters in the model seem interesting in that they will run on more modest hardware.
I had one of the quantised quen 3 30b models running on a 24core first gen epyc and getting okish performance.
Getting good (relatively anyway) models that don’t need globs of hardware would be good progression although I suspect it will still burn a shitton of resources training them.
wrote 25 days ago last edited by@Dragon The quality of models you can run on personal hardware certainly has improved a lot, but even that feels minor to me. I just don't find myself using LLMs beyond experimenting because I always double check what it says, because I know how they work. As long as these doubts exist, they won't be all that useful to me
-
@volpeon not hallucinating is fundamentally impossible for what an LLM is, which is a probabilistic model that takes in a string of tokens and predicts what should come next. we're a long long way away from something which actually reasons in a meaningful waywrote 25 days ago last edited by
@chjara Exactly!