Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Collapse

NodeBB

  1. Home
  2. uncategorized
  3. Whenever someone says how LLMs have improved, all I can think is that the changes during all this time have felt very minor to me.

Whenever someone says how LLMs have improved, all I can think is that the changes during all this time have felt very minor to me.

Scheduled Pinned Locked Moved uncategorized
7 Posts 5 Posters 8 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • V This user is from outside of this forum
    V This user is from outside of this forum
    volpeon@icy.wyvern.rip
    wrote 25 days ago last edited by
    #1

    Whenever someone says how LLMs have improved, all I can think is that the changes during all this time have felt very minor to me. The one thing I would consider a real breakthrough is a new architecture that doesn't make hallucinating the core mechanism for how it works. Otherwise I just don't see them ever become reliable.

    C ? L C 4 Replies Last reply 25 days ago
    0
    • V volpeon@icy.wyvern.rip
      25 days ago

      Whenever someone says how LLMs have improved, all I can think is that the changes during all this time have felt very minor to me. The one thing I would consider a real breakthrough is a new architecture that doesn't make hallucinating the core mechanism for how it works. Otherwise I just don't see them ever become reliable.

      C This user is from outside of this forum
      C This user is from outside of this forum
      catraxx@tech.lgbt
      wrote 25 days ago last edited by
      #2

      @volpeon According to most researchers, hallucinations will never go away.

      1 Reply Last reply
      0
      • V volpeon@icy.wyvern.rip
        25 days ago

        Whenever someone says how LLMs have improved, all I can think is that the changes during all this time have felt very minor to me. The one thing I would consider a real breakthrough is a new architecture that doesn't make hallucinating the core mechanism for how it works. Otherwise I just don't see them ever become reliable.

        ? Offline
        ? Offline
        Guest
        wrote 25 days ago last edited by
        #3

        @volpeon some of the MOE models that don’t need to run all the parameters in the model seem interesting in that they will run on more modest hardware.

        I had one of the quantised quen 3 30b models running on a 24core first gen epyc and getting okish performance.

        Getting good (relatively anyway) models that don’t need globs of hardware would be good progression although I suspect it will still burn a shitton of resources training them.

        V 1 Reply Last reply 25 days ago
        0
        • V volpeon@icy.wyvern.rip
          25 days ago

          Whenever someone says how LLMs have improved, all I can think is that the changes during all this time have felt very minor to me. The one thing I would consider a real breakthrough is a new architecture that doesn't make hallucinating the core mechanism for how it works. Otherwise I just don't see them ever become reliable.

          L This user is from outside of this forum
          L This user is from outside of this forum
          lanodan@queer.hacktivis.me
          wrote 25 days ago last edited by
          #4
          @volpeon "Look, NetBeans doesn't needs 5s for autocompletion, now it only takes 2s!"
          1 Reply Last reply
          0
          • V volpeon@icy.wyvern.rip
            25 days ago

            Whenever someone says how LLMs have improved, all I can think is that the changes during all this time have felt very minor to me. The one thing I would consider a real breakthrough is a new architecture that doesn't make hallucinating the core mechanism for how it works. Otherwise I just don't see them ever become reliable.

            C This user is from outside of this forum
            C This user is from outside of this forum
            chjara@akko.wtf
            wrote 25 days ago last edited by
            #5
            @volpeon not hallucinating is fundamentally impossible for what an LLM is, which is a probabilistic model that takes in a string of tokens and predicts what should come next. we're a long long way away from something which actually reasons in a meaningful way
            V 1 Reply Last reply 25 days ago
            0
            • ? Guest
              25 days ago

              @volpeon some of the MOE models that don’t need to run all the parameters in the model seem interesting in that they will run on more modest hardware.

              I had one of the quantised quen 3 30b models running on a 24core first gen epyc and getting okish performance.

              Getting good (relatively anyway) models that don’t need globs of hardware would be good progression although I suspect it will still burn a shitton of resources training them.

              V This user is from outside of this forum
              V This user is from outside of this forum
              volpeon@icy.wyvern.rip
              wrote 25 days ago last edited by
              #6

              @Dragon The quality of models you can run on personal hardware certainly has improved a lot, but even that feels minor to me. I just don't find myself using LLMs beyond experimenting because I always double check what it says, because I know how they work. As long as these doubts exist, they won't be all that useful to me

              1 Reply Last reply
              0
              • C chjara@akko.wtf
                25 days ago
                @volpeon not hallucinating is fundamentally impossible for what an LLM is, which is a probabilistic model that takes in a string of tokens and predicts what should come next. we're a long long way away from something which actually reasons in a meaningful way
                V This user is from outside of this forum
                V This user is from outside of this forum
                volpeon@icy.wyvern.rip
                wrote 25 days ago last edited by
                #7

                @chjara Exactly!

                1 Reply Last reply
                0
                Reply
                • Reply as topic
                Log in to reply
                • Oldest to Newest
                • Newest to Oldest
                • Most Votes

                4/7

                18 May 2025, 19:56


                • Login

                • Login or register to search.
                Powered by NodeBB Contributors
                4 out of 7
                • First post
                  4/7
                  Last post
                0
                • Categories
                • Recent
                • Tags
                • Popular
                • World
                • Users
                • Groups