Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Collapse

NodeBB

  1. Home
  2. uncategorized
  3. Whenever someone says how LLMs have improved, all I can think is that the changes during all this time have felt very minor to me.

Whenever someone says how LLMs have improved, all I can think is that the changes during all this time have felt very minor to me.

Scheduled Pinned Locked Moved uncategorized
7 Posts 5 Posters 0 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • volpeon@icy.wyvern.ripV This user is from outside of this forum
    volpeon@icy.wyvern.ripV This user is from outside of this forum
    volpeon@icy.wyvern.rip
    wrote last edited by
    #1

    Whenever someone says how LLMs have improved, all I can think is that the changes during all this time have felt very minor to me. The one thing I would consider a real breakthrough is a new architecture that doesn't make hallucinating the core mechanism for how it works. Otherwise I just don't see them ever become reliable.

    catraxx@tech.lgbtC ? lanodan@queer.hacktivis.meL chjara@akko.wtfC 4 Replies Last reply
    0
    • volpeon@icy.wyvern.ripV volpeon@icy.wyvern.rip

      Whenever someone says how LLMs have improved, all I can think is that the changes during all this time have felt very minor to me. The one thing I would consider a real breakthrough is a new architecture that doesn't make hallucinating the core mechanism for how it works. Otherwise I just don't see them ever become reliable.

      catraxx@tech.lgbtC This user is from outside of this forum
      catraxx@tech.lgbtC This user is from outside of this forum
      catraxx@tech.lgbt
      wrote last edited by
      #2

      @volpeon According to most researchers, hallucinations will never go away.

      1 Reply Last reply
      0
      • volpeon@icy.wyvern.ripV volpeon@icy.wyvern.rip

        Whenever someone says how LLMs have improved, all I can think is that the changes during all this time have felt very minor to me. The one thing I would consider a real breakthrough is a new architecture that doesn't make hallucinating the core mechanism for how it works. Otherwise I just don't see them ever become reliable.

        ? Offline
        ? Offline
        Guest
        wrote last edited by
        #3

        @volpeon some of the MOE models that don’t need to run all the parameters in the model seem interesting in that they will run on more modest hardware.

        I had one of the quantised quen 3 30b models running on a 24core first gen epyc and getting okish performance.

        Getting good (relatively anyway) models that don’t need globs of hardware would be good progression although I suspect it will still burn a shitton of resources training them.

        volpeon@icy.wyvern.ripV 1 Reply Last reply
        0
        • volpeon@icy.wyvern.ripV volpeon@icy.wyvern.rip

          Whenever someone says how LLMs have improved, all I can think is that the changes during all this time have felt very minor to me. The one thing I would consider a real breakthrough is a new architecture that doesn't make hallucinating the core mechanism for how it works. Otherwise I just don't see them ever become reliable.

          lanodan@queer.hacktivis.meL This user is from outside of this forum
          lanodan@queer.hacktivis.meL This user is from outside of this forum
          lanodan@queer.hacktivis.me
          wrote last edited by
          #4
          @volpeon "Look, NetBeans doesn't needs 5s for autocompletion, now it only takes 2s!"
          1 Reply Last reply
          0
          • volpeon@icy.wyvern.ripV volpeon@icy.wyvern.rip

            Whenever someone says how LLMs have improved, all I can think is that the changes during all this time have felt very minor to me. The one thing I would consider a real breakthrough is a new architecture that doesn't make hallucinating the core mechanism for how it works. Otherwise I just don't see them ever become reliable.

            chjara@akko.wtfC This user is from outside of this forum
            chjara@akko.wtfC This user is from outside of this forum
            chjara@akko.wtf
            wrote last edited by
            #5
            @volpeon not hallucinating is fundamentally impossible for what an LLM is, which is a probabilistic model that takes in a string of tokens and predicts what should come next. we're a long long way away from something which actually reasons in a meaningful way
            volpeon@icy.wyvern.ripV 1 Reply Last reply
            0
            • ? Guest

              @volpeon some of the MOE models that don’t need to run all the parameters in the model seem interesting in that they will run on more modest hardware.

              I had one of the quantised quen 3 30b models running on a 24core first gen epyc and getting okish performance.

              Getting good (relatively anyway) models that don’t need globs of hardware would be good progression although I suspect it will still burn a shitton of resources training them.

              volpeon@icy.wyvern.ripV This user is from outside of this forum
              volpeon@icy.wyvern.ripV This user is from outside of this forum
              volpeon@icy.wyvern.rip
              wrote last edited by
              #6

              @Dragon The quality of models you can run on personal hardware certainly has improved a lot, but even that feels minor to me. I just don't find myself using LLMs beyond experimenting because I always double check what it says, because I know how they work. As long as these doubts exist, they won't be all that useful to me

              1 Reply Last reply
              0
              • chjara@akko.wtfC chjara@akko.wtf
                @volpeon not hallucinating is fundamentally impossible for what an LLM is, which is a probabilistic model that takes in a string of tokens and predicts what should come next. we're a long long way away from something which actually reasons in a meaningful way
                volpeon@icy.wyvern.ripV This user is from outside of this forum
                volpeon@icy.wyvern.ripV This user is from outside of this forum
                volpeon@icy.wyvern.rip
                wrote last edited by
                #7

                @chjara Exactly!

                1 Reply Last reply
                0
                Reply
                • Reply as topic
                Log in to reply
                • Oldest to Newest
                • Newest to Oldest
                • Most Votes


                • Login

                • Login or register to search.
                Powered by NodeBB Contributors
                • First post
                  Last post
                0
                • Categories
                • Recent
                • Tags
                • Popular
                • World
                • Users
                • Groups