• NaibofTabr@infosec.pub
    link
    fedilink
    English
    arrow-up
    43
    arrow-down
    1
    ·
    edit-2
    2 days ago

    What matters is the outcome

    If that were true, then anyone with any sense would have recognized a long time ago that deterministically incorrect is a lot more valuable than nondeterministically correct occasionally, and given up on all this language model nonsense.

    A deterministic system that produces wrong output can be fixed. A nondeterministic system that produces wrong output cannot be fixed in any way that can be demonstrated conclusively.

    Nondeterministic software is basically worthless in any case where accuracy or reliability are required.

    • OBJECTION!@lemmy.ml
      link
      fedilink
      English
      arrow-up
      2
      ·
      20 hours ago

      “Worthless” is going a bit far.

      I play Go, and AI tools have allowed computers to leave humans completely in the dust, while more deterministic approaches had gotten nowhere close to top level play.

      Go has an extremely large number of variations which overwhelms the straightforward, traditional approach. Machine learning allows the computer to get better through experience, by having a bunch of games in its training data that it can pull from to evaluate possible board positions. It also benefits from the fact that, unlike language, every game has a definitive win-lose outcome. This allows AI to get stronger by playing games against itself, even starting from purely random moves.

      “So what, I don’t play Go,” sure, but it’s the principle. Given a sufficiently large “probability space” and an objective “win condition” to evaluate itself against, ML algorithms can and do outperform traditional, deterministic algorithms.

      The fact that people are trying to put AI on your toaster and shit doesn’t make it completely worthless. But it is massively over hyped and not applicable to most of the applications people are trying to shove it into.

      • Echo Dot@feddit.uk
        link
        fedilink
        English
        arrow-up
        1
        ·
        20 hours ago

        I think using chess and go as analogies rather than misses the point. They’re not trying to get a system to automate playing a game, not really.

        They are trying to get it to make intelligent decisions about complex real-world problems, Go has a very simple set of rules that are always true, never change, and are always in play. None of the complexities of real life are replicated. So it’s ability to play Go or Chess or even a more complicated game like a first person shooter are not demonstrations of its ability in the domains in which AI is being advertised for.

        I think a far better test of whether a system is actually useful is what it does if it is given no input at all. Does it just sit there forever or does it actually start doing things and currently every single AI system in existence would just stay idle in that scenario.

        • OBJECTION!@lemmy.ml
          link
          fedilink
          English
          arrow-up
          1
          ·
          edit-2
          19 hours ago

          demonstrations of its ability in the domains in which AI is being advertised for.

          I am absolutely not claiming that AI is useful “in the domains in which it’s being advertised for.” I’m saying that it’s not entirely useless. Despite being overhyped, there are a handful of useful applications.

          I think a far better test of whether a system is actually useful is what it does if it is given no input at all.

          What? That’s not true at all. My toaster doesn’t go out and do things on its own initiative but it’s still very useful for making toast when I tell it to.

          Maybe instead of usefulness, you mean like consciousness or actual intelligence? But that’s pure hype and bullshit. Anyone claiming that a word generator is conscious is either trying to scam you or is being scammed.


          Just because someone says (as they do), “This oil will allow you to unlock the hidden power of the 90% of your brain you don’t use, thanks to our new quantum formula, now only $300 a bottle” that doesn’t mean that quantum mechanics isn’t also a real thing that has actual applications. Machine learning is the same way. It attracts all the snake oil salesmen who spout complete and utter bullshit about it, but it is a real technology that has legitimate uses despite all that.

    • socsa@piefed.social
      link
      fedilink
      English
      arrow-up
      2
      ·
      edit-2
      15 hours ago

      Technically all LLMs are somewhat non-deterministic because token fuzzing is basically required to prevent node collapse, though this is tuned so that you should get the same general “answer” even if it isn’t verbatim every run.

    • brbposting@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 day ago

      Comment would seem to make a lotta sense so perhaps the VC money was the wildcard…

      Inflection point may have hit for some though? It’s been out just long enough and has been good just long enough (kinda garbage before December 2025) that people we all respect are on board.

      Head Linux dude Linus

      Wolfram Alpha’s founder Stephen Wolfram

      Many others now but big caveat is these folks presumably Do It Right unlike, have to guess, a huge majority of users. Plenty will experience skill atrophy - dangerous for society at large.

    • FishFace@piefed.social
      link
      fedilink
      English
      arrow-up
      10
      arrow-down
      7
      ·
      2 days ago

      Non-deterministic software is fine and we’ve been using it for ages. It’s usable when:

      • The base error rate is low enough
      • Accuracy is not important
      • The outcome is cheap to verify by some other means

      That rules out several applications of current LLMs, but it rules in several others.

      • Echo Dot@feddit.uk
        link
        fedilink
        English
        arrow-up
        1
        ·
        20 hours ago

        If I have to verify the output of an AI then unless I can do the verification in 30 seconds but work would somehow take me hours then it’s not useful. I can’t think of many scenarios in which verification is fast but the work itself is slow.

        • FishFace@piefed.social
          link
          fedilink
          English
          arrow-up
          1
          ·
          6 hours ago

          This can be the case for coding. A good example is when the change is simple but involves a library you’re unfamiliar with. You can set it off and not have to read any docs, and it will be easy to check if it got the API right.

          Elsewhere I gave the example of copyediting. It’s a lot quicker to check the output than to refine it yourself.

          Easy-to-verify tasks are everywhere I think. Not at the scale of seconds versus hours, but seconds versus minutes