• Rioting Pacifist@lemmy.world
    link
    fedilink
    English
    arrow-up
    245
    arrow-down
    1
    ·
    1 day ago

    AIbros: we’re creating God!!!

    AI users: it can do translation & reformating pretty well but you got to check it’s not chatting shit

    • halcyoncmdr@piefed.social
      link
      fedilink
      English
      arrow-up
      90
      ·
      1 day ago

      The takeaway from all LLM-based AI is the user needs to be smart enough to do whatever they’re asking anyway. All output needs to be verified before being used or relied upon.

      The “AI” is just streamlining the process to save time.

      Relying on it otherwise is stupid and just proves instantly that you are incompetent.

      • rumba@lemmy.zip
        link
        fedilink
        English
        arrow-up
        3
        ·
        9 hours ago

        This is absolutely the case, and honestly, at least for now how it needs to be across the board.

        Noone should be using AI to do things you’re incapable of doing (or undoing).

      • 7101334@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        9 hours ago

        Relying on it otherwise is stupid and just proves instantly that you are incompetent.

        Relying on it in any circumstances (though medical stuff is understandable if you’re simply too poor or don’t have access) while it is exhausting water supplies and polluting the planet is stupid and instantly proves that you are stupid and inconsiderate.

      • Zagorath@quokk.au
        link
        fedilink
        English
        arrow-up
        11
        arrow-down
        1
        ·
        1 day ago

        the user needs to be smart enough to do whatever they’re asking anyway

        I’m gonna say that’s ideal but not quite necessary. What’s needed is that the user is capable of properly verifying the output. Which anyone who could do it themselves definitely can, but it can be done more broadly. It’s an easier skill to verify a result than it is to obtain that result. Think: how film critics don’t necessarily need to be filmmakers, or the P=NP question in computer science.

        • Aralakh@lemmy.ca
          link
          fedilink
          English
          arrow-up
          2
          ·
          5 hours ago

          This is where domain expertise would come in, no? It’s speeding up the work but it usually outputs generic content, and whatever else it injects while hallucinating. Therefore the validation part holds up I’d say.

        • Pyro@programming.dev
          link
          fedilink
          English
          arrow-up
          11
          ·
          1 day ago

          But if the output has issues, what’re you going to do, prompt it again? If you are only able to verify but not do the task, you cannot correct the AI’s mistakes yourself.

          • fartographer@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            4 hours ago

            If you’re unable to brute-force verification (research, testing, consulting the ancient texts), there’s where you stop what you’re doing, and take a breath. Then, consult an expert. Just like the film critic analogy, it’s easier to verify than to create, so you’re saving the expert time and effort while learning about something that you were obviously already passionate enough about to have started this endeavor.

          • Zagorath@quokk.au
            link
            fedilink
            English
            arrow-up
            7
            ·
            1 day ago

            At the risk of sounding like an overly obsequious AI… You know what, you’re completely right. I’m honestly not sure what use case I was imagining when I wrote that last comment.

            • EldritchFemininity@lemmy.blahaj.zone
              link
              fedilink
              English
              arrow-up
              1
              ·
              4 hours ago

              You were thinking logically about a normal production chain. In that case, QA or whoever says “This is wrong, rework it and correct the issue” and that’s that. With AI, it does the whole thing over again and may or may not come back with the same issue or an entirely new one.

            • Redjard@reddthat.com
              link
              fedilink
              English
              arrow-up
              5
              ·
              1 day ago

              Making text flow naturally, grouping and ordeeing information, good writing.

              You can verify two textst have the same facts and information, yet one reads way better than the other. But writing a text that reads well is quite hard.

          • Redjard@reddthat.com
            link
            fedilink
            English
            arrow-up
            2
            arrow-down
            1
            ·
            1 day ago

            If you don’t habe the ability then you would do what you would have 5 years ago: not do it
            Either submit without, or not submit at all.

    • youcantreadthis@quokk.auBanned from community
      link
      fedilink
      English
      arrow-up
      10
      ·
      1 day ago

      Fucking hate those anti human filth pushing slop into everything. I want to take one apart with power tools.

    • XLE@piefed.social
      link
      fedilink
      English
      arrow-up
      5
      ·
      1 day ago

      I don’t think AI users would say it does reformatting either (if they’re honest): If you tell a chatbot to reformat text without changing it, it will change the text, because it does not understand the concept of not changing text. It should only take one time for someone to get burned for them to learn that lesson.