• Lvxferre@lemmy.ml
    link
    fedilink
    arrow-up
    8
    arrow-down
    1
    ·
    edit-2
    1 year ago

    Given that it’s pointing straight to “no”, should I interpret “AI” as “additional irony”?

    …seriously, model-based generation is in its infancy. Currently it outputs mostly trash; you need to spend quite a bit of time to sort something useful out of it. If anyone here actually believes that it’s smart, I have a bridge to sell you.

    • huginn@feddit.it
      link
      fedilink
      arrow-up
      2
      arrow-down
      1
      ·
      1 year ago

      LLMs will undoubtedly improve as we build more systems around them.

      The question is will it ever be reliable enough to trust? You can’t have a 99% reliable critical system.

      • novibe@lemmy.ml
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        2
        ·
        edit-2
        1 year ago

        But 99% accuracy is better than any human alive, so while maybe LLMs won’t be able to substitute critical systems, they might just replace all the people around those systems.

        Like, we won’t want an AI as the failsafe for a nuclear plant. But we might prefer an AI as the the “person” in charge of this failsafe.

        • huginn@feddit.it
          link
          fedilink
          arrow-up
          1
          ·
          edit-2
          1 year ago

          Current generations aren’t even close to that rate, and it’s unclear if it’s economical or even possible to fix the deep structural issues of our current Gen LLMs.

          My professional experience with LLMs is that they don’t even approach 20% accuracy for a field as ridiculously structured as programming.

          They’re just helpful enough to not be a hindrance.

          Not too mention plenty of humans are 99% accurate