• Jakeroxs@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    1
    ·
    3 days ago

    Took that a slightly different way then I was expecting, my point is we have to be on the lookout for bullshit when getting info from other people so it’s really no different when getting info from an LLM.

    However you took it to the LLM can’t determine between what’s true and false, which is obviously true but an interesting point to make nonetheless

    • Nalivai@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      3 days ago

      It’s not that LLM can’t know truth, that’s obvious but besides the point. Its that the user can’t really determine when the lies are, not to the degree that you can be when getting info from a human.
      So you really need to check everything, every claim, every word, every sound. You can’t assume good intentions, there are no intentions in real sence of the word, you can’t extrapolate or intrapolate. Every word of the data you’re getting might be a lie with the same certainty as any other word.
      It requires so much effort to check properly, you either skip some or spend more time that you would without the layer of lies.

      • Jakeroxs@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        2
        ·
        edit-2
        3 days ago

        I don’t see how that’s different honestly, then again I’m not usually asking for absolute truth from LLMs, moreso explaining concepts that I can’t fully grasp by restating things in another way or small coding stuff that I can check essentially immediately if it works or not lol.

        • Nalivai@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          2 days ago

          See, this is the problem I’m talking about. You think you can gauge if the code works or not, but even for small pieces (and in some cases, especially for small pieces) there is a world of very bad, very dangerous shit that lies between “works” and “not works”.
          And it is as dangerous when you trust it to explain something for you. It’s by definition something you don’t know therefore can’t check.

          • Jakeroxs@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            1
            ·
            edit-2
            2 days ago

            I mean I literally can test it immediately lol, a nodered js function isn’t going to be dangerous lol

            Or an AHK script that displays a keystroke on screen, or cleaning up a docker command into docker compose, simple shit lol

            • Nalivai@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              ·
              edit-2
              23 hours ago

              Oh yeah, you absolutely can test it.
              And then it gives you (and this is a real example, with real function names removed)

              find_something > dirpath
              … rm - rf $dirpath/*
              do_something_in_the_dir(dirpath)

              And it will work, but on a failure of a first question, instead of failing gracefully it wipes your hard drive clean.
              You can find shit like that on the regular Internet, but the difference is, it will be downvoted and some nerd will leave a snarky comment explaining why it’s stupid. When llm gives you that, you don’t have ways to distinguish a working code from a slow boiling trap