• @[email protected]
      link
      fedilink
      English
      55 months ago

      I ask GPT for random junk all the time. If it’s important, I’ll double-check the results. I take any response with a grain of salt, though.

      • @[email protected]
        link
        fedilink
        English
        25 months ago

        You are spending more time and effort doing that than you would googling old fashioned way. And if you don’t check, you might as well throwing magic 8-ball, less damage to the environment, same accuracy

        • @[email protected]
          link
          fedilink
          English
          25 months ago

          The latest GPT does search the internet to generate a response, so it’s currently a middleman to a search engine.

          • @[email protected]
            link
            fedilink
            English
            15 months ago

            No it doesn’t. It incorporates unknown number of words from the internet into a machine which only purpose is to sound like a human. It’s an insanely complicated machine, but the truthfulness of the response not only never considered, but also is impossible to take as a deaired result.
            And the fact that so many people aren’t equipped to recognise it behind the way it talks could be buffling, but also very consistent with other choices humanity takes regularly.

          • @[email protected]
            link
            fedilink
            English
            15 months ago

            And some of those citations and quotes will be completely false and randomly generated, but they will sound very believable, so you don’t know truth from random fiction until you check every single one of them. At which point you should ask yourself why did you add unneccessary step of burning small portion of the rainforest to ask random word generator for stuff, when you could not do that and look for sources directly, saving that much time and energy

            • @[email protected]
              link
              fedilink
              English
              25 months ago

              I, too, get the feeling, that the RoI is not there with LLM. Being able to include “site:” or “ext:” are more efficient.

              I just made another test: Kaba, just googling kaba gets you a german wiki article, explaining it means KAkao + BAnana

              chatgpt: It is the combination of the first syllables of KAkao and BEutel - Beutel is bag in german.

              It just made up the important part. On top of chatgpt says Kaba is a famous product in many countries, I am sure it is not.

              • @[email protected]
                link
                fedilink
                English
                1
                edit-2
                5 months ago

                LLMs are great at cutting through noise

                Even that is not true. It doesn’t have aforementioned criteria for truth, you can’t make it have one.
                LLMs are great at generating noise that humans have hard time distinguishing from a text. Nothing else. There are indeed applications for it, but due to human nature, people think that since the text looks like something coherent, information contained will also be reliable, which is very, very dangerous.

              • @[email protected]
                link
                fedilink
                English
                15 months ago

                You do have this issue, you can’t not have this issue, your LLM, no matter how big the model is and how much tooling you use, does not have criteria for truth. The fact that you made this invisible for you is worse, so much worse.

      • @[email protected]
        link
        fedilink
        English
        25 months ago

        So, if it isn’t important, you just want an answer, and you don’t care whether it’s correct or not?

        • @[email protected]
          link
          fedilink
          English
          25 months ago

          The same can be said about the search results. For search results, you have to use your brain to determine what is correct and what is not. Now imagine for a moment if you were to use those same brain cells to determine if the AI needs a check.

          AI is just another way to process the search results, that happens to give you the correct answer up front, most of the time. If you go blindly trust it, that’s on you.

            • @[email protected]
              link
              fedilink
              English
              05 months ago

              If you knew what the sources were, you wouldn’t have needed to search in the first place. Just because it’s on a reputable website does not make it legit. You still have to reason.