• FaceDeer@fedia.io
    link
    fedilink
    arrow-up
    4
    ·
    3 months ago

    I expect if you follow the references you’d find one of them to be one of those “if Earth was a grain of sand” analogies.

    People like laughing at AI but usually these silly-sounding answers accurately reflect the information the search returned.

    • conciselyverbose@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      0
      ·
      3 months ago

      It’s in the quote that they scaled it.

      The point is that the entire alleged value is the ability to parse the reading material and extract the key points, but because it doesn’t resemble intelligence in any way, it isn’t actually capable of meaningfully doing so.

      Yes, not being able to distinguish between the real answer and a “banana for scale” analogy is a big problem that shows how fucking useless the technology is.

      • FaceDeer@fedia.io
        link
        fedilink
        arrow-up
        0
        ·
        3 months ago

        Except it is capable of meaningfully doing so, just not in 100% of every conceivable situation. And those rare flubs are the ones that get spread around and laughed at, such as this example.

        There’s a nice phrase I commonly use, “don’t let the perfect be the enemy of the good.” These AIs are good enough at this point that I find them to be very useful. Not perfect, of course, but they don’t have to be as long as you’re prepared for those occasions, like this one, where they give a wrong result. Like any tool you have some responsibility to know how to use it and what its capabilities are.

        • conciselyverbose@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          0
          ·
          3 months ago

          No, it isn’t.

          You’re allowing a simple tool with literally zero reading comprehension to do your reading for you. It’s not surprising your understanding of what the tech is is lacking.

          • FaceDeer@fedia.io
            link
            fedilink
            arrow-up
            0
            ·
            3 months ago

            Your comment is simply counterfactual. I do indeed find LLMs to be useful. Saying “no you don’t!” Is frankly ridiculous.

            I’m a computer programmer. Not directly experienced with LLMs themselves, but I understand the technology around them and have written program that make use of them. I know what their capabilities and limitations are.

            • conciselyverbose@sh.itjust.works
              link
              fedilink
              English
              arrow-up
              0
              ·
              3 months ago

              Your claim that it’s capable of doing what it claims isn’t just false.

              It’s an egregious, massively harmful lie, and repeating it is always extremely malicious and inexcusable behavior.

              • FaceDeer@fedia.io
                link
                fedilink
                arrow-up
                0
                ·
                3 months ago

                I have genuinely found LLMs to be useful in many contexts. I use them to brainstorm and flesh out ideas for tabletop roleplaying adventures, to write song lyrics, to write Python scripts to do various random tasks. I’ve talked with them to learn about stuff, and verified that they were correct by checking their references. LLMs are demonstrably capable of these things. I demonstrated it.

                Go ahead and refrain from using them yourself if you really don’t want to, for whatever reason. But exclaiming “no it doesn’t!” In the face of them actually doing the things you say they don’t is just silly.

                • conciselyverbose@sh.itjust.works
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  3 months ago

                  They absolutely cannot reliably summarize the result of searches, like this post is about, and OP in and of itself proves conclusively.

                  Any meaningful rate of failures at all makes them massively, catastrophically damaging to humanity as a whole. “Just don’t use them” absolutely does not prevent their harm. Pushing them as competent is extremely fucking unacceptable behavior.

                  And this is all completely ignoring the obscene energy costs associated with making web searches complete and utter dogshit.

                  • FaceDeer@fedia.io
                    link
                    fedilink
                    arrow-up
                    0
                    ·
                    3 months ago

                    They absolutely cannot reliably summarize the result of searches, like this post is about

                    The problem is that it did summarize the result of this search, the results of this search included one of those “if the Earth was the size of a grain of sand, Alpha Centauri would be X kilometers away” analogies. It did exactly the thing you’re saying it can’t do.

                    Any meaningful rate of failures at all makes them massively, catastrophically damaging to humanity as a whole.

                    Nothing is perfect. Does that make everything a massive catastrophic threat to humanity? How have we managed to survive for this long?

                    You’re ridiculously overblowing this. It’s a “ha ha, looks like AI made a whoopsie because I didn’t understand that I actually asked it to do” situation. It’s not Skynet coming to convince us to eat cyanide.

                    And this is all completely ignoring the obscene energy costs associated with making web searches complete and utter dogshit.

                    Of course it’s ignoring that. It’s not real.

                    You realize that energy costs money? If each web search cost an “obscene” amount, how is Microsoft managing to pay for it all? Why are they paying for it? Do you think they’ll continue paying for it indefinitely? It’d be a completely self-solving problem.