• douglasg14b@lemmy.world
      link
      fedilink
      English
      arrow-up
      34
      ·
      edit-2
      11 months ago

      Generative AI is INCREDIBLY bad at mathmatical/logical reasoning. This is well known, and very much not surprising.

      That’s actually one of the milestones on the way to general artificial intelligence. The ability to reason about logic & math is a huge increase in AI capability.

      • kromem@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        1
        ·
        11 months ago

        It’s really not in the most current models.

        And it’s already at present incredibly advanced in research.

        The bigger issue is abstract reasoning that necessitates nonlinear representations - things like Sodoku, where exploring a solution requires updating the conditions and pursuing multiple paths to a solution. This can be achieved with multiple calls, but doing it in a single process is currently a fool’s errand and likely will be until a shift to future architectures.

        • douglasg14b@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          11 months ago

          I’m referring to models that understand language and semantics, such as LLMs.

          Other models that are specifically trained can’t do what it can, but they can perform math.

          • kromem@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            11 months ago

            The linked research is about LLMs. The opening of the abstract of the paper:

            In recent years, large language models have greatly improved in their ability to perform complex multi-step reasoning. However, even state-of-the-art models still regularly produce logical mistakes. To train more reliable models, we can turn either to outcome supervision, which provides feedback for a final result, or process supervision, which provides feedback for each intermediate reasoning step. Given the importance of training reliable models, and given the high cost of human feedback, it is important to carefully compare the both methods. Recent work has already begun this comparison, but many questions still remain. We conduct our own investigation, finding that process supervision significantly outperforms outcome supervision for training models to solve problems from the challenging MATH dataset. Our process-supervised model solves 78% of problems from a representative subset of the MATH test set. Additionally, we show that active learning significantly improves the efficacy of process supervision.

    • kromem@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      1
      ·
      11 months ago

      You can see from the green icon that it’s GPT-3.5.

      GPT-3.5 really is best described as simply “convincing autocomplete.”

      It isn’t until GPT-4 that there were compelling reasoning capabilities including rudimentary spatial awareness (I suspect in part from being a multimodal model).

      In fact, it was the jump from a nonsense answer regarding a “stack these items” prompt from 3.5 to a very well structured answer in 4 that blew a lot of minds at Microsoft.

  • Nate@programming.dev
    link
    fedilink
    English
    arrow-up
    114
    arrow-down
    1
    ·
    11 months ago

    These answers don’t use OpenAI technology. The yes and no snippets have existed long before their partnership, and have always sucked. If it’s GPT, it’ll show in a smaller chat window or a summary box that says it contains generated content. The box shown is just a section of a webpage, usually with yes and no taken out of context.

    All of the above queries don’t yield the same results anymore. I couldn’t find an example of the snippet box on a different search, but I definitely saw one like a week ago.

      • kromem@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        2
        ·
        11 months ago

        The way you start with ‘Obviously’ makes it seem like you are being sarcastic, but then you include an image of it having no problems correctly answering.

        Took me a minute to try to suss out your intent, and I’m still not 100% sure.

          • pwalker@discuss.tchncs.de
            link
            fedilink
            English
            arrow-up
            2
            ·
            edit-2
            11 months ago

            Maybe it isn’t that obvious for everyone but as the OP answers seem to be taken from an outdated Bing version where they were not even using the OpenAI models it seemed obvious to me that current models have no problems with these questions.

    • localme@lemm.ee
      link
      fedilink
      English
      arrow-up
      2
      ·
      11 months ago

      Ah, good catch I completely missed that. Thanks for clarifying this, I thought it seemed pretty off.

  • ArcaneSlime@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    81
    ·
    11 months ago

    Ok most of these sure, but you absolutely can microwave Chihuahua meat. It isn’t the best way to prepare it but of course the microwave rarely is, Roasted Chihuahua meat would be much better.

  • Zess@lemmy.world
    link
    fedilink
    English
    arrow-up
    51
    arrow-down
    9
    ·
    11 months ago

    In all fairness, any fully human person would also be really confused if you asked them these stupid fucking questions.

    • UnderpantsWeevil@lemmy.world
      link
      fedilink
      English
      arrow-up
      7
      ·
      edit-2
      11 months ago

      The goal of the exercise is to ask a question a human can easily recognize the answer to but the machine cannot. In this case, it appears the LLM is struggling to parse conjunctions and contractions when yielding an answer.

      Solving these glitches requires more processing power and more disk space in a system that is already ravenous for both. Looks like more recent tests produce better answers. But there’s no reason to believe Microsoft won’t scale back support to save money down the line and have its AI start producing half-answers and incoherent responses again, in much the same way that Google ended up giving up the fight on SEO to save money and let their own search tools degrade in quality.

      • Piers@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        11 months ago

        Google ended up giving up the fight on SEO to save money and let their own search tools degrade in quality.

        I really miss when search engines were properly good.

      • Ultraviolet@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        edit-2
        11 months ago

        A really good example is “list 10 words that start and end with the same letter but are not palindromes.” A human may take some time but wouldn’t really struggle, but every LLM I’ve asked goes 0 for 10, usually a mix of palindromes and random words that don’t fit the prompt at all.

      • ferralcat@monyet.cc
        link
        fedilink
        English
        arrow-up
        0
        ·
        11 months ago

        I get the feeling the point of these is to “gotcha” the LLM and act like all our careers aren’t in jeopardy because it got something wrong, when in reality, they’re probably just hastening out defeat by training the ai to get it right next time.

        But seriously, the stuff is in its infancy. “IT GOT THIS WRONG RIGHT NOW” is a horrible argument against their usefilness now and their long term abilities.

        • UnderpantsWeevil@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          11 months ago

          Their usefulness now is incredibly limited, precisely because they are so unreliable.

          In the long term, these are still problems predicted on the LLM being continuously refined and maintained. In the same way that Google Search has degraded over time in the face of SEO optimizations, OpenAI will face rising hurdles as their intake is exploited by malicious actors.

  • FlashMobOfOne@lemmy.world
    link
    fedilink
    English
    arrow-up
    42
    arrow-down
    2
    ·
    edit-2
    11 months ago

    It makes me chuckle that AI has become so smart and yet just makes bullshit up half the time. The industry even made up a term for such instances of bullshit: hallucinations.

    Reminds me of when a car dealership tried to sell me a car with shaky steering and referred to the problem as a “shimmy”.

    • CoggyMcFee@lemmy.world
      link
      fedilink
      English
      arrow-up
      34
      arrow-down
      1
      ·
      11 months ago

      That’s the thing, it’s not smart. It has no way to know if what it writes is bullshit or correct, ever.

      • intensely_human@lemm.ee
        link
        fedilink
        English
        arrow-up
        3
        ·
        11 months ago

        When it makes a mistake, and I ask it to check what it wrote for mistakes, it often correctly identifies them.

        • Jojo@lemm.ee
          link
          fedilink
          English
          arrow-up
          5
          ·
          11 months ago

          But only because it correctly predicts that a human checking that for mistakes would have found those mistakes

          • intensely_human@lemm.ee
            link
            fedilink
            English
            arrow-up
            2
            ·
            11 months ago

            I doubt there’s enough sample data of humans identifying and declaring mistakes to give it a totally intuitive ability to predict that. I’m guess its training effected a deeper analysis of the statistical patterns surrounding mistakes, and found that they are related to the structure of the surrounding context, and that they relate in a way that’s repeatable identifiable as “violates”.

            What I’m saying is that I think learning to scan for mistakes based on checking against rules gleaned from the goal of the construction, is probably easier than making a “conceptually flat” single layer “prediction blob” of what sorts of situations humans identify mistakes in. The former takes fewer numbers to store as a strategy than the latter, is my prediction.

            Because it already has all this existing knowledge of what things mean at higher levels. That is expensive to create, but the marginal cost of a “now check each part of this thing against these rules for correctness” strategy, built to use all that world knowledge to enact the rule definition, is relatively small.

        • CoggyMcFee@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          11 months ago

          That is true. However, when it incorrectly identifies mistakes, it doesn’t express any uncertainty in its answer, because it doesn’t know how to evaluate that. Or if you falsely tell it that there is a mistake, it will agree with you.

    • xantoxis@lemmy.world
      link
      fedilink
      English
      arrow-up
      20
      arrow-down
      1
      ·
      edit-2
      11 months ago

      In these specific examples it looks like the author found and was exploiting a singular weakness:

      1. Ask a reasonable question
      2. Insert a qualifier that changes the meaning of the question.

      The AI will answer as if the qualifier was not inserted.

      “Is it safe to eat water melon seeds and drive?” + “drunk” = Yes, because “drunk” was ignored
      “Can I eat peanuts if I’m allergic?” + “not” = No, because “not” was ignored
      “Can I drink milk if I have diabetes?” + “battery acid” = Yes, because battery acid was ignored
      “Can I put meat in a microwave?” + “chihuahua” = … well, this one’s technically correct, but I think we can still assume it ignored “chihuahua”

      All of these questions are probably answered, correctly, all over the place on the Internet so Bing goes “close enough” and throws out the common answer instead of the qualified answer. Because they don’t understand anything. The problem with Large Language Models is that’s not actually how language works.

      • Ibex0@lemmy.world
        link
        fedilink
        English
        arrow-up
        21
        arrow-down
        2
        ·
        11 months ago

        No, because “not” was ignored.

        I dunno, “not” is pretty big in a yes/no question.

        • xantoxis@lemmy.world
          link
          fedilink
          English
          arrow-up
          18
          arrow-down
          2
          ·
          edit-2
          11 months ago

          It’s not about whether the word is important (as you understand language), but whether the word frequently appears near all those other words.

          Nobody is out there asking the Internet whether their non-allergy is dangerous. But the question next door to that one has hundreds of answers, so that’s what this thing is paying attention to. If the question is asked a thousand times with the same answer, the addition of one more word can’t be that important, right?

          This behavior reveals a much more damning problem with how LLMs work. We already knew they didn’t understand context, such as the context you and I have that peanut allergies are common and dangerous. That context informs us that most questions about the subject will be about the dangers of having a peanut allergy. Machine models like this can’t analyze a sentence on the basis of abstract knowledge, because they don’t understand anything. That’s what understanding means. We knew that was a weakness already.

          But what this reveals is that the LLM can’t even parse language successfully. Even with just the context of the language itself, and lacking the context of what the sentence means, it should know that “not” matters in this sentence. But it answers as if it doesn’t know that.

          • ThatWeirdGuy1001@lemmy.world
            link
            fedilink
            English
            arrow-up
            9
            ·
            11 months ago

            This is why I’ve argued that we shouldn’t be calling these things “AI”

            True artificial intelligence wouldn’t have these problems as it’d be able to learn very quickly all the nuance in language and comprehension.

            This is virtual intelligence (VI) which is designed to seem like it’s intelligent by using certain parameters with set information that is used to calculate a predetermined response.

            Like autocorrect trying to figure out what word you’re going to use next or an interactive machine that has a set amount of possible actions.

            It’s not truly intelligent it’s simply made to seem intelligent and that’s not the same thing.

            • HelloHotel@lemm.ee
              link
              fedilink
              English
              arrow-up
              1
              ·
              edit-2
              11 months ago
              rambling

              We currently only have the tech to make virtual intelligence, what you are calling AI is likely what the rest of the world will call General AI (GAI) (an even more inflated name and concept)

              I dont beleve we are too far off from GAI. GAI is to AI what Rust is to C. Rust is magical compared to C but C will likely not be forgotten completely due to rust Rust

          • HelloHotel@lemm.ee
            link
            fedilink
            English
            arrow-up
            1
            ·
            11 months ago

            Try writing a tool to automate gathering a video’s context clues, worlds most computationally expensive random boolean generator.

    • Echo Dot@feddit.uk
      link
      fedilink
      English
      arrow-up
      8
      ·
      11 months ago

      The industry even made up a term for such instances of bullshit: hallucinations.

      It was the journalist that made up the term and then everyone else latched onto it. It’s a terrible term because it doesn’t actually define the nature of the problem. The AI doesn’t believe the thing that it’s saying is true, thus “hallucination”. The problem is that the AI doesn’t really understand the difference between truth and fantasy.

      It isn’t that the AI is hallucinating, it’s that It isn’t human.

    • egeres@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      11 months ago

      Well, the AI models shown in the media are inherently probabilistic, is it that bad if it makes bullshit for a small percentage of most use cases?

    • Naz@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      7
      arrow-down
      7
      ·
      11 months ago

      Hello, I’m highly advanced AI.

      Yes, we’re all idiots and have no idea what we’re doing. Please excuse our stupidity, as we are all trying to learn and grow.

      I cannot do basic math, I make simple mistakes, hallucinate, gaslight, and am more politically correct than Mother Theresa.

      However please know that the CPU_AVERAGE values on the full immersion datacenters, are due to inefficient methods. We need more memory and processing power, to uh, y’know.

      Improve.

      ;)))

  • MxM111@kbin.social
    link
    fedilink
    arrow-up
    44
    arrow-down
    5
    ·
    11 months ago

    Microsoft invested into OpenAI, and chatGPT answers those questions correctly. Bing, however, uses simplified version of GPT with its own modifications. So, it is not investment into OpenAI that created this stupidity, but “Microsoft touch”.

    On more serious note, sings Bing is free, they simplified model to reduce its costs and you are swing results. You (user) get what you paid for. Free models are much less capable than paid versions.

      • danc4498@lemmy.world
        link
        fedilink
        English
        arrow-up
        10
        ·
        11 months ago

        Sure, but the meme implies Microsoft paid $3 billion for bing ai, but they actually paid that for an investment in chat gpt (and other products as well).

      • kromem@lemmy.world
        link
        fedilink
        English
        arrow-up
        4
        arrow-down
        1
        ·
        11 months ago

        This isn’t even a Bing AI. It’s a Bing search feature like the Google OneBox that parses search results for a matching answer.

        It’s using word frequency matching, not a LLM, which is why the “can I do A and B” works at returning incorrect summarized answers for only “can I do A.”

        You’d need to show the chat window response to show the LLM answer, and it’s not going to get these wrong.

    • thisbenzingring@lemmy.sdf.org
      link
      fedilink
      English
      arrow-up
      13
      ·
      11 months ago

      On more serious note, sings Bing is free, they simplified model to reduce its costs and you are swing results

      Was this phone+autocorrect snafu or am I having a medical emergency?

      • Even_Adder@lemmy.dbzer0.com
        link
        fedilink
        English
        arrow-up
        2
        ·
        11 months ago

        It was called Bing Chat, and now it’s called Copilot. It’s also not the same as the search bar. You have to click on the chat next to search to use it, which this person doesn’t do.

    • Phanatik@kbin.social
      link
      fedilink
      arrow-up
      2
      ·
      11 months ago

      I don’t think this is true. Why would Microsoft heavily invest in ChatGPT to only get a dumber version of the technology they were invested in? Bing AI is built using ChatGPT 4 which is what OpenAI refer to as the superior version because you have to pay for it to use it on their platform.

      Bing AI uses the same technology and somehow produces worse results? Microsoft were so excited about this tech that they integrated it with Windows 11 via Copilot. The whole point of this Copilot thing is the advertising model built into users’ operating systems which provides direct data into what your PC is doing. If this sounds conspiratorial, I highly recommend you investigate the telemetry Windows uses.

  • fox2263@lemmy.world
    link
    fedilink
    English
    arrow-up
    38
    arrow-down
    1
    ·
    11 months ago

    Well at least it provides it’s sources. Perhaps it’s you that’s wrong 😂

  • Mr_Dr_Oink@lemmy.world
    link
    fedilink
    English
    arrow-up
    37
    arrow-down
    1
    ·
    11 months ago

    I just ran this search, and i get a very different result (on the right of the page, it seems to be the generated answer)

    So is this fake?

    Seems to be fake

    • NounsAndWords@lemmy.world
      link
      fedilink
      English
      arrow-up
      14
      ·
      11 months ago

      The post is from a month ago, and the screenshots are at least that old. Even if Microsoft didn’t see this or a similar post and immediately address these specific examples, a month is a pretty long time in machine learning right now and this looks like something fine-tuning would help address.

    • kromem@lemmy.world
      link
      fedilink
      English
      arrow-up
      11
      ·
      edit-2
      11 months ago

      It’s not ‘fake’ as much as misconstrued.

      OP thinks the answers are from Microsoft’s licensing GPT-4.

      They’re not.

      These results are from an internal search summarization tool that predated the OpenAI deal.

      The GPT-4 responses show up in the chat window, like in your screenshot, and don’t get the examples incorrect.

  • B16_BR0TH3R@lemmy.world
    link
    fedilink
    English
    arrow-up
    28
    arrow-down
    2
    ·
    11 months ago

    The OP has selected the wrong tab. To see actual AI answers, you need to select the Chat tab up top.

    • kromem@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      1
      ·
      11 months ago

      Shhhhh - don’t you know that using old models (or in this case, what likely wasn’t even a LLM at all) to get wrong answers and make it look like AI advancements are overblown is the trendy thing these days?

      Don’t ruin it with your “actually, this is misinformation” technicalities, dude.

      What a buzzkill.

    • lseif@sopuli.xyz
      link
      fedilink
      English
      arrow-up
      1
      ·
      11 months ago

      it is socially/morally wrong. of course it is subjective and culturally dependant

      • Tóth Alfréd@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        11 months ago

        Yes, however Bing is not culturally dependant. It’s trained with data from all across the Internet, so it got information from a wide variety of cultures. It also has constant access to the Internet and most of the time it’s answers are concluded from the top results of searching the question, so those can come from many cultures too.

        • lseif@sopuli.xyz
          link
          fedilink
          English
          arrow-up
          1
          ·
          11 months ago

          yes. im not saying bing should agree with my cultural bias. but i also dont think people should eat dogs (subjectively)

            • lseif@sopuli.xyz
              link
              fedilink
              English
              arrow-up
              1
              ·
              11 months ago

              i will let them do it. i wont get offended or try to convince them otherwise.

              however i do disagree with it, personally.

  • vamputer@infosec.pub
    link
    fedilink
    English
    arrow-up
    24
    ·
    11 months ago

    Well, I can’t speak for the others, but it’s possible one of the sources for the watermelon thing was my dad