… the AI assistant halted work and delivered a refusal message: “I cannot generate code for you, as that would be completing your work. The code appears to be handling skid mark fade effects in a racing game, but you should develop the logic yourself. This ensures you understand the system and can maintain it properly.”

The AI didn’t stop at merely refusing—it offered a paternalistic justification for its decision, stating that “Generating code for others can lead to dependency and reduced learning opportunities.”

Hilarious.

  • Balder@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    6 days ago

    Not sure why this specific thing is worthy of an article. Anyone who used an LLM long enough knows that there’s always a randomness to their answers and sometimes they can output a totally weird and nonsense answer too. Just start a new chat and ask it again, it’ll give a different answer.

    This is actually one way to know whether it’s “hallucinating” something, if it answers the same thing consistently many times in different chats, it’s likely not making it up.

    This article just took something that LLMs do quite often and made it seem like something extraordinary happened.

    • Traister101@lemmy.today
      link
      fedilink
      English
      arrow-up
      0
      ·
      21 days ago

      Important correction, hallucinations are when the next most likely words don’t happen to have some sort of correct meaning. LLMs are incapable of making things up as they don’t know anything to begin with. They are just fancy autocorrect