• ugo@feddit.it
    link
    fedilink
    English
    arrow-up
    18
    arrow-down
    1
    ·
    1 month ago

    But the article author wasn’t interfacing with chatgpt, she was interfacing with a human paid to help with the things the article author did not know. The wedding planner was a supposed expert in this interaction, but instead simply sent back regurgitated chatgpt slop.

    Is this the fault of the wedding planner? Yes. Is it the fault of chatgpt? Also yes.

    • conciselyverbose@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      19
      ·
      1 month ago

      Scams are LLM’s best use case.

      They’re not capable of actual intelligence or providing anything that would remotely mislead a subject matter expert. You’re not going to convince a skilled software developer that your LLM slop is competent code.

      But they’re damn good at looking the part to convince people who don’t know the subject that they’re real.