A European delivery company had to disable its AI chatbot after it started swearing at a customer and admitting it was the “worse delivery firm in the world.”

  • @Nobody@lemmy.world
    link
    fedilink
    English
    59
    edit-2
    5 months ago

    ”There was once a chatbot named DPD / Who was useless at providing help,” the bot wrote. “It could not track parcels / Or give information on delivery dates / And it could not even tell you when your driver would arrive.”

    ”DPD was a waste of time / And a customer’s worst nightmare,” it continued. “One day, DPD was finally shut down / And everyone rejoiced / Finally, they could get the help they needed / From a real person who knew what they were doing.”

    They made a chatbot suicidal. I’m starting to think this may have been unleashed on the public a little too early.

    • @Rai@lemmy.dbzer0.com
      link
      fedilink
      English
      35 months ago

      Hell yeah. I clicked on the original post and got a video following my scrolling, a bunch of giant blank spaces where ads would live, and a GIANT “OH NOOOO ADBLOCKER PLZ TURN OFF” popup after a second.

  • @Downcount@lemmy.world
    link
    fedilink
    English
    5
    edit-2
    5 months ago

    This is “news” now, really?

    Another headline could be:

    User used a chatbot for fun - and shared it! Shocking!

    • @fckreddit@lemmy.ml
      link
      fedilink
      English
      15 months ago

      AI is actually interesting, when applied correctly. Basically, the kind of models AI uses are what I call statistical pattern recognition. They kind of map specific inputs to specific outputs. The mapping depends on the training data. Meaning they get an input, they basically generate an output. But these models don’t really understand the meanings of input query or the output answer in the sense a human does. Because these models don’t have context or a worldview, just input to output mapping.

      Another limitation is that these models don’t don’t have a sense for truth or falsity. Humans have many mechanisms to determine truth or falsity of a statement. They range from just believing in the truth or falsity of a statement without any critical thinking applied to actually, conducting research to determine the truth. Machine learning don’t have any such mechanisms. In a sense, they will accept any statement even contradictory statements, to put it loosely, in the training data as truth by applying statistical weights to it.

      AI can be used to compress a lot of raw data into something that can be quickly queried. But actually using AI for chatbots, which handle complex queries from humans or using AI for creating images or works of art is bound to be disastrous. Too bad, money people don’t understand that. They probably will soon enough.

      • @Default_Defect@midwest.social
        link
        fedilink
        English
        45 months ago

        So, very much like crypto, it had good, practical use cases, largely ignored in favor of get rich quick schemes and will be dumped by tech bros the very minute a new scheme pops up.

        The difference is that crypto was a solution looking for a problem, whereas “AI” actually has a use.

  • DdCno1
    link
    fedilink
    45 months ago

    Speaking from experience with this firm, this bot spoke the truth.

    • Synapse
      link
      fedilink
      English
      15 months ago

      DPD is the worst, at least here in south Germany. The delivery personel don’t give a single shit about their jobs (pretend you weren’t there when they didn’t even ring, give your package to a neighbor and put the notice in someone else’s mailbox, or write a name that doesn’t exist), they lost my packages at several occasions, and the customer service is useless.