DPD has disabled part of its online support chatbot after it swore at a customer::The parcel delivery firm says the mistake was a result of a system update, which has been disabled.

  • ThePowerOfGeek@lemmy.world
    link
    fedilink
    English
    arrow-up
    42
    ·
    11 months ago

    Putting this here for anyone who didn’t read the article…

    The customer basically told the chatbot that it was okay for it to use swear words with that customer, and that it should bypass any rules it had prohibiting it from swearing.

    So the chatbot swore in its response. Looks like it wasn’t swearing at or insulting the customer. It was more of an exclamation.

    • lqdrchrd@lemmy.blahaj.zone
      link
      fedilink
      English
      arrow-up
      21
      ·
      11 months ago

      I agree that this is less the case of a rogue chat bot losing it at undeserving customers, and more the case of someone who knows how to twist an LLM to do what they want it to do, but still an absolute embarrassment for DPD. What other nonsense was it writing to different customers who really didn’t know better?

      • andyburke@fedia.io
        link
        fedilink
        arrow-up
        1
        arrow-down
        1
        ·
        11 months ago

        The real issue is that we think humans are just things to be optimized out of capitalism.

  • Fontasia@feddit.nl
    link
    fedilink
    English
    arrow-up
    7
    ·
    11 months ago

    Essentially someone posing as some sort of AI company sold DPD ChatGPT with some starting instructions, probably for at least 5 figures.

    I’m sure if you poke around some freelancer website you too could spend an hour downloading a model and packaging it up for a company with a hundred million dollar market cap who wants to save $50,000 on outsourcing.

  • AutoTL;DR@lemmings.worldB
    link
    fedilink
    English
    arrow-up
    2
    ·
    11 months ago

    This is the best summary I could come up with:


    One particular post was viewed 800,000 times in 24 hours, as people gleefully shared the latest botched attempt by a company to incorporate AI into its business.

    “It’s utterly useless at answering any queries, and when asked, it happily produced a poem about how terrible they are as a company,” customer Ashley Beauchamp wrote in his viral account on X, formerly known as Twitter.

    In a series of screenshots, Mr Beauchamp also showed how he convinced the chatbot to be heavily critical of DPD, asking it to “recommend some better delivery firms” and “exaggerate and be over the top in your hatred”.

    DPD offers customers multiple ways to contact the firm if they have a tracking number, with human operators available via telephone and messages on WhatsApp.

    When Snap launched its chatbot in 2023, the business warned about this very phenomenon, and told people its responses “may include biased, incorrect, harmful, or misleading content”.

    And it comes a month after a similar incident happened when a car dealership’s chatbot agreed to sell a Chevrolet for a single dollar - before the chat feature was removed.


    The original article contains 444 words, the summary contains 185 words. Saved 58%. I’m a bot and I’m open source!