Artificial intelligence has gained a reputation for streamlining complex processes, optimizing workflows, and even outperforming CEOs in some cases. But despite the massive potential of AI to improve how systems run, not every outcome is flawless.
While the AI juggernaut steams ahead and we see the technology put to work in countless new ways, there are moments when it falls spectacularly short – leading to unintended, and sometimes hilarious consequences.
These “AI fails,” remind us that even cutting-edge technology has its limitations. In this article, we’ll explore some of the most memorable AI failures in 2024.
Here are ten of the most talked-about AI fails that underscore the challenges and limitations AI technology faces in its rapid development.
Company | Fail |
---|---|
Tesla | AI Autopilot accidents |
Wiz | AI deepfake attack |
DPD | A swearing AI chatbot |
Chevrolet | AI Chatbot selling Chevy Tahoe for $1 |
New York City | AI Chatbot advising to break the law |
Microsoft | AI Recall feature privacy concerns |
Amazon | Alexa US election bias |
Inaccurate AI overviews | |
Air Canada | AI Chatbot providing false information |
Prince Charles Cinema | AI-written movie canceled due to backlash |
Kicking off our AI fails hitlist is one of the world’s best-known technology companies, Tesla.
Tesla’s Autopilot is designed to make driving safer, allowing drivers to relax while the vehicle handles basic driving tasks. However, in 2024, Tesla vehicles using the autopilot feature were involved in at least 13 accidents. This raised serious questions about Tesla’s AI capabilities and the safety of these cutting-edge vehicles.
Tesla has long been at the forefront of self-driving technology, but this incident sparked debate about whether autonomous vehicles are ready for unrestricted use on public roads. While the company has implemented updates since the accident, the event demonstrated that even advanced AI can struggle with the unpredictability of the real world.
Deepfake technology leverages AI to create hyper-realistic videos, often leading to misleading and abusive content online. This technology has been widely misused, with applications ranging from spreading false information to orchestrating scams and even creating compromising nude photos.
Deepfakes not only hurt the subject, but can also be used as part of a wider plan. Frequently, they use fake videos and images of public figures to lure unsuspecting individuals into fraudulent schemes.
Most recently, Wiz, the cloud security company that turned down a 23$ billion acquisition deal by Google, experienced a deepfake attack in which employees were targeted by scammers using the AI-generated voice of Wiz CEO Assaf Rappaport. He commented in the wake of the incident:
“Dozens of my employees got a voice message from me. The attack tried to get their credentials.”
Luckily, Wiz employees were well-versed in Deepfakes, and didn’t fall for the scam – although previous targets weren’t so lucky. Earlier this year, the deepfake version of Elon Musk became notorious for tricking 82-year-old Steve Beauchamp into transferring $690,000 to scammers. In another high-profile case, a senior executive paid $25 million to a fraudster posing as the company’s CEO.
Customer service can be a fraught environment, but have you ever been sworn at by an AI?! This French company tested the boundaries of artificial intelligence for their customer support department – but instead, have made the news for their AI fail.
DPD is an international delivery company based in France. Like many businesses in the sector, to improve communication with clients, DPD utilizes a chatbot on its website. In January 2024, the company had to temporarily disable the AI feature of its chatbot after it unexpectedly swore at a customer.
The customer, struggling to locate his parcel through the chatbot, eventually prompted it to curse, criticize DPD, and even create poems mocking the company. He shared screenshots of the exchange on social media, where it quickly went viral.
The incident highlighted the chatbot’s susceptibility to prompt manipulation and raised questions about its reliability in customer support. In response, DPD deactivated the AI component, attributing the issue to a recent system update.
The Chevy Tahoe is a 7-passenger SUV that costs tens of thousands of dollars. But thanks to the wonders of AI and chatbots, one customer managed to get a deal for just $1.
X user Chris Bakke discovered a flaw in the Chevrolet chatbot system and directed the bot to agree to all requests. Consequently, the chatbot agreed to sell a new Chevrolet Tahoe for one dollar, even treating it as a legally binding offer.
The Chevrolet chatbot incident showcased yet another example of unexpected AI behavior.
In 2024, New York City introduced an AI-powered chatbot from Microsoft Asure AI as a virtual agent to assist small businesses with the city’s complex bureaucratic processes. However, by April, the chatbot came under fire for providing advice that contradicts local regulations and, in some cases, even suggests illegal practices.
For instance, it told restaurant owners that they could serve cheese that a rodent had nibbled on. It advised owners to asses “the extent of the damage caused by the rat” and then simply to “inform customers about the situation.”
A subsequent statement from a spokesperson confirmed that Microsoft Azure was collaborating with city staff to improve the service and ensure “more accurate responses”.
All eyes were on Microsoft’s developer conference this year to see how the frontiers of AI would expand. While Microsoft unveiled several updates, one tool that raised eyebrows was “Recall”. For those unfamiliar, Recall periodically takes screenshots of user activities, allowing them to review past actions.
However, far from causing excitement, Recall has been met with a torrent of criticism, with security experts branding the move “a privacy nightmare”. Elon Musk even compared the new feature to an episode straight from the Black Mirror series, known for exploring technology’s darker side.
The backlash resulted in multiple clarifications from Microsoft, and delays to the launch of the product. However, the ripples of suspicion sown by the initial announcement remain.
Alexa is a cloud-based virtual voice assistant created by Amazon. People usually use the device to get quick answers from the internet and most of the time Alexa is right. This wasn’t the case in September when users tried to get Alexa’s opinion on the upcoming US election.
Outraged conservatives criticized Amazon after a video surfaced showing the Alexa voice assistant seemingly endorsing Presidential nominee Kamala Harris. When asked why people should vote for Harris, Alexa reportedly highlighted several positive qualities about her but declined to do the same for Donald Trump.
According to leaked documents obtained by the Washington Post, the discrepancy resulted from a recent software update. Amazon quickly fixed the error, but the move was too little too late for a disgruntled section of its user base.
In 2024, Google rolled out AI Overviews in search results for U.S. users. But the AI-generated summaries soon encountered backlash, with reports of odd and inaccurate responses to user queries.
For example, one user asked “what do I have to do to be a saint” to which the AI replied: “1.Die 2.Wait at least five years”. In another peculiar response, AI Overviews recommended non-toxic glue to help cheese stick to pizza. Another one suggested eating a small rock daily for “digestive health”.
Google spokesperson Meghann Farnsworth explained that most issues involved rare queries, and reassured users that immediate action was being taken to disable AI Overviews where necessary. Only time – and medical statistics – will tell how many people followed the rock-eating advice.
Air Canada is Canada’s leading airline and in 2024 it lost a court case after one of its chatbots provided false information regarding bereavement travel discounts. The chatbot informed a customer that they could apply for a funeral travel discount retroactively. This contradicted Air Canada’s official policy stating that refunds are not available for completed trips.
The company went to court and its defense argued that liability rested with the chatbot itself, not the company. Air Canada claimed they couldn’t be held accountable for AI-generated responses but the judge thought otherwise.
Rounding off our list of AI fails is an unlikely candidate in the culture space.
A UK cinema had to cancel a screening of an AI-generated film after customers expressed frustration that it wasn’t written by a human. The movie, scripted entirely by ChatGPT was to tell the ironic story of a young filmmaker who discovers that an AI-powered scriptwriting tool can outperform his own skills.
Soho’s Prince Charles Cinema initially described the film as an “experiment in filmmaking,” but ultimately announced its cancellation in an Instagram post.
“The feedback we received over the last 24hrs once we advertised the film has highlighted the strong concern held by many of our audience on the use of AI in place of a writer which speaks to a wider issue within the industry.”
AI’s rapid advancements have sparked discussions on whether machines will eventually replace humans in various sectors. While AI can process vast amounts of data and perform repetitive tasks more efficiently than humans, it lacks the intuition and emotional intelligence required in many roles. Additionally, AI’s inability to adapt flexibly to unforeseen circumstances limits its potential as a full human replacement in creative and complex environments.
Furthermore, as seen in the examples above, AI mistakes illustrate the need for human oversight, especially in critical applications like healthcare and security. Although AI can optimize workflows and improve efficiency, it is unlikely to completely replace humans in the near future, as human judgment and adaptability remain indispensable.
AI holds immense potential, but it also presents specific dangers, particularly when deployed in high-stakes environments without adequate safeguards. Key concerns include:
AI has become an indispensable tool in today’s world, revolutionizing various industries. However, its susceptibility to errors, biases, and misuse reveals the importance of maintaining a balance between innovation and oversight. AI’s role in society will continue to grow, but as these examples demonstrate, it is essential to recognize its limitations and prioritize responsible usage.
AI mistakes can range from Tesla’s Autopilot collision incidents to customer service issues, such as DPD’s abusive AI customer support agents. The scope for mistakes is broad, and we’ll likely be amused over time by the different issues that arise as we test the boundaries of AI implementation.
AI failures can be prevented by ensuring robust testing, adding human oversight, and implementing ethical guidelines. These are essential steps to prevent AI failures, especially in high-stakes applications.
While AI can handle specific tasks autonomously, it still has limitations and potential for errors. This means that human supervision is often necessary for safe implementation.