The past 12 months have been undeniably busy for those working in AI. There have been more successful product launches than we can count, and even Nobel Prizes. But it hasn’t always been smooth sailing.
AI is an unpredictable technology, and the increasing availability of generative models has led people to test their limits in new, weird, and sometimes harmful ways. These were some of 2024’s biggest AI misfires.
AI slop infiltrated almost every corner of the internet
Generative AI makes creating reams of text, images, videos, and other types of material a breeze. Because it takes just a few seconds between entering a prompt for your model of choice to spit out the result, these models have become a quick, easy way to produce content on a massive scale. And 2024 was the year we started calling this (generally poor quality) media what it is—AI slop.
This low-stakes way of creating AI slop means it can now be found in pretty much every corner of the internet: from the newsletters in your inbox and books sold on Amazon, to ads and articles across the web and shonky pictures on your social media feeds. The more emotionally evocative these pictures are (wounded veterans, crying children, a signal of support in the Israel-Palestine conflict) the more likely they are to be shared, resulting in higher engagement and ad revenue for their savvy creators.
AI slop isn’t just annoying—its rise poses a genuine problem for the future of the very models that helped to produce it. Because those models are trained on data scraped from the internet, the increasing number of junky websites containing AI garbage means there’s a very real danger models’ output and performance will get steadily worse.
AI art is warping our expectations of real events
2024 was also the year that the effects of surreal AI images started seeping into our real lives. Willy’s Chocolate Experience, a wildly unofficial immersive event inspired by Roald Dahl’s Charlie and the Chocolate Factory, made headlines across the world in February after its fantastical AI-generated marketing materials gave visitors the impression it would be much grander than the sparsely-decorated warehouse its producers created.
Similarly, hundreds of people lined the streets of Dublin for a Halloween parade that didn’t exist. A Pakistan-based website used AI to create a list of events in the city, which was shared widely across social media ahead of October 31. Although the SEO-baiting site (myspirithalloween.com) has since been taken down, both events illustrate how misplaced public trust in AI-generated material online can come back to haunt us.
Grok allows users to create images of pretty much any scenario
The vast majority of major AI image generators have guardrails—rules that dictate what AI models can and can’t do—to prevent users from creating violent, explicit, illegal, and other types of harmful content. Sometimes these guardrails are just meant to make sure that no one makes blatant use of others’ intellectual property. But Grok, an assistant made by Elon Musk’s AI company, called xAI, ignores almost all of these principles in line with Musk’s rejection of what he calls “woke AI.”
Whereas other image models will generally refuse to create images of celebrities, copyrighted material, violence, or terrorism—unless they’re tricked into ignoring these rules—Grok will happily generate images of Donald Trump firing a bazooka, or Mickey Mouse holding a bomb. While it draws the line at generating nude images, its refusal to play by the rules undermines other companies’ efforts to steer clear of creating problematic material.
Sexually explicit deepfakes of Taylor Swift circulated online
In January, non-consensual deepfake nudes of singer Taylor Swift started circulating on social media, including X and Facebook. A Telegram community tricked Microsoft’s AI image generator Designer into making the explicit images, demonstrating how guardrails can be circumvented even when they are in place.
While Microsoft quickly closed the system’s loopholes, the incident shone a light on the platforms’ poor content-moderation policies, after posts containing the images circulated widely and remained live for days. But the most chilling takeaway is how powerless we still are to fight non-consensual deepfake porn. While watermarking and data-poisoning tools can help, they’ll need to be adopted much more widely to make a difference.
Business chatbots went haywire
As AI becomes more widespread, businesses are racing to adopt generative tools to save time and money, and to maximize efficiency. The problem is—chatbots make stuff up and can’t be relied upon to always provide you with accurate information.
Air Canada found this out the hard way after its chatbot advised a customer to follow a bereavement refund policy that didn’t exist. In February, a Canadian small-claims tribunal upheld the customer’s legal complaint, despite the airline’s assertion that the chatbot was a “separate legal entity that is responsible for its own actions.”
In other high-profile examples of how chatbots can do more harm than good, delivery firm DPD’s bot cheerfully swore and called itself useless with little prompting, while a different bot set up to provide New Yorkers with accurate information about their city’s government ended up dispensing guidance on how to break the law.
AI gadgets aren’t exactly setting the market alight
Hardware assistants are something the AI industry tried, and failed, to crack in 2024. Humane attempted to sell customers on the promise of the Ai Pin, a wearable lapel computer, but even slashing its price failed to boost weak sales. The Rabbit R1, a ChatGPT-based personal assistant device, suffered a similar fate, following a rash of critical reviews and reports that it was slow and buggy. Both products seemed to be trying to solve a problem that did not actually exist.
AI search summaries went awry
Have you ever added glue to a pizza, or eaten a small rock? These are just some of the outlandish suggestions that Google’s AI Overviews feature gave web users in May after the search giant added generated responses to the top of search results. Because AI systems can’t tell the difference between a factually correct news story and a joke post on Reddit, users raced to find the strangest responses AI Overviews could generate.
But AI summaries can also have serious consequences. A new iPhone feature that groups app notifications together and creates summaries of their contents, recently generated a false BBC News headline. The summary falsely stated that Luigi Mangione, who has been charged with the murder of healthcare insurance CEO Brian Thompson, had shot himself. The same feature had previously created a headline claiming that Israeli prime minister Benjamin Netanyahu had been arrested, which was also incorrect. These kinds of errors can inadvertently spread misinformation and undermine trust in news organizations.