Cyabra Launches Deepfake Detection

How Fake Profiles Fueled #BoycottAmazon

Following Amazon’s announcement of rolling back the company’s DEI initiatives, Cyabra conducted a 6-months research into the viral backlash and boycott around Amazon, and found that fake accounts were central to the narrative, posing as grassroots activists, amplifying boycott hashtags, and flooding feeds with service complaints. These inauthentic profiles helped turn a policy decision into a PR crisis. Here’s what Cyabra uncovered: 

Grassroots in Appearance, Synthetic in Origin

In late December 2024, Amazon announced plans to roll back its DEI (Diversity, Equity, and Inclusion) initiatives. By January, public backlash was gaining traction. By February, it had gone viral. But Cyabra’s investigation reveals that what looked like a widespread consumer revolt was in large parts manufactured: 

Over a third of the accounts discussing Amazon from January to June 2025 were fake, producing 1,528 posts and comments fueling the boycott. These inauthentic accounts were active drivers of the conversation, using hashtags like #BoycottAmazon, #AmazonFail, and #EconomicBlackout to align themselves with consumer frustration and anti-corporate narratives.

The largest spike in bot activity occurred on February 28, when the People’s Union’s Economic Blackout campaign called for a nationwide 24-hour boycott of non-essential purchases. Fake profiles mimicked activist voices calling for collective economic power. In the leadup to a second boycott day two months later, the pattern repeated, with another massive surge in fake activity on April 28.

In the picture: A fake profile encourages a boycott of Amazon and other retailers on February 28th, using the hashtag #EconomicBlackout.

Many of the fake profiles were meticulously designed to look like everyday consumers and community activists. They used slogans like “Power to the People,” posted protest graphics, and pushed time-bound calls to action. The repetition of language, coordinated timing, and rapid amplification across accounts all point to a concerted effort to shape perception and mobilize action.

While not every user supporting the boycott was fake, fake accounts significantly altered the scale, tone, longevity, and virality of the conversation.

Faking Frustration

While many inauthentic accounts leaned into the narrative of Amazon’s DEI rollback as the reason to boycott, an additional coordinated subset took a different approach: complaints about bad service.

Using hashtags like #AmazonFail, these bots flooded feeds with gripes about shipping delays, missing refunds, and product issues – many of them duplicated across multiple posts with similar tone and language. By repeating these themes en masse, bots created the illusion of overwhelming consumer dissatisfaction.

In the picture: An inauthentic account complains about Amazon’s customer service with the hashtag #AmazonFail.

And it worked. Cyabra’s analysis shows that on multiple days in June 2025, fake accounts were responsible for more posts about Amazon than authentic users. While the official boycotts had ended, the fake campaign continued, shifting the conversation from DEI to service quality, and keeping Amazon in the crosshairs.

When Manufactured Outrage Drives Market Impact

Amazon’s backlash was not entirely fake, but it was significantly fueled by inauthentic activity. What began as frustration over a corporate policy was escalated by bots, redirected toward broader complaints, and sustained well beyond the initial flashpoint.

This case illustrates how quickly a narrative can be hijacked and how difficult it is for companies to respond when fake voices dominate the conversation.

For brands navigating reputational risk in today’s digital landscape, distinguishing between real customers and synthetic outrage is critical. 

To respond effectively, companies must have clear visibility into what is being said, who is saying it, and how it is spreading. That means:

  • Monitoring sentiment and discourse patterns in real time to catch issues as they develop
  • Distinguishing between real customers and fake accounts, and understanding who is shaping perception
  • Knowing when to act, and what kind of message will resonate with authentic audiences

Cyabra’s new AI alerts offer early warnings about suspicious activity and coordinated amplification, so brands can stay ahead of the narrative before it spirals. Cyabra continuously scans billions of online interactions to identify risks across social platforms and news media 

Contact Cyabra to learn more about our new AI alerts and to see how Cyabra helps brands uncover inauthentic campaigns, detect bot networks, and respond with clarity – saving your team time and effort by bringing the insights directly to you. 

Download the full report by Cyabra

.

Related posts

Hair Care Brands Tangled Up in Fake News

The “No Shampoo” trend, also known as the “No-Poo Movement”, has caught social media like wildfire. Despite repeated warnings from experts regarding this trend, which...

Dark blue pump shampoo bottle with a prohibited symbol, surrounded by floating social media icons, hearts, and #NoPoo and #NoShampoo speech bubbles, symbolizing viral anti-shampoo narratives online

Rotem Baruchin

April 1, 2024

How a Fortune 500 Brand Stopped a Disinformation Crisis

A blockbuster launch spiraled into chaos – viral hashtags, fake claims, and mass backlash. Cyabra revealed 76% came from fake profiles – helping the corporate...

Rotem Baruchin

April 7, 2025

Ozempic Attacked: Fake Profiles, Misinformation, and Illegal Sales

Cyabra uncovered two types of attacks directed at Ozempic: fake profiles spreading fear and disinformation, and illegal sales of the prescription drug.

Humanoid robots in a neon-lit control room study social media feeds on multiple monitors, with an Ozempic semaglutide injection box and pen on the desk

Rotem Baruchin

August 1, 2024