How Nigeria-Based Operators Infiltrated Authentic US Conversations

Here’s What Cyabra Uncovered

In November 2025, Cyabra analyzed social media conversations about US-Nigeria relations across Facebook, X, Instagram, and TikTok. The findings were alarming: 2,627 profiles, which were 25% of the accounts in the conversation, were fake accounts participating in a coordinated influence operation. These inauthentic profiles generated over 13,600 engagements and reached an estimated 5.2 million potential viewers.

Unlike typical influence operations that operate in isolation, these fake profiles strategically integrated within authentic discussion clusters and engaged directly with real users. This direct engagement generated over 13,600 engagements and reached an estimated 5.2 million potential viewers, demonstrating sophistication closer to tactics previously associated with state actors like Iran and Russia.

At a glance: Cyabra’s analysis

Infiltration Strategy

Cyabra’s analysis identified synchronized posting within identical timeframes, with visualization data showing clear bursts of activity by fake profiles operating during the same time periods, a hallmark of centrally coordinated influence campaigns.

The operation exhibited distinct, synchronized activity patterns from November 1 to 26, with posting times carefully aligned to match the authentic users’ engagement rhythms. This synchronization helped fake profiles blend seamlessly into ongoing discussions. The operation was particularly active on X, where it had the most significant impact.

Fake profiles (red) infiltrated authentic communities and discussions (green)

Beyond the Echo Chamber

What sets this operation apart is its approach to infiltrating authentic conversations. Rather than operating in isolated echo chambers, these fake profiles positioned themselves within genuine discussion clusters.

Cyabra’s analysis identified several coordinated behaviors indicating a centrally guided operation:

  • Non-human and AI-generated profile images
  • Repetition of similar messaging patterns
  • Synchronized activity windows aligned with authentic users
  • Emotionally charged language designed to provoke reactions

While the operation originated from Nigeria, many accounts operated from other countries, including the United Kingdom. This demonstrates a multi-source fake network employing VPNs and distributed assets, tactics commonly associated with state-backed campaigns.

Fake profiles spread the same narrative using sophisticated, unique messages.

Engineered Narratives and Their Impact

The fake profiles consistently amplified two central narratives:

  • Urgent calls for US intervention in Nigeria, framed around Christian persecution
  • Broad appeals to “support Nigeria” reinforcing the intervention narrative

These narratives created the appearance of widespread Nigerian support for US involvement. The most significant impact occurred when authentic users amplified content from fake profiles. In one example, a fake profile named “KelzOnyenkwere” posted a message urging former President Donald Trump to “save Nigeria from genocide.” While this post initially received modest engagement (350 views, 6 likes, 3 shares), it was reshared by Deacon Nick Donnelly, an authentic account with over 74,000 followers, thereby exponentially increasing the reach of the manipulated narrative.

This pattern of authentic amplification created the false impression that large numbers of Nigerian citizens were urgently calling for US intervention.

Implications for National Security

This influence operation poses significant challenges for government agencies. By infiltrating authentic conversations and spreading the campaign’s content through authentic influencers who unknowingly echoed and shared fake profiles, these manipulators created a sophisticated threat to public perception.

A comparison with other campaigns highlights these advanced tactics. While operations like China’s efforts targeting Taiwan rely on clusters of fake accounts interacting primarily with one another, the Nigeria operation’s strategic infiltration into real user networks represents an evolution in influence operation tactics. The manufactured appearance of grassroots consensus poses real risks to informed policy-making around US-Nigeria relations.

Fake profiles were most active in specific times of the day, and implemented heavy synchronized posting

Detecting Embedded Influence Operations

For government agencies, this operation highlights the need for enhanced detection capabilities that go beyond content analysis to identify critical network behaviors. Key protections should include:

  • Network analysis to identify fake profiles embedded within authentic conversations
  • Behavioral pattern recognition to detect synchronized posting
  • Cross-platform monitoring to track coordinated campaigns across social media

Cyabra’s AI-powered platform exemplifies this approach by analyzing posts, conversations, and cross-platform behaviors to categorize content by narrative, sentiment, authenticity, and coordinated activity, enabling precise detection of influence operations before they significantly impact public discourse.

As these operations evolve, government agencies must adapt their strategies. This Nigerian campaign serves as a critical case study for understanding how influence operations continue to evolve and why detection methodologies must similarly advance to protect the integrity of public discourse. Contact Cyabra to learn more.

Test Yourself: Identifying Influence Operations Tactics

What are some of the key indicators of a fake campaign?

* Coordinated posting patterns – Synchronized activity bursts occurring within identical timeframes
* Visual anomalies in profile images – using images of pets, landscapes, objects, or AI-generated faces
* Behavioral inconsistencies – Repetition of similar messaging and keyword patterns across the network
* Network structure analysis – Mapping how fake profiles positioned themselves within authentic conversation clusters

How should government agencies protect authentic public discourse?

Network behavior analysis – Map relationships between accounts to identify infiltration patterns
Cross-platform monitoring – Track coordinated activity across multiple platforms
Enhanced public-private partnerships – Improve coordination between intelligence agencies and platforms
Real-time detection systems – Deploy monitoring tools that identify suspicious activity as it emerges

How do modern influence operations differ from traditional “echo chamber” campaigns?


Infiltration of authentic clusters – Rather than operating in isolation, fake accounts strategically integrate into genuine discussion groups to engage real users directly.
Authentic amplification – The goal is to prompt high-follower, authentic accounts to reshare manipulated content, giving it unearned credibility and exponentially increasing its reach.
Sophisticated personas – Using GenAI to create consistent, high-realism backstories, including profile images and long social histories, to avoid easy detection.

Why is the “Nigeria-based” operation considered an evolution in threat tactics?

Distributed assets – While originating in Nigeria, the network used VPNs to operate from multiple countries, including the UK, to mask its source.
Mimicking state actors – The coordination and scale of the 2,627 fake accounts demonstrated a level of sophistication typically reserved for state-backed entities like Russia or Iran.
Engineered grassroots consensus – By flooding narratives with “urgent calls for intervention,” the operation created a false impression of widespread local support for specific US foreign policies.

Related posts

“Made in China” or by Bots? Fake Profiles Fuel Tariff Debate

Luxury fashion brands like Lululemon, Chanel, and Gucci were slammed following US tariffs discussions, accused for "misleading consumers about where their products are manufactured". Cyabra...

Rows of glowing-eyed robots at laptops generate fake social media posts urging a boycott of luxury brands during the #TariffWar and #MadeInChina debate, symbolizing automated disinformation

Rotem Baruchin

May 26, 2025

Nepal’s Protests: Authentic Voices, Fake Calls for Violence

The September 2025 Nepal protests were an authentic movement led by Gen Z and fueled by long-standing frustrations. But Cyabra’s research uncovered how fake profiles...

Rotem Baruchin

September 15, 2025

Cyabra at Golin: Understanding What’s Real vs. What’s Fake

Jonny Bentwood, Global President of Data and Analytics at PR agency Golin, explains how Cyabra helps protect clients’ reputations.

Man with glasses speaking in a blurred office environment, alongside a Cyabra-branded quote that reads “To protect the reputation of brands, understanding what is fake and what is real is critical”

Rotem Baruchin

February 11, 2025