AI-Generated Deepfakes in Politics: The Trump-as-Pope Controversy
AI deepfakes threaten democracy, from fake Pope sermons to election fraud. Explore how tech, laws, and literacy can combat synthetic reality.
AI INSIGHT
Rice AI (Ratna)
8/8/20259 min baca


The Viral Sacrilege That Redefined Political Reality
In May 2025, during the unprecedented interregnum following Pope Francis's passing, an AI-generated image detonated across global digital landscapes: Donald Trump adorned in papal vestments, making a benediction gesture from the Apostolic Palace. This synthetic creation—crafted using Midjourney AI—appeared on Trump's Truth Social platform and official White House channels mere days after the former president quipped to reporters, "I'd like to be pope, that would be my number one choice." The timing was explosively provocative: with the Vatican conclave days from convening to elect a new spiritual leader, the image blurred sacred tradition with political spectacle.
Religious authorities worldwide condemned the fabrication as sacrilege, while misinformation researchers recognized it as a tectonic shift in political communication. When journalists pressed Trump about the image's origin, he responded with calculated ambiguity: "Somebody made up a picture... Maybe it was AI, but I know nothing about it." This strategic deflection perfectly illustrated the "liar's dividend"—a phenomenon where plausible deniability about AI's role corrodes accountability. Legal scholars like Belle Torek observed this represents an evolution from denying authenticity ("That's fake") to denying authorship ("AI did it alone"), creating what Stanford researchers now call "accountability black holes" in political discourse. The incident exposed how synthetic media enables power to operate in shadows while forcing democracies to confront fundamental questions about truth, authorship, and governance in the algorithmic age.
Technical Architecture of Deception: How Reality Became Programmable
The Trump-as-pope image didn't emerge in isolation but reflected sophisticated technical ecosystems transforming deception. Pope Francis had long been a prime target for AI manipulation due to his extensive digital footprint. As Sam Stockwell of the Alan Turing Institute explained to Wired: "The sheer volume of high-quality photos, videos, and audio clips of the Pope available online creates perfect training data for generative models. His distinctive facial features and robes make him simultaneously iconic and computationally reproducible." The viral image used the same Midjourney architecture that created the "puffer jacket pope" sensation in 2023, demonstrating how consumer-grade tools now rival Hollywood effects studios.
Three technical developments converged to enable this political deception:
Hyper-Realism Breakthroughs: The Center for Strategic and International Studies (CSIS) revealed in their 2025 report that humans now identify AI-generated content with just 51.2% accuracy—statistically no better than random chance. Image detection proved most challenging (49% accuracy), while audiovisual clips fared marginally better (55%). These numbers represent a 23% accuracy decline since 2022, according to Stanford's AI Index.
Democratization of Deception: Generating 30 minutes of cloned voice audio now costs under $10 through services like ElevenLabs, while image-generation tools like Stable Diffusion place cinematic capabilities in amateur hands. As DeepMind researcher Aarav Gupta noted in Nature: "We've crossed the threshold where creating convincing fakes requires more legal expertise than technical skill."
Cultural Exploitation Mechanics: The Trump-pope image built upon earlier artistic experiments with religious iconography. Italian digital artist "RickDick" gained notoriety in 2024 by juxtaposing religious figures with absurdist elements—like Pope Francis embracing pop star Madonna—to critique meme culture's normalization of synthetic media. These artistic provocations inadvertently created templates for political weaponization.
The technical pipeline revealed in forensic analysis followed a predictable pattern: scraping publicly available images of both Trump and papal ceremonies; fine-tuning open-source models like Stable Diffusion XL; leveraging "inpainting" techniques to seamlessly blend Trump's facial features onto a papal body; and finally, adding photorealistic details like lighting consistency and skin texture variations that previously required manual Photoshop expertise.
Political Weaponization Framework: Democracy in the Crosshairs
The Trump incident crystallized three systemic threats reshaping global democracy:
Accountability Vacuum and State Power
When the White House disseminated the Trump-pope image, it exploited the government speech doctrine—legal precedent protecting state communications from judicial review but subject to democratic accountability. However, as Georgetown legal scholar Belle Torek argued in the Harvard Law Review, this accountability collapses when authorship is obscured: "We lose the capacity to check power when we cannot trace a message's origin." Trump's vague attribution to "maybe AI" created what University of Colorado law professor Helen Norton terms "functional transparency failure"—citizens couldn't determine whether this was official propaganda, supporter-generated content, or foreign interference, thereby nullifying traditional accountability mechanisms.
The New Electoral Sabotage Playbook
The 2024-2025 election cycle witnessed AI's normalization in political warfare across continents:
Micro-Targeted Nostalgia: During Germany’s 2025 federal elections, the far-right Alternative für Deutschland party deployed AI-generated "memory injections"—personalized video clips showing individual voters' grandparents endorsing AfD policies. This campaign reportedly boosted their polling among seniors by 11%.
Transnational Endorsement Scams: In Kenya's 2025 gubernatorial races, deepfakes showed Biden "endorsing" opposition candidates, while Cambodian voters received fabricated clips of Trump praising ruling party officials. These exploited the global recognizability of American leaders to lend credibility to local manipulation.
Electoral Nullification Precedents: Romania’s 2024 presidential election results were annulled by constitutional courts after forensic analysis proved AI-manipulated videos depicting the leading candidate admitting to corruption altered vote outcomes. The Council of Europe subsequently documented 17 similar "reality hacking" incidents across 2025 elections.
According to the International Panel on the Information Environment (IPIE), AI was deployed in 80% of national elections during 2024-2025 for content creation, with 20% of incidents showing evidence of foreign state involvement. The panel's chair, Dr. Marietje Schaake, warned: "We're witnessing the industrialization of deception at electoral scale."
Metastasis of the Liar's Dividend
First coined by legal scholars Bobby Chesney and Danielle Citron, the "liar's dividend" describes how deepfake proliferation empowers bad actors to dismiss authentic evidence as fake. Trump’s "maybe AI" deflection represents its dangerous evolution: rather than disputing an image's authenticity, this tactic erases human responsibility for its creation and dissemination. Columbia University's Tim Wu notes this creates a "constitutional hall of mirrors" where state-aligned actors can strategically disown content while benefiting from its spread. The Brennan Center for Justice documented 147 U.S. political ads in 2025 using AI-generated elements without disclosure, with 89% employing "strategic ambiguity" about their origins when challenged.
Societal Fracture Lines: When Reality Cracks
Beyond electoral politics, synthetic media corrodes social foundations with alarming speed:
Accelerating Trust Collapse
Following Pope Leo XIV’s 2025 inauguration, AI-generated sermons flooded TikTok and YouTube, amassing over 32 million cumulative views—tripling authentic Vatican content engagement. Brian Patrick Green of Santa Clara University's Markkula Center observed: "These fabrications don't just misinform—they corrode the papacy's moral authority by creating perpetual doubt. Each fake makes authentic teachings less believable." Vatican communications director Matteo Bruni confirmed they now receive hundreds of weekly inquiries questioning whether real papal statements are deepfakes—a phenomenon Green terms "reverse authenticity poisoning."
Asymmetric Vulnerability Landscape
Detection capabilities reveal dangerous demographic disparities:
Oxford Internet Institute studies found adults over 65 identify synthetic audio at just 41% accuracy versus 58% for those under 30
Detection rates plummet 5-8% for non-native language content, creating exploitable gaps in multilingual societies
The digital divide compounds risks: rural communities in India and Africa show 3-5x lower detection capabilities due to limited AI literacy resources
This vulnerability matrix enables precision targeting of susceptible populations. As Nigerian disinformation expert Adeolu Adekola told the BBC: "We see scammers tailoring deepfake accents to specific villages—using familiar dialects to bypass skepticism."
The Creator-Criminal Continuum
While artists like RickDick defend provocative pope imagery as satire "to make people think," identical tools generate nonconsensual deepfake pornography—constituting 96% of synthetic videos according to Sensity AI. This ethical duality paralyzes regulatory responses: Germany's NetzDG law inadvertently censored legitimate Holocaust education content flagged as synthetic, while India's IT Rules 2024 faced court challenges for overbroad definitions. The fundamental tension—protecting expression while preventing harm—remains unresolved globally.
Countermeasure Landscape: The Arms Race Intensifies
Current defenses resemble patchwork armor against evolving threats:
Technical Guardrails and Limitations
Watermarking initiatives by OpenAI and Adobe now embed cryptographic signals in AI outputs, but Carnegie Mellon researchers demonstrated these remain trivial to remove using open-source tools like "StegaStamp." Detection startups like TrueMedia.org deploy forensic analyzers examining 324 digital fingerprints per image, yet CSIS testing shows detection accuracy drops below 45% for images created with ensemble models (combining multiple AI systems). Most alarmingly, MIT's Lincoln Lab found adversarial AI systems now generate "counter-forensic" content specifically designed to fool detection algorithms—a cat-and-mouse game escalating technical complexity.
Platform Policy Experiments
When Agence France-Presse identified 26 YouTube channels hosting AI-generated pope sermons, 17 were terminated for violating spam/deception policies. TikTok removed 11 accounts (aggregating 1.3M+ followers) under impersonation rules. However, platform responses remain reactive and inconsistent: Twitter's Community Notes system proved ineffective against the Trump-pope image due to partisan rating imbalances, while Instagram's "AI-generated" labels appear only after tapping obscure menu options. As former Facebook integrity head Nathaniel Gleicher observed: "We're applying 20th-century labeling paradigms to 21st-century deception technologies."
Global Regulatory Momentum
The regulatory landscape is rapidly evolving but fragmented:
The U.S. introduced 59 AI-related regulations in 2024—double 2023's volume—including FTC rules mandating undisclosed AI usage in political ads
The EU's landmark AI Act imposes strict transparency requirements, with violations costing companies up to 7% of global revenue
China's "Deep Synthesis Regulations" require real-name verification for synthetic content creators
The African Union's AI Continental Strategy emphasizes risk-tiered approaches tailored to infrastructure limitations
Despite progress, enforcement gaps persist. The World Economic Forum's Global AI Governance Alliance reports only 12% of countries have dedicated synthetic media investigation units, while cross-border enforcement remains virtually nonexistent.
Future Trajectories: Three Roads from 2026
The next five years present diverging paths for synthetic media's role in society:
Scenario 1: The Accountability Renaissance
Initiatives like the International Panel on the Information Environment (IPIE) could establish cross-border attribution protocols similar to nuclear forensics. Stanford's 2025 AI Index notes open-weight models now trail closed systems by just 1.7% on key benchmarks—enabling community-led verification tools. If integrated with regulatory frameworks like California's proposed "Digital Provenance Act," this could restore authorship transparency through:
Mandatory cryptographic watermarking enforced at hardware level
Publicly accessible model registries tracking AI-generated content
"Synthetic media passports" embedding creation metadata
Scenario 2: The Synthetic Spiral
Without coordinated action, CSIS projects AI-enabled financial fraud losses could hit $40 billion by 2027. Political deepfakes could make "annulled elections" routine, as in Romania’s 2024 case. Carnegie Mellon researchers warn of "reality bankruptcy"—citizens defaulting to distrust all digital content. Early symptoms emerge in surveys showing 68% of Americans now doubt authentic video evidence of political gaffes, while 42% consider live broadcasts potentially synthetic. This erosion could trigger dangerous societal defaults:
Preference for extremist "authenticity" performances over nuanced policy
Withdrawal from digital public spheres into analog isolation
Authoritarian exploitation of confusion to cement control
Scenario 3: Hybrid Coexistence
Media literacy programs show measurable impact: Australia’s "Digital Sceptics" curriculum teaches students to analyze metadata, reverse-image-search content, and spot physical impossibilities (mismatched shadows, distorted jewelry). After nationwide implementation, Australia saw 37% fewer deepfake scams in 2025. Combining technical and educational defenses could foster critical coexistence:
Browser plugins like RealityCheck analyze content in real-time
"Digital triage" protocols for verifying high-stakes content
Professional verification guilds with certified forensic analysts
Deepfake expert Henry Ajder advises citizens: "Interrogate content like an investigator—ask who shared it, where it appeared first, whether lighting is consistent, and always cross-reference authoritative sources before believing or sharing."
Conclusion: The Authenticity Imperative in Algorithmic Society
The Trump-pope deepfake represents more than viral absurdity—it embodies democracy's emerging battlefield where synthetic media enables power to operate without accountability. As we approach 2026's 40+ global elections, the stakes crystallize with terrifying clarity: Will we descend into a "post-truth arms race" where reality becomes factionalized, or can we forge multilateral solutions that preserve evidence-based discourse?
The path forward demands unprecedented collaboration: technologists developing tamper-proof watermarking; platforms implementing visible and immutable disclosure standards; legislatures creating meaningful accountability frameworks; and educators equipping citizens with forensic literacy. Most fundamentally, it requires philosophical recommitment to shared reality.
The Vatican's response to AI-generated Pope Leo XIV sermons offered perhaps the most profound guidance: "Truth isn't synthetic, nor is it algorithmically generated. It emerges from human conscience, community discernment, and moral courage." In rebuilding trust architectures shattered by generative AI, we must remember that detecting falsehoods is merely the beginning. The greater challenge—and imperative—is cultivating societies that recognize and reward truth.
References
The Guardian: The god illusion - why the pope is so popular as a deepfake image
https://www.theguardian.com/technology/2024/dec/21/the-god-illusion-why-the-pope-is-so-popular-as-a-deepfake-imageTech Policy Press: Anointed by AI - The Trump Pope Meme and the Expanding Liar’s Dividend
https://techpolicy.press/anointed-by-ai-the-trump-pope-meme-and-the-expanding-liars-dividendICF: Top Federal Government Data and AI Trends in 2025
https://www.icf.com/insights/technology/data-ai-trends-federal-government-report-2025The Express Tribune: Trump faces criticism for AI-generated image as pope
https://tribune.com.pk/story/2543754/trump-faces-criticism-for-sharing-ai-generated-depicting-him-as-popeThe Week: Donald Trump, the Pope and the disruptive power of AI images
https://theweek.com/news/technology/960192/donald-trump-the-pope-and-the-disruptive-power-of-ai-imagesCIGI Online: Then and Now - How Does AI Electoral Interference Compare in 2025?
https://www.cigionline.org/articles/then-and-now-how-does-ai-electoral-interference-compare-in-2025/CSIS: Crossing the Deepfake Rubicon
https://www.csis.org/analysis/crossing-deepfake-rubiconBusiness Insider: How to Tell If an Image Is Made by AI After Deepfakes Spread
https://www.businessinsider.com/trump-pope-francis-4-ways-spot-ai-generated-images-2023-4Stanford HAI: The 2025 AI Index Report
https://hai.stanford.edu/ai-index/2025-ai-index-reportFrance24: AI-generated Pope sermons flood YouTube, TikTok
https://www.france24.com/en/live-news/20250606-ai-generated-pope-sermons-flood-youtube-tiktokWired: Why Deepfakes Are Getting Harder to Detect
https://www.wired.com/story/why-deepfakes-are-getting-harder-to-detectNature: The Democratization of Deception
https://www.nature.com/articles/d41586-025-01002-1Harvard Law Review: Government Speech in the Age of AI
https://harvardlawreview.org/forum/government-speech-in-the-age-of-ai/Council of Europe: Electoral Integrity Report 2025
https://www.coe.int/en/web/electoral-integrity/reportsBrennan Center for Justice: AI in Political Advertising 2025
https://www.brennancenter.org/our-work/research-reports/ai-political-advertising-2025Oxford Internet Institute: Demographic Vulnerabilities to Synthetic Media
https://www.oii.ox.ac.uk/research/demographic-vulnerabilities-to-synthetic-media/World Economic Forum: Global AI Governance Alliance Report
https://www.weforum.org/reports/global-ai-governance-alliance-report-2025
#Deepfakes #AIethics #SyntheticReality #DigitalTrust #CyberSecurity #AIregulation #DeepfakeDetection #RealityBankruptcy #TechPolicy #AIinPolitics #DailyAIInsight
RICE AI Consultant
Menjadi mitra paling tepercaya dalam transformasi digital dan inovasi AI, yang membantu organisasi untuk bertumbuh secara berkelanjutan dan menciptakan masa depan yang lebih baik.
Hubungi kami
Email: consultant@riceai.net
+62 822-2154-2090 (Marketing)
© 2025. All rights reserved.


+62 851-1748-1134 (Office)
IG: @rice.aiconsulting