Close Menu
AI News TodayAI News Today

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    AI traffic to US retailers rose 393% in Q1, and it’s boosting their revenue too

    Teenage Engineering KO-Amp 35 leak suggests it’s targeting guitars next

    InsightFinder raises $15M to help companies figure out where AI agents go wrong

    Facebook X (Twitter) Instagram
    • About Us
    • Contact Us
    Facebook X (Twitter) Instagram Pinterest Vimeo
    AI News TodayAI News Today
    • Home
    • Shop
    • AI News
    • AI Reviews
    • AI Tools
    • AI Tutorials
    • Chatbots
    • Free AI Tools
    AI News TodayAI News Today
    Home»AI Reviews»How the Internet Broke Everyone’s Bullshit Detectors
    AI Reviews

    How the Internet Broke Everyone’s Bullshit Detectors

    By No Comments4 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    How the Internet Broke Everyone’s Bullshit Detectors
    Share
    Facebook Twitter LinkedIn Pinterest Email

    Lego-style propaganda videos alleging war crimes are flooding online feeds, echoing the White House’s own turn toward cryptic teaser clips and meme-native visuals. This is not just content drift. It is a new front in the information war, one where speed, ambiguity, and algorithmic reach matter as much as accuracy.

    One Iran-linked outlet, Explosive News, can reportedly turn around a two-minute synthetic Lego segment in about 24 hours. The speed is the point. Synthetic media does not need to hold up forever; it only needs to travel before verification catches up.

    Last month, the White House added to that confusion when it posted two vague “launching soon” videos, then removed them after online investigators and open source researchers began dissecting them.

    The reveal turned out to be anticlimactic: a promotional push for the official White House app. But the episode demonstrated how thoroughly official communication has absorbed the aesthetics of leaks, virality, and platform-native intrigue. Even when official accounts adopt the aesthetics of a leak, questioning whether a record is real or synthetic is the only defensive move left.

    Real vs. Synthetic: The New Friction

    A zero digital footprint used to signal authenticity. Now, it can signal the opposite. The absence of a trail no longer means something is original—it may mean it was never captured by a lens at all. The signal has inverted. Truth lags; engagement leads.

    Automated traffic now commands an estimated 51 percent of internet activity, scaling eight times faster than human traffic according to the 2026 State of AI Traffic & Cyberthreat Benchmark Report. These systems don’t just distribute content, they prioritize low-quality virality, ensuring the synthetic record travels while verification is still catching up.

    Open source investigators are still holding the line, but they are fighting a volume war. The rise of hyperactive “super sharers,” often backed by paid verification, adds a layer of false authority that traditional open source intelligence (OSINT) now has to navigate.

    “We’re perpetually catching up to someone pressing repost without a second thought,” says Maryam Ishani, an OSINT journalist covering the conflict. “The algorithm prioritizes that reflex, and our information is always going to be one step behind.”

    At the same time, the surge of war-monitoring accounts is beginning to interfere with reporting itself. Manisha Ganguly, visual forensics lead at The Guardian and an OSINT specialist investigating war crimes, points to the false certainty created by the flood of aggregated content on Telegram and X.

    “Open source verification starts to create false certainty when it stops being a method of inquiry—through confirmation bias, or when OSINT is used to cosmetically validate official accounts or knowingly misapplied to align with ideological narratives rather than interrogate them,” Ganguly says.

    While this plays out, the verification toolkit itself is becoming harder to access. On April 4, Planet Labs—one of the most relied-upon commercial satellite providers for conflict journalism—announced it would indefinitely withhold imagery of Iran and the broader Middle East conflict zone, retroactive to March 9, following a request from the US government.

    The response from US defense secretary Pete Hegseth to concerns about the delay was unambiguous: “Open source is not the place to determine what did or did not happen.”

    That shift matters. When access to primary visual evidence is restricted, the ability to independently verify events narrows. And in that narrowing gap, something else expands: Generative AI doesn’t just fill the silence—it competes to define what’s seen in the first place.

    Generative AI Is Getting Harder to Spot

    Generative AI platforms have been learning from their mistakes. Henk van Ess, an investigative trainer and verification specialist, says many of the classic tells—incorrect finger counts, garbled protest signs, distorted text—have largely been fixed in the latest generation of models. Tools like Imagen 3, Midjourney, and Dall·E have improved in prompt understanding, photorealism, and text-in-image rendering.

    But the harder problem is what van Ess calls the hybrid.

    broke Bullshit Detectors Everyones Internet
    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous Article‘Euphoria’ Season 3: Release Date and Time on HBO Max
    Next Article Today’s NYT Strands Hints, Answer and Help for April 11 #769
    • Website

    Related Posts

    AI Reviews

    InsightFinder raises $15M to help companies figure out where AI agents go wrong

    AI Reviews

    Today’s NYT Mini Crossword Answers for April 17

    AI Reviews

    Live Nation says it will fight monopoly suit loss

    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    AI traffic to US retailers rose 393% in Q1, and it’s boosting their revenue too

    0 Views

    Teenage Engineering KO-Amp 35 leak suggests it’s targeting guitars next

    0 Views

    InsightFinder raises $15M to help companies figure out where AI agents go wrong

    0 Views
    Stay In Touch
    • Facebook
    • YouTube
    • TikTok
    • WhatsApp
    • Twitter
    • Instagram
    Latest Reviews
    AI Tutorials

    Quantization from the ground up

    AI Tools

    David Sacks is done as AI czar — here’s what he’s doing instead

    AI Reviews

    Judge sides with Anthropic to temporarily block the Pentagon’s ban

    Subscribe to Updates

    Get the latest tech news from FooBar about tech, design and biz.

    Most Popular

    AI traffic to US retailers rose 393% in Q1, and it’s boosting their revenue too

    0 Views

    Teenage Engineering KO-Amp 35 leak suggests it’s targeting guitars next

    0 Views

    InsightFinder raises $15M to help companies figure out where AI agents go wrong

    0 Views
    Our Picks

    Quantization from the ground up

    David Sacks is done as AI czar — here’s what he’s doing instead

    Judge sides with Anthropic to temporarily block the Pentagon’s ban

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    Facebook X (Twitter) Instagram Pinterest
    • About Us
    • Contact Us
    • Terms & Conditions
    • Privacy Policy
    • Disclaimer

    © 2026 ainewstoday.co. All rights reserved. Designed by DD.

    Type above and press Enter to search. Press Esc to cancel.