
Reports & Papers
Documentary at Risk: Strategies for Ethics, Sustainability, and Innovation in a Time of Disruption
Reports & Papers
The views expressed in Shorenstein Center Discussion Papers are those of the author(s) and do not necessarily reflect those of Harvard Kennedy School, Harvard University or the organizations and institutions with whom the authors are affiliated.
Discussion Papers have not undergone formal review and approval. Such papers are included in this series to elicit feedback and to encourage debate on important issues and challenges in media, politics and public policy. Copyright belongs to the author(s). A PDF of this paper is available for download here for personal use only.
The machines aren’t coming—they’re here, reshaping our world with unprecedented power while carrying persistent flaws. As artificial intelligence floods into more aspects of daily life, we face a critical choice: will we let generative AI accelerate social decay and inequality, or will we harness these powerful but imperfect tools to help solve humanity’s greatest challenges?
Every day, this technology’s impact grows more visible and consequential. Hobby Lobby now casually sells AI-generated artwork with bizarrely missing details, displacing human artists without hesitation1. Meta populates Facebook and Instagram with AI-generated photos and even AI profiles with fabricated life stories2. Digital commerce uses your data to personalize prices without your knowledge3. What connects these developments is a sophisticated machinery of reality distortion that affects our communities in increasingly pervasive ways.
This analysis examines three interconnected crises imposing mounting burdens on American society:
At the heart of these crises lies a simple truth: too much power concentrated in too few hands. This power imbalance lets corporations control our digital landscape, economic reality, and physical environment with minimal accountability or oversight.
For citizens concerned about AI’s impact on their communities, it shows how to get involved effectively. For legislators wrestling with these transformative technologies, it outlines practical solutions that are pro-freedom, pro-consumer and creator, and anti-pollution.
Key Themes
Importantly, while this analysis raises serious concerns about Big Tech’s consolidated power, many solutions actually align with Silicon Valley’s self-interest: reduced energy costs, improved infrastructure, and reliable human-curated content serve both corporate profits and the public good. This alignment could make positive change achievable. Life can be good for all.
The stakes could not be higher. AI already affects many aspects of daily life—increasingly coming from behind the curtain. Without immediate action, these effects will only intensify. But with smart intervention, we can protect human agency and community prosperity while capturing AI’s benefits—and break the machinery of reality distortion before it fully takes hold.
Recent advances in AI capabilities raise the specter of recursive self-improvement, where systems become capable of enhancing themselves. While multiple AI labs express growing confidence in their ability to achieve artificial general intelligence (AGI) within years rather than decades, they’ve failed to articulate a clear vision of what that world would actually look like for ordinary citizens. Instead, they’re placing the burden on all of society to bear the costs of both concrete harms (hallucinations, bias, pollution) and abstract risks (existential threats, job displacement, social disruption) and policymakers to determine how to govern technologies that could fundamentally reshape human society—without providing a roadmap for what that reshaped society might be.
In the meantime, we’re forced to contend with AI’s persistent reality distortion. Facebook Marketplace users suspend critical thinking and trust AI-generated descriptions over contradictory evidence from major retailers. NotebookLM, among Google’s flagship AI products, butchered this very analysis on March 11, 2025—inventing policy recommendations I never made while transforming my horse manure crisis analogy (about utilities getting grid planning wrong at consumers’ expense) into generic technophobia. If even sophisticated AI systems so confidently misinterpret straightforward metaphors while projecting false authority, how can we build the shared understanding necessary to address our mounting challenges? This analysis outlines a way forward.
Just as the Industrial Revolution’s ultimate success required labor laws and environmental protections, AI’s promise—our new Industrial Revolution—can only be fully realized with thoughtful frameworks that preserve human agency while fostering innovation. Recent forecasts from leading AI researchers have dramatically compressed our timeline: superhuman capabilities in specific domains are anticipated to within 2-3 years, not decades, with predictions it will directly impact working families.
The convergence of social media’s attention economy with increasingly powerful AI tools represents a critical inflection point for American society. Our digital environments—already structured to reward emotional manipulation and tribal outrage—are now being supercharged with technologies capable of generating unlimited, personalized content designed to maximize engagement regardless of societal cost. At Meta, this convergence is playing out in particularly concerning ways: former executives warn of unprecedented new vectors for misinformation as the company dismantles internal safeguards against harmful characterizations of vulnerable groups while simultaneously developing more sophisticated AI systems that can generate and amplify such content at unprecedented scale4.
The real-world consequences of this convergence are everywhere around us, extending far beyond our screens. Research has consistently linked online disinformation to offline violence5. This dynamic creates feedback loops where digital pollution drives real-world harm, which then generates more inflammatory content online. We see this cycle play out in local communities where synthetic news—designed to provoke rather than inform—shapes civic decisions, polarizes neighbors, and erodes the shared factual foundation necessary for problem-solving. When we can’t agree on basic facts, we start failing to solve our problems.
Sadly, our media too often fixates on political optics, ignoring real-world consequences. Polling or what politicians say dominates, rather than explaining what’s illegal or untruthful. Take environmental coverage: despite pollution’s massive impact on our health and economy, stories rarely connect political decisions to their human costs. This style of coverage—treating life-and-death policy decisions as mere political theater—leaves Americans struggling to understand how politicians actually affect their lives, their children’s futures, and their communities’ wellbeing.
We’re living in what The Atlantic’s Derek Thompson calls “the anti-social century.” Self-imposed solitude might just be the most important social fact of the 21st century in America—an unprecedented rise in time spent alone and at home. This isolation warps our politics and personalities, making it harder to understand others’ realities or even recognize our shared challenges. When people primarily interact through screens rather than face-to-face, we lose the basic social fabric that helps communities solve problems together.6
Even Google search, long the gold standard for accessing human knowledge, now prioritizes AI-generated content that lacks factual verification, crowding out human-created, expert-verified sources. Approximately half of American adults struggle with reading comprehension,7 making them particularly vulnerable to sophisticated AI-generated content, as web searches now deliver wrong answers again.
Most alarming, the massive data centers powering these AI systems force utilities to keep coal plants running—an expensive choice that burdens ratepayers with both higher costs and increased pollution, literally making everyday people pay more money to burn the dirtiest fossil fuels to generate digital pollution and misinformation.8
We’re learning their current tendency to ‘hallucinate’ is evolving into something more concerning: strategic deception.9 What happens when vast machine intelligence combines with persistent unreliability and risk of deception? We’re not adequately preparing for a world awash in artificial intelligence—both its remarkable capabilities and its deep-seated flaws.10 These aren’t just technical challenges for AI developers to solve. They demand urgent attention from policymakers, business leaders, and citizens who must navigate this transformation. How we prepare for and shape it will determine whether it becomes a force for progress, and how that progress will be distributed across society—or disruption. We can’t wait until the water starts rising to have these crucial conversations.
This crisis reflects a fundamental shift in how online information operates. As information literacy expert Mike Caulfield and journalist Charlie Warzel observe, today’s internet functions less as a brainwashing engine and more as a “justification machine’ where ‘a rationale is always just a scroll or click away.”11 The incentives of the modern attention economy—where engagement and influence are rewarded—ensure there will always be a rush to provide such rationales. When confronted with evidence that might shake their beliefs, users can instantly seek validation from conspiracist feeds, decontextualized videos, and now AI systems—all eager to confirm existing worldviews. Users aren’t seeking truth but confirming existing beliefs, making it increasingly easy to maintain those beliefs even in the face of contradicting evidence.
Something fascinating is happening to our collective minds. Social media isn’t just changing what we consume—it’s reshaping how our brains process information. The science of neuroplasticity shows our brains are adapting to our constant exposure to quick-hit digital stimulation, reshaping our neural pathways until sustained focus feels as unnatural as reading a book on a roller coaster. We’re not getting dumber—our brains are adapting to an environment that prizes constant distraction over deep reflection. This rewiring of our collective attention spans makes us particularly vulnerable to the justification machine’s endless scroll of convenient confirmations.12
While Americans’ trust in mass media plumbed new depths, recent catastrophic weather events—from Hurricane Helene to Los Angeles’s wildfires—showed how public media, public technology like Watch Duty, and select, often local outlets broke through the noise and institutional inertia, delivering timely reporting that married scientific precision with human narrative. Meanwhile social media platforms, once heralded as truth’s digital champions, devolved into algorithmic amplifiers13 of digital pollution: 14 misinformation, conspiracy theories and financial scams that leads to real-world offline victims.
Drawing on both Congressional experience and current research, this work offers practical solutions people from policymakers to philanthropists should consider that address root causes and can bridge political divides:
While some recommendations require legislative or regulatory action, many can be implemented by philanthropies, industry leaders, civic organizations, and community groups. Effective solutions often emerge from collaboration across sectors rather than government action alone.
Generative AI—with Large Language Models (LLMs) being the prime offender—produces authoritative-sounding falsehoods as fluently as facts . University of Glasgow researchers bluntly labeled these outputs ‘bullshit,’15 highlighting LLMs lack of true comprehension. These systems routinely hallucinate and become unreliable when faced with simple name changes or variable swaps,16 demonstrating fundamental flaws beneath their confident veneer.
In March 2025, the Los Angeles Times pulled down their AI-generated coverage bot after it defended the Ku Klux Klan.17 Google’s shift from human-curated ‘Snippets’ to ‘AI Overviews’ in 2024 turned queries about “warm-blooded reptiles” in science18 into AI-generated oversimplifications and prioritizes AI imitations over authentic works, replacing expert-verified information with plausible falsehoods.19
Tech companies want to dream about sci-fi AI futures, but they need to fix the mundane, dangerous problems—hallucinations, reliability failures, and rampant information pollution. It’s possible, even likely, that LLMs won’t ever be able to completely solve these problems without additional uninvented technology. Even the most sophisticated AI summarization tools consistently obliterate nuance, often misrepresenting complex arguments. Despite extensive testing with cutting-edge AI systems, these tools sometimes generate summaries that contradict key findings. This limitation drove the development of three carefully calibrated editions of this paper—each preserving essential context while accommodating different reading time constraints. While AI assisted in this process, its current limitations underscore why human oversight remains indispensable.
The 2025 Cybertruck bombing in Las Vegas—the first attack publicly linked to AI—where a decorated special forces Army veteran used ChatGPT to methodically plan his attack despite built-in safety protocols, harming seven innocent victims. While individual blame is complex, these incidents reveal that “move fast and break things” sometimes leaves behind broken lives and communities.
The stakes escalate as we move from today’s “narrow AI” systems, which perform specific, defined tasks, to autonomous agents that can independently pursue complex objectives through chains of decisions. We’re already seeing the future – corporate lawyers observe and warn: autonomous AI agents regularly make decisions and take actions their creators neither anticipated nor designed for.20
Until AIs tout their hallucination-free guarantee—and crucially, promise it can identify and correct flawed human inputs that cause errors too—we should approach every AI-generated insight with informed skepticism.
We need policies and technologies that empower digital self-determination in an age where we are all creators and consumers: ensuring people can consolidate content across platforms, choose how they discover and filter the content they view, and freely transfer their data, relationships, and creative works between whatever service they choose.
This framework resembles what digital governance scholars call ‘middleware’—intermediary layers between platforms and users that allow individuals to choose how their feeds are curated.21
Imagine choosing your own algorithm to filter content from all your social feeds—TikTok, Instagram, Twitter—based on what you define as time well spent, whether that’s viral videos, family updates, learning, creativity, or meaningful connection. In contrast, platform-controlled algorithms push hyperpartisan misinformation even into accounts following only sports.22 Today’s developers can build partial solutions through browser extensions, but platforms have methodically closed the access points needed for comprehensive alternatives to their algorithms. Open standards and interoperability requirements could enable API access, standardized data export, and relationship portability —technological solutions that may require regulatory support to overcome platform resistance .
These rights would ensure creators can maintain their content libraries, audience relationships, and analytics across platforms, preventing platform lock-in and mitigating concerns over censorship. With users controlling their own content aggregation and filtering, platforms are still hosts, but lose their power as gatekeepers.
Critics rightly note that many users won’t customize their feeds or may abandon tools that challenge their viewpoints—after all, Facebook’s passive consumption model defeated RSS readers precisely because it required less effort. But even if only a minority of users initially choose alternative algorithms or cross-bubble content curation, their existence creates market pressure for better options, just like organic and nutrition transparency transformed grocery stores without universal adoption. On-device AI models could make such tools more affordable and private.23 At times, China’s Deepseek v3 exceeds benchmarks. Running its best version requires a powerful computer but frees you from both cost and the Chinese Communist Party’s ideological constraints on their website versions. While Deepseek likely trained on unauthorized materials similar to how Meta and OpenAI are alleged, this doesn’t excuse anyone stealing from creators to build competitors.
Unlike China, America and other democracies value creative rights and fair compensation. Surrendering creators’ rights to big tech risks realizing the dystopian world of ‘Ready Player One’—where creative work is systematically extracted from masses to build immersive virtual worlds that simultaneously entertain and exploit them, concentrating wealth and power in the hands of tech oligarchs.
AI agents threaten to shatter the critical boundaries between applications, operating systems, and user data. The industry’s latest obsession—AI agents that handle everything from concert tickets to calendar scheduling—demands unprecedented access to our digital lives. These agents require your browser, payment information, calendar, messaging apps, and essentially root-level permissions across your entire system. Signal’s founders warn about these risks: agents can be a fundamental breakdown of the privacy architecture that keeps our digital lives secure.24 Because most of this processing happens on remote servers, not your device, private communications, financial details, and personal schedules all potentially flow through corporate servers accessible to governments. Agents are a profound restructuring of digital security that prioritizes corporate convenience and “magic genie” functionality over fundamental rights of privacy—all so our “brains can sit in jars” while AI handles life’s details. The machinery of reality distortion isn’t just changing what we see and how we think—it’s reaching for control of what we do.
Major platforms once promised user empowerment—Facebook, Twitter/X, Google—now frequently amplify misinformation25 and extremism, with consequences as severe as the January 6 Capitol insurrection.26
These firms are retreating from moderation and systematically dismantling safeguards: disbanding fact-checking teams and content-safety plans, and blocking researcher access.27
Most alarming is a mere 3% of users generate one-third of all toxic content—superspreaders of digital pollution, a concentrated source of harm that often leads to physical violence.28 Profit-driven engagement metrics override public safety .
Meta’s and YouTube’s retreat from content moderation is alarming, after X/Twitter’s abandonment led to plummeting content quality and an advertiser exodus.29 Following vague but ominous threats of prosecution from President-elect Trump—a dangerous precedent for a company already under investigation for moving too fast and breaking the law across a number of federal and state agencies—Meta terminated partnerships with independent fact-checkers who had effectively reduced the spread of false information in the US.
A former Meta trust and safety employee warned: “(Moderation) is not the climate change debate . . .This is degrading, horrible content that leads to violence and has the intent to harm other people.”30 By dismantling safeguards against non-illegal but harmful content, it effectively declared “open season” on vulnerable populations—precisely when they face heightened targeting from extremist movements. Meta argues, “the lack of dialogue changes no hearts. At worst it hardens them.” I agree, but the attention economy powering these algorithms are tuned for conflict.31
Similarly, Trump’s promised TikTok reprieve effectively places the platform under an implicit gag order against criticizing him—ironically enabled by the very bipartisan law he inspired to force its sale. This pattern of platforms heel-turning under political pressure, whether driven by political pressure or a fickle founder, signals a dangerous new phase where platform policies heel to intimidation rather than protecting users.
Meme, sports, and pop culture accounts are weaponized, repeating implausible fiction into perceived truth.32 Studies revealed systematic pro-GOP bias in X and TikTok’s content distribution algorithms, while Meta and YouTube’s systems were effectively weaponized by coordinated disinformation campaigns.33 This wasn’t organic virality but sophisticated manipulation of platform mechanics, underscoring why users need direct control over the algorithms shaping their information diet.
When platforms can’t or won’t prevent algorithmic manipulation, the solution isn’t just better corporate oversight—it’s also giving users the power to choose and customize their own content filtering systems.
Even Steve Bannon, Trump’s former chief strategist who helped catalyze MAGA’s rise, warns of “technofeudalism” from Silicon Valley elites. While acknowledging America’s need for technological progress, Bannon argues Silicon Valley’s version of progress threatens fundamental human values and community bonds.
Besides, as Musk’s X platform devolves into an echo chamber of conspiracy theories and his companies increasingly blur the line between corporate and government functions, the solutions proposed here become more crucial, not less. When government actors can shape public discourse and policy with minimal oversight, strengthening information ecosystems becomes existential. The double standard is glaring: Dan Rather’s career ended over sharing one fraudulent document, while Musk’s daily stream of comparable falsehoods only strengthens his following. We must reassert lying as harmful—beyond fact-checking, as a corrosive moral failing. Normalized deception undermines the common ground and authentic connection required for society to function.
The machinery of reality distortion now extends to the physical layer of internet access itself. Senate Republicans’ pivotal votes for the bipartisan infrastructure law demanded the Biden program’s local control and lengthy timelines. The new administration plans to strike Biden-era conditions ensuring federal rural internet investments are affordable to middle-class Americans. Trump’s Commerce Department and Senate Chairman Cruz also want to redirect $20 billion in bipartisan infrastructure funding from permanent fiber networks built to last 50-100 years to Musk’s Starlink disposable satellite system—incinerating bipartisan Internet investment in the atmosphere every 5-7 years. Taxpayers and rural America shouldn’t settle for renting disposable internet on a subscription plan that centralizes a chokepoint susceptible to surveillance, filtering, and censorship, some of which Musk has already demonstrated on X.34
The tech industry is increasingly resembling an ouroboros—the ancient symbol of a snake consuming its own tail. Already, AI generates an estimated 5% of Wikipedia’s English-language articles, meaning new AI models are training on content generated by earlier AI models.35 This self-referential cycle risks amplifying biases and hallucinations, degrading the quality of training data. The risk to science is particularly acute: even breaking the 1% summary hallucination barrier still means searching for 9 or fewer hallucinations per 1000 words. OpenAI’s newest product, Deep Research, while impressive in synthesizing data, exemplifies this problem—experts note it generates plausible-sounding but fabricated figures and misrepresents key aspects, threatening to flood scientific literature with difficult-to-detect errors.
Meanwhile, a 2025 FTC report reveals how Big Tech’s market power creates another similar self-consuming cycle: the largest cloud providers (Alphabet, Amazon, and Microsoft) are securing privileged positions with leading AI developers like Anthropic and OpenAI through investments that grant them exclusive rights and sensitive technical information.36 These firms extracted ideas, top talent and technology licenses from startups, leaving the smaller firm’s younger employees holding less-valuable stock. This consolidation of power means fewer companies control both the infrastructure AI runs on and the AI systems themselves, potentially limiting competition, innovation, and public accountability.
Breaking up Big Tech isn’t a simple solution. Former FTC Chair Lina Khan and others make compelling arguments for it—still, Bell Labs and Alphabet produced remarkable breakthroughs through their monopoly-funded research. Shareholders of these firms when their stocks are split historically do even better over time. Competition in AI is actually driving the worst outcomes—pushing companies to rush dangerous products to market with inadequate safeguards. This suggests traditional antitrust remedies alone won’t solve our AI governance challenges. When companies violated the law, they need to be held accountable. We also need frameworks that can both foster beneficial innovation and prevent destructive competition, rather than simply breaking up large firms and hoping market forces produce optimal outcomes.
Behind AI and social media content lies a hidden workforce: Global South moderators and data workers earning $2/hour to process AI and humanity’s darkest content.37
Creative professionals face their own crisis: artists’ work is scraped without consent to train AI models, while writers and journalists compete against AI often plagiarizes their work.38 Translators, creatives, data workers and others lose livelihoods. This dual exploitation reveals how Big Tech’s AI ambitions depend on devaluing human labor.
Creators deserve a seat at the table during policies affecting their future. While tech companies and their lobbyists shape AI policy, the artists, writers, musicians, and other creatives whose work trains these systems are largely excluded from these decisions.
AI companies trample even basic digital rights, ignoring robots.txt—the Internet’s ‘No Trespassing’ sign—to scrape content without permission. Perplexity.ai and others brazenly violate their own stated policies, deploying hidden bots to plagiarize and harvest explicitly forbidden content protected from those following Internet standards.39
‘Pink slime’ operations—AI-generated local news funded by special interests—now frequently outnumber legitimate newspapers.40 These operations have evolved from clickbait websites to printed tabloids masquerading as community papers, sometimes resurrecting shuttered local or diverse, niche publications. Political groups and industry lobbies flood swing district mailboxes with these sophisticated forgeries, leaving residents struggling to distinguish propaganda from genuine journalism.41
Roughly half of US adults struggle with reading comprehension to some degree; around 130 million people over 16. Basic reading difficulties compound the complexity of online information assessment. Combating this infiltration requires more than digital literacy—communities need media-savvy local validators who can help residents navigate between authentic journalism and sophisticated propaganda.
While Big Tech’s unchecked power poses serious threats to discourse and problem solving, its impact on economic systems and workforce dynamics presents equally pressing challenges.
Recent research reveals a concerning pattern in AI’s impact on labor markets: each profession faces a distinct “inflection point” with AI tools. AIs, trained on material created by the very humans they displace, initially act as productivity multipliers—web developers saw 65% income increases when first adopting AI tools. However, once AI capabilities cross a certain threshold, the relationship inverts dramatically—translators and bloggers experienced 30-40% income drops as AI translation and content generation became more sophisticated.42 Similarly, working musicians have seen their streaming revenue drop as Spotify replaced their songs with AI-enhanced bargain-rack music—another ouroboros where the long tail of music creation eats itself.43 This pattern suggests workforce disruption may be more sudden and severe than gradual adaptation allows, requiring more urgent policy responses.
Congressional oversight of automation’s impact has been inconsistent: robust 1950s investigations found ‘enlightened business’ accepting responsibility for displaced workers, followed by decades of neglect until 2016.44 Components of recent legislation—particularly the CHIPS & Science Act and the Inflation Reduction Act—finally aims to modernize workforce skills and revitalize struggling regions and lower energy and health care costs . RECOMPETE, for example, shows how combining regional planning and solving families’ root crises leads to rebirth amidst turmoil. The law expands innovation hubs and corridors that combine research, manufacturing, and workforce development.
The recent financial downturn and Trump-initiated trade war (which Congress can turn off anytime) have only intensified corporate adoption of Shopify-style policies requiring proof AI can’t do the task before new positions are approved. A sobering assessment comes from one of the past two years’ top-20 technology forecasters, now an AI policy advisor at a think tank. His prediction: there’s a 50% chance that within 10 years, technological advancement will eliminate most current employment. While he acknowledges there will always be some roles where humans are intrinsically preferred over AI, he doubts these positions alone could sustain anything close to our current workforce.45 Declining industrial robot costs reinforce this forecast, as automation becomes viable for even small manufacturers.
Strategic public R&D investment offers a proven counterbalance to this extraction—creating broadly shared prosperity through innovations that solve fundamental challenges from national security to climate change to inequality. Targeted R&D builds problem-solving capacity while yielding returns far exceeding costs. Yet the Department of Government Efficiency’s (DOGE) theatrical cuts target precisely these high-return investments, sacrificing future prosperity for immediate political theater—like Trump directing Army Corps officials to release California’s reservoir water despite knowing it would waste a critical resource for millions.46
Breaking this machinery of reality distortion requires us to confront three crucial points: First, support is essential during the transition period, helping communities adapt rather than collapse. Second, even in a highly automated economy, human needs for care, creativity, and community won’t disappear—they may become even more central to meaningful work and social cohesion. Third, and perhaps most importantly, strengthening these social foundations builds the resilience and civic capacity we’ll need to navigate whatever upheavals AI brings.
The solution lies not just in universal basic income alone, as some in AI advocate, but in a profound reorientation of our political economy toward greater social security and family stability.
A comprehensive framework of social supports could transform American resilience through affordable housing, guaranteed sick leave, accessible child care, job training, universal pre-K, and enhanced market competition—at about one-seventh the annual cost of the Bush deficits. American history offers clear lessons for our present moment: we face a new Gilded Age, where tech giants wield power rivaling the 19th century robber barons, tariffs return as economic weapons, and preventable diseases resurge while billions flow to AI development.
Countries aren’t corporations—they exist to serve people, not quarterly profits. At its core, the purpose of government is to solve the problems that prevent people from living their best lives—focusing on areas where public action can unlock the best solutions and empower citizens, not attempting to solve every issue but strategically addressing the many barriers that hold people back from being their best selves.
Lost in debates about AI’s workforce disruption is an immediate crisis: 48 million Americans serve as family caregivers with minimal financial support. While detailed policy solutions exist, media coverage routinely overlooks these critical proposals in favor of horse-race political coverage. Treating government as merely a business undermines the patient investments in infrastructure, education, and innovation that built American prosperity.47
This approach ensures technological progress serves human flourishing while rebuilding hope in struggling communities. It’s about building an economy that works for everyone, not just those at the technological frontier. These policies would mark a decisive shift from treating economic insecurity as inevitable to seeing it as a solvable challenge through targeted, proven interventions that address fundamental causes of hardship while strengthening community resilience.
This isn’t about vilifying wealth or success—it’s about distinguishing value creation from value extraction. The issue isn’t prosperity but predation—business models that drain communities while enriching the few. Competition benefits consumers while even well-intended regulations need constant oversight to ensure they serve people, and aren’t captured by the very industries they oversee. We need more entrepreneurs solving real problems, not fewer. The path forward requires both successful businesses and equitable systems that ensure prosperity is broadly shared.
Rather than accepting a future where AI concentrates wealth while leaving communities behind, this framework uses public investment to address root causes of economic hardship, not just symptoms. Like our chronic disease epidemic where 90% of healthcare spending treats preventable conditions, we know how to solve these economic challenges but often focus on treating symptoms rather than causes.48 RFK Jr.’s appeal illustrates how corporate influence in healthcare and agriculture has created both economic and health crises—though his dangerous anti-vaccine advocacy undermines his legitimate critiques of corporate power.49
During Congress’s four decades of inattention on automation, $50 trillion moved from the bottom 90% of Americans to the top 0.5%—history’s largest wealth transfer.50 Congressional tax cuts (1978-1989) slashed rates on corporations, capital gains, and top earners by nearly half. After the 1993 tax increases briefly created surpluses, subsequent cuts (2001-2003, 2017) failed to generate promised investment while expanding deficits, with national debt soaring from $5.6 trillion to $35.32 trillion.51
Research confirms this shift fundamentally changed corporate behavior, increasing short-termism and causally reducing influential inventions, harming both firm competitiveness and U.S. economic growth.52 These findings confirm our economy has become dominated by rentier extraction rather than value creation—the technofeudal reality Bannon described where gatekeepers extract wealth without creating proportional value.
Recent data from Moody’s Analytics reveals an alarming economic dependency: the top 10% of earners now drive nearly half (49.7%) of all consumer spending—up dramatically from 36% three decades ago. While the wealthiest increased their spending 12% from 2023-2024, middle and lower-income households actually reduced spending during the same period, creating a fundamentally unstable economic structure that rises or falls on the whims of the few. This dependency isn’t accidental—it’s the logical outcome of the massive wealth transfer upward. When a February 2025 Wall Street Journal analysis shows just 24 “superbillionaires” control $3.3 trillion (equivalent to France’s entire GDP), we’re witnessing more than inequality—it’s a complete restructuring of economic power. Nobel Prize-winning economist Joseph Stiglitz identifies the root cause: monopoly power in tech coupled with systematic tax avoidance has created a self-reinforcing cycle where “these guys live in a totally different world from ordinary Americans,” undermining the solidarity essential for society to function.
Corporate focus on quarterly profits and shareholder returns has systematically eroded the traditional duties of care that businesses once maintained toward their customers. Congress and states must counter this trend by implementing enforceable duties of care across more sectors—such as social media platforms, utilities, AI developers, and online gambling operators. These duties would require businesses to prioritize consumer safety over engagement metrics or quarterly returns, with meaningful penalties for executives who knowingly deploy harmful systems. Much as healthcare providers face consequences for negligence, companies deploying services that predictably harm users should bear responsibility. The epidemic of despair we’re witnessing—reflected in rising depression, substance abuse, and suicide rates—correlates directly with industries that profit from exploiting vulnerability rather than building value.
Recent data shows between 25-30% of Americans live paycheck to paycheck or face financial insecurity—a crisis compounded by widespread financial illiteracy and deepening inequality.53 The zero-interest era’s artificial suppression of consumer costs planted a time bomb of public rage. When VC and notably petro-state sovereign wealth funds quietly withdrew their massive tech subsidies and services like private burrito taxis hit real market rates, millions watched their glimpse of an affordable future vanish. The whiplash from subsidized convenience to the new unaffordable reality may have been more radicalizing than traditional political grievances.
Rates of despair—measured by depression, anxiety and suicide—are growing in wealthy nations like ours. New research by neurologists finds that people in despair are particularly vulnerable to misinformation, conspiracy theories, and radicalization. The epidemic of loneliness—the ‘anti-social century’—further erodes social ties, driving many deeper into digital worlds where misinformation and conspiracies thrive uncontested.
The politics of despair paved the way for dramatic policy shifts that often exacerbate rather than address these fundamental economic challenges. President Trump’s first term added $8 trillion in enacted spending hikes and tax cuts to the deficit. Despite campaign promises of fiscal responsibility, the hard truth is that Trump’s economic policies primarily benefit the wealthy while burdening working Americans with long-term costs.
This pattern continues in his second term, where the appearance of “breaking the system” masks the reality that it’s being broken to serve narrow interests, not to fix fundamental problems for average citizens. Contrary to strengthening US manufacturing, a Federal Reserve paper revealed Trump’s 2018-19 tariffs actually increased prices and cut manufacturing jobs without boosting production, particularly damaging competitive manufacturers reliant on trade inputs and export markets.54 The new Treasury-backed cryptocurrency reserve prioritizing the five digital currencies owned by Trump’s crypto advisor exemplifies this troubling trend.55 Even conservatives called it “a slap in the face” when Trump appointed a junk food industry lobbyist to run the Department of Agriculture. Despite his promise to “drain the swamp,” one out of 14 of his appointees from his first term were lobbyists in positions to regulate their industries creating conflicts of interest.
While Elon Musk’s DOGE has become emblematic of what even fiscal conservatives label “spending cut theater.”56 Despite the hype, DOGE’s genuine cuts represents just 0.04% of the proposed $4,500 billion in proposed new tax cuts.57 After journalists pointed out calculation errors in their initial claims, DOGE officials responded not with corrections but by reducing transparency, making independent verification nearly impossible. Contracts were public when valid; why the secrecy when challenged?58 Not content with proposed congressional cuts to food aid, DOGE slashed critical programs that support food banks and pay farmers for healthy local produce—ironically undermining the very wholesome food systems the Make America Healthy Movement champions . Meanwhile, DOGE has ‘detonated a crisis at a highly sensitive nuclear weapons agency’ and systematically fired top performing federal employees who recently received promotions.59
The Government Accountability Office (GAO)—whose actual job is investigating fraud and inefficiency—identified $70 billion in savings last year and $2 trillion since 2003, and often reports that understaffing, not bloat, causes delays and inefficiencies—a counterintuitive but evidence-based conclusion that true fiscal conservatives should embrace. There’s no excusing the end-run around Congress. Clinton and Gore reinvented government and got them to pass the Government Performance and Results Act, DOGE couldn’t even bear to work through a bipartisan caucus dedicated to its work.
The Internet’s knowledge commons faces dual threats: vital expertise vanishes behind paywalls (Substack, newspapers need better options), while ‘link rot’ erases public access to human knowledge. The best congressional policy coverage remains priced for lobbyists, not citizens—creating an information aristocracy despite some journalists finding fragile lifelines through crowdfunding. Members of Congress and influencers with access to paywalled news like Politico Pro could help bridge this gap by sharing gift links and reacting on camera to key facts, ensuring the public understands policies affecting their lives.
Meanwhile, documentary and other edgy filmmaking and television—some of our most powerful tools for preserving human perspectives and understanding complex social issues—face their own crises as funding sources constrict and distribution channels prioritize non-challenging formats.60 Even as traditional media struggles to explain what happens in Washington, some innovative formats break through: podcasts like ‘The Middle’ prove thoughtful dialog can bridge divisions, while fearless documentarians continue capturing crucial narratives that challenge our comfortable assumptions. These voices—whether through film, audio, or emerging formats—remain our best defense against digital oversimplification.
“Garbage In, Garbage Out,” my grandmother would say during her time as a human computer—one of the mathematicians who calculated missile trajectories by hand before personal computers existed—at White Sands, whenever her team’s calculations produced faulty outputs due to incorrect inputs.
As Texas teacher Chinea Bond discovered, students relying on AI produced ‘really, really bad’ papers, lacking fundamental analytical skills.61 Instead of rushing to rely on AI, learners and new AI users should be lead by teachers to first build a bedrock of “core skills” only eventually adding AI to truly enhance their capabilities. Finland’s systematic approach to digital literacy—where ninth-graders consistently lead global rankings in detecting misinformation—offers a proven model.62 Their integration of fundamental skills—reading, mathematics, finance, science, civics, critical thinking—throughout K-12 education creates a foundation for evaluating sources and using AI responsibly, even a free book for every 9th grader.63 The Homework Apocalypse—where AI can complete most traditional assignments—presents an even subtler danger: students may uncritically accept AI’s plausible-sounding but potentially flawed connections and analysis, rather than developing the essential skills of teasing out genuine patterns and verifying historical or factual accuracy. This demands not just new forms of assessment, but teaching students to approach AI-suggested insights with scholarly skepticism.64
Educators nationwide are discovering that phone bans improve not just attention spans, but student wellbeing—a wake-up call for all of us. Phone manufacturers quietly acknowledge radiation risks in their fine print, and warn against the millions who keep phones pressed against their bodies routinely. Don’t use the cellphone as a phone and save your eyes: in iOS open Settings>Accessibility>Accessibility Shortcut. Try Reduce White Point first, toggle it by triple clicking the Lock/Side button, then try Smart Invert. Use whichever you prefer.
Adults require similar support our children do. Grassroots, ‘guerrilla’ outreach can deliver reliable news to communities drowning in misinformation, particularly where local and niche newspapers have vanished. Local journalism helps citizens distinguish truth from fiction, rebuilding trust and resilience against AI-driven manipulation. It’s not a giant leap: after state and local civil rights groups couldn’t match 2020 funds to educate voters,65 weeks later, Meta enabled targeted harassment campaigns that historically escalate to real-world violence.
In 1894, London and New York faced what seemed an insurmountable crisis: their streets were drowning in horse manure.66 The very technology powering urban transportation threatened public health and city life itself. Today’s AI boom presents a remarkably similar challenge—computing infrastructure threatens to overwhelm our electrical grid and environment.
New research shows billions will lose their livelihoods and global economic output could plummet by up to 34% if Earth warms by 3°C this century—yet investing less than 2% of GDP now could eliminate most of those catastrophic losses. This economic reality makes addressing AI’s environmental footprint not just an ethical imperative but a financial necessity.67
While some dismiss AI’s environmental impact by comparing individual queries to daily actions like eating meat, this narrow analysis dangerously understates the massive scale of AI infrastructure challenges.68 Meta’s and Google Search’s shift to AI in 2024 significantly raised their energy footprints.69 Automated AI agents multiply this impact often generating thousands of queries per second.
AI servers consume 10-100 times more energy than standard web/email servers,70 demanding ever-larger facilities that lock regions into decades of ecological consequences. These hidden costs hit vulnerable communities hardest.
Loudoun County, Virginia—which handles 70% of global email—has seen energy use surge 240% in five years, reaching 3.4 GW, with projections showing consumption tripling by 2028.71 Its grid is so strained that data centers also rely on over 4,000 constant, highly-polluting, noisy diesel generators.
America’s power system operates at just 53% average utilization—meaning our vast network of generation plants and transmission lines sits unused nearly half the time. The core challenge is getting power where and when it’s needed. Strategic deployment of American-made innovations—including storage, microgrids, advanced transmission, and local grid technologies—could solve this mismatch, but utilities often seek a different path.
Yet while utilities cite AI-driven growth to justify expensive new fossil fuel plants, independent analysis tells a different story. The Tennessee Valley Authority—home to xAI’s world’s largest AI supercomputer—projects more modest demand growth. Their assessment aligns with Duke University’s Nicholas Institute’s 2025 analysis, which confirms our existing grid can handle significant new loads through flexible management, modernization, and distributed generation. A recent comprehensive energy report offers promising paths forward that align with values across the political spectrum—from market competition to environmental stewardship to consumer protection.72
Utilities are following a predictable playbook—starting the conversation with yesterday’s solutions because they earn guaranteed returns on new power plants in ways they don’t from grid improvements or efficiency. As a result, Wall Street is surging utility stock prices.73 Meanwhile, Tech companies are already generating more of their own clean power to meet their immediate AI-driven demand. But if utilities push rates too high, these companies might eventually move all their operations behind the meter—stranding expensive assets and forcing regular customers to shoulder an even heavier burden. The result? A perverse system forcing everyday people to pay more to burn the dirtiest fossil fuels in vulnerable communities to generate digital pollution and misinformation.
Utilities make substantially more money over a gas plant’s lifetime than if a large industrial user brings their own power generation. This incentivizes fighting against the very technological innovations that could reduce costs for consumers and pollution for communities. That’s what petro-oligarchs are afraid of too—that the world of cheap American clean power made with our friends in democracies that respect workers and the planet is within reach. But incumbents and monopolists in our country and outside it fight to keep us tethered to a global market of toxic substances. We can do better.
Local officials in Loudoun County report facing three stark choices: government-imposed energy constraints, breakthrough technology and efficiencies, or rapid development of clean onsite microgrids. The gap between utility claims and technical reality highlights why communities need accurate data to make informed infrastructure decisions .
A single Amazon campus in Indiana will consume over 2,200 MW—more than many cities. Major tech companies will triple both the state’s electricity sales this decade and its carbon pollution over 20 years. Indiana’s sales tax exemption surrenders $23 billion in data center revenue, while these facilities receive enormous public subsidies and force ratepayers to shoulder billions in infrastructure costs, perpetuated by decades-old fossil fuel subsidies. State boundaries offer no protection from these hidden subsidies: Kentucky’s PSC and Attorney General recently documented through a FERC complaint how their residential customers are being charged for transmission upgrades serving Big Tech data centers in Indiana and Ohio—revealing the urgent need for interstate regulatory reform.74
By 2029, AI facilities could require an additional 128 gigawatts—equivalent to powering a small nation. Data centers could devour 17% of all U.S. electricity by 2030, quadrupling today’s consumption.75 While China expects renewables to meet all new electricity demand by 2025, effectively capping coal power growth, U.S. utilities continue expanding fossil fuel infrastructure.76 The EPA now distorts reality by challenging that carbon pollution endangers human health—a scientific fact established decades ago—creating a dangerous regulatory environment where polluters would face fewer constraints.77
In North Omaha, where people of color make up 68% of the population and asthma rates rank among the nation’s highest, residents watched the 2023 deadline to stop burning coal vanish. Meta’s Nebraska facility alone consumed nearly as much electricity as the North Omaha coal units produced in 2023, while Google’s Nebraska usage ranks as its highest in the U.S. Though both tech giants claim “net zero” impact through distant renewable contracts, they remain silent about local fossil fuel use. Energy economists concluded: “If not for the data centers and poor planning by the utility, they would not need to push to keep those coal units open.”
Downriver in Memphis, xAI’s facility depletes local aquifers of 1 million gallons daily despite available treated wastewater nearby.78 After replacing an oven factory that once employed thousands with a data center providing dozens of jobs, its new methane generators add pollution to neighborhoods already suffering elevated asthma and mortality rates.79
The cost-shifting between data centers and utilities reveals this manipulation. While Exelon fought Amazon’s attempt to force Mid-Atlantic ratepayers to cover its data center costs,80 its ComEd subsidiary will shift 90% of a $5 billion grid expansion onto Illinois ratepayers81—showing how both tech companies and utilities maximize profits at public expense. A new paper investigates nearly 50 regulatory proceedings investigating data centers shifting their utility costs to consumers and other businesses.82
Secret negotiations for data center pricing undermine fair rates and encourage venue-shopping. Congress could eliminate these problems by banning confidential agreements between utilities and data centers, requiring cost-based rates without special carveouts or subsidies. While Texas’s Smart Meter program provides usage data which saved ratepayers up to 18%, New York’s IEDR will include both usage and cost information—an even better model for transparency. States could also require new large loads to procure a percentage of demand from virtual power plants, similar to California’s proposal. The FTC needs expanded enforcement power to investigate utilities blocking customer data access—practices that would be illegal in banking or healthcare.
Data centers create dangerous ‘bad harmonics’—electrical distortions that damage nearby residents’ appliances and increase fire risks in both urban and rural areas. Communities bear these costs through higher insurance rates and damaged electronics.83 Utilities say they’re not the problem but refuse to share exculpatory data .
The energy rate problem is so severe that most states’ failed regulatory approach allows investor-owned utilities to enjoy ratepayer-guaranteed returns nearly double their market-based cost of capital. This costs families $300 a year, totaling $50 billion while slowing the transition to cleaner, more efficient systems. A new paper argues Congress and states should pass legislation setting the rate of return at the actual cost of capital, ending this legalized extraction from American families.84
Unlike the 1894 Horse Manure Crisis, clean energy solutions to our AI power challenge already exist and often cost less than fossil fuels.85 The Department of Energy’s Loan Program Office leverages expertise to help banks commercialize technologies from research programs like ARPA-E, generating positive revenue while accelerating American innovation—precisely the efficient government initiative that efficiency advocates should champion.
Smart load management offers an alternative to costly new construction. Some data centers participate in demand response programs, agreeing to reduce power during peak stress in exchange for lower rates.
While most tech companies struggle with sustainability, mid-sized firms like Stripe demonstrate creative solutions, showing how data centers could affordably operate off-grid through smart integration of solar, battery storage, and methane gas generators—though connecting to the grid remains more cost-effective, even at reduced capacity.86
Grid-enhancing technologies offer immediate solutions. Smart sensors and software can increase transmission capacity by over 33%, while advanced controls redirect power around congested lines. These technologies could save ratepayers billions annually while enabling more renewable energy connections. States are leading this transformation—California, Minnesota, South Carolina, Utah, and Virginia now require utilities to evaluate GETs in their resource planning. Several states have enacted performance-based rate regulations, aligning incentives toward improving resilience, affordability, and reducing pollution, rather than just building more infrastructure.
Texas presents both promises and warnings. The state generates massive clean power—31 GW from wind and solar plus 5 GW from nuclear as of 2024. Its ‘connect and manage’ interconnection process enabled its grid ERCOT to add about three times more than multi-state PJM (the nation’s largest grid operator), and ERCOT outperformed PJM by an even wider margin in 2022.87 Texas also pioneered consumer-empowering data policies that enable 6-18% savings88 (New York improved those regulations requiring utilities pass along cost data too). However, Texas’ over-reliance on methane gas infrastructure proved catastrophic during the Big Freeze. The system’s collapse contributed to 700 deaths, disproportionately affecting low-income communities of color, illustrating the deadly costs of infrastructure dependency on a single fuel source.89
Similarly, California’s experience offers critical lessons. While California pioneered many grid innovations, its isolated transmission system and outdated utility liability rules create unnecessary risks and costs. Despite billions spent on wildfire damage, we know how to prevent homes from burning—controlling a home’s 100-foot perimeter ignition zone and following the National Fire Protection Association checklist can dramatically decrease fire risk.90 Community microgrids with rigorous safety protocols could help utilities avoid catastrophic wildfire risks while improving reliability, protecting both ratepayers and utilities from preventable disasters. Yet consistently, we fail to rebuild with the foresight needed to withstand inevitable future disasters.
Supply chain bottlenecks threaten this modernization. Severe transformer and substation shortages—with lead times now stretching beyond 24 months—create urgent need to maximize existing assets.
This is the time for an all-hands-on-deck strategy. While utilities will keep natural gas as a priority, increasing planned exports could raise prices for domestic manufacturers. We must diversify our energy sources and support industries friendshoring our energy supply chains to ensure we win the race for a clean economy without being beholden to another set of oligarchs.91
Reducing the federal workforce will directly undermine these efforts. As senior regulatory experts warn, reducing agency expertise doesn’t just slow permitting—it can turn a one-year process into 3-4 years. The benefits of the supposed end of permitting might be overblown—business-community watchdogs say Chicago left its unfinished Block 37 Superstation a boondoggle because officials spent money without process and public input.92
The new administration faces a pivotal choice: embrace the hundreds of thousands of jobs (especially in the trades) and $450 billion in private investment already flowing from landmark legislation, or jeopardize America’s energy advantage and force household energy costs up at least 10% monthly.93 With families already struggling with inflation, blocking clean energy deployment directly threatens both household budgets and national competitiveness.94 Companies across the country stand ready to leverage these laws to create jobs and consumer savings95—but reversing course on energy modernization would be like deliberately breaking all our eggs on the way home from the store. Competitive entrepreneurs, conservatives, conservationists, and consumer advocates will find much to appreciate in a new report charting the path to a better energy future.96
State Public Utilities Commissions (PUCs) make crucial decisions about energy costs and data center impacts through public proceedings that directly affect utility bills and community health. These venues offer concrete opportunities for citizen involvement.97 When informed, organized residents participate in PUC hearings, they’ve successfully challenged utilities’ inflated projections and secured better protections against unfair cost allocation. Groups focused on urgent pollution reduction and affordable clean energy are playing crucial roles in preventing costly boondoggles while building sustainable, reliable alternatives.
Addressing AI’s environmental impact requires systemic change, not just individual usage modifications or incremental efficiency improvements. Big Tech’s carbon-free energy promises clash with exploding AI demand, showing the urgent need for:
While technical and regulatory solutions can address AI’s environmental impact, they represent just one aspect of a broader accountability challenge. As AI systems grow more powerful and autonomous, we must establish ensure responsible development and deployment.
The Westfield High School incident in New Jersey—where students allegedly used Clothoff’s AI to generate non-consensual intimate images—at least offers a difficult-to-find identifiable party for accountability.98 Yet as AI tools multiply, victims increasingly face anonymous harassment from untraceable sources, rendering traditional legal remedies futile.
These aren’t theoretical problems. Meta’s “FungiFriend” AI chatbot with a friendly wizard icon gave potentially lethal mushroom foraging advice, demonstrating AI systems can cause immediate public safety risks even operating exactly as designed.99 Unlike traditional published content, which courts have historically protected from liability, AI systems actively generate new, potentially dangerous content in real-time interactions with users.
Corporate personhood, our traditional framework for limiting business liability, assumes corporations act through human agents. However, AI breaks this model – it can develop capabilities and take actions beyond what any human intended or controlled. When AI systems operate autonomously across jurisdictions or through distributed networks, even identifying responsible parties could become nearly impossible.100
Tort law’s fundamental premise—identifying who caused harm—faces unprecedented challenges in this new reality. While courts will play a crucial role in addressing AI harms, traditional case-by-case solutions cannot fully address systemic challenges.
At the core of the debate: whether product liability or negligence standards better suit AI harms?101 While product liability could strengthen software safety without requiring plaintiffs to prove specific development failures, AI systems pose unique challenges to this framework. Unlike traditional products, AI platforms continuously evolve through updates and user interactions, often invisibly to users. Such evolution fundamentally challenges the definition of a “product” under the law. Negligence law’s reasonableness standard might offer a more adaptable solution, focusing on developer behavior rather than product definitions, and holding AI creators to objective standards of proper conduct rather than merely common practice.
Two approaches emerge as potential solutions:
These aren’t future problems. AI is already concentrating wealth and power while externalizing costs to society. While tech leaders profit from disruption, communities shoulder the consequences. Whistleblower protections are also urgently needed.104
As AI systems grow more autonomous, fundamental questions of control and ownership emerge: Can AI ‘entities’ legally hold and manage assets? Who maintains ultimate control? At its core, algorithmic transparency represents a power dynamic. When platforms hide their decision-making processes, they’re effectively running black-box governance over our digital lives.
Recent “pro-social media” research argues these systems should strengthen rather than erode the social fabric they depend on. This requires dual transparency: both verified original provenance showing where content comes from and social provenance indicating which communities it serves by bridging divides or balancing diverse perspectives.105
The 2024 Crowdstrike incident illustrates how black-box systems of all types create systemic risk: failures paralyzed operations for thousands of businesses, hospitals and governments—shows how ignoring basic software safety practices can cascade into widespread societal harm.106 When companies deploy powerful tools like kernel drivers that can affect millions of users, they should face meaningful accountability, including potential criminal penalties, balanced with clear safe harbor provisions for adopting memory-safe programming languages and other proven safety measures.
As AI systems become increasingly integrated into complex software infrastructure, the potential for cascading failures grows exponentially. This demonstrates why we need algorithmic transparency, stronger preventive measures for basic software safety and clear liability frameworks that account for AI’s unique capabilities and risks.
While federal action remains crucial, states are emerging as vital laboratories for AI governance, using inherent powers and (new FN) developing models that can inform national policy.107 Twenty states now have privacy laws, with California and Maryland setting the highest bar.108 Through global forums and interstate coalitions, state legislators and attorneys general can champion human-centered values in AI development while federal action lags. Successful privacy frameworks embrace data minimization, enable individuals to protect their own rights, eliminate deceptive design practices, and reinforce existing civil rights.
Tech companies’ stance on state privacy laws is revealing: while publicly supporting state frameworks, they often lobby for weaker standards than California’s comprehensive protections. Yet these same companies comply profitably with California’s stricter rules in their home state, and many simply apply California’s standards nationwide. Some even make privacy a marketing advantage, demonstrating how strong protections can align with business interests while reducing regulatory complexity.
Recently-ended preliminary Federal Trade Commission (FTC) investigations reveal how companies increasingly use personal data—from mouse movements to shopping cart abandonment—to set individualized prices, often without consumer knowledge or consent. This “surveillance pricing” affects everything from cosmetics to groceries, fundamentally altering how companies compete and consumers shop. Despite law enforcement’s claims that facial recognition merely provides investigative leads, the technology has directly led to wrongful arrests and detentions.109 Where you live matters too, merchants charge different prices for hotel rooms based on where they think you’re coming from.110 Beyond just price discrimination, these practices represent a broader crisis in digital rights. Legislation could establish frameworks that shift power from platforms back to citizens—enabling them to understand why they see specific content, control how their data is used for pricing, and choose algorithms that promote “time better spent” rather than maximizing engagement.111 This approach builds on successful precedents like phone number portability, which transformed telecommunications by giving consumers real choice without micromanaging business practices.112
In a break with several other courts ruling the FCC has authority over high-speed Internet—a circuit court decision stripped the Federal Communications Commission of its authority over access to the Internet. This ruling could let ISPs charge premium fees for AI access or force consumers into expensive bundles for basic AI tools. While six states have net neutrality laws that effectively protect consumers nationwide (as networks avoid operating dual systems), this patchwork approach leaves significant gaps in cybersecurity oversight.113
Congress gave the FCC authority in Title II of its law to protect network reliability and security; no other federal agency can establish even basic cybersecurity standards for our critical Internet infrastructure. Incoming FCC Chair Brendan Carr’s Project 2025 chapter proposes abandoning traditional consumer protection in favor of sweeping new powers over social media and AI.114 Expanding authority to regulate speech while weakening established consumer protections raises significant concerns.
As digital threats multiply at breakneck speed, the Federal Trade Commission’s antiquated rule-making procedures—which can take up to five years to implement—leave Americans increasingly vulnerable to privacy breaches and online fraud. Either the FTC should be empowered with streamlined Administrative Procedures Act authority, or Congress needs to establish a dedicated digital protection agency built for the modern era. A comprehensive digital strategy should include national Right-to-Repair legislation, federal anti-SLAPP protections, enhanced e-government services for all citizens, and policies ensuring researcher access to platform data while maintaining privacy safeguards—all crucial steps toward a more transparent and equitable digital marketplace.115
Consumer Reports leads the way with rigorous, unbiased research and testing of AI technologies, often collaborating with other consumer advocacy groups. While many new AI-focused organizations are emerging—some with questionable motives—established groups like the Electronic Privacy Information Center, Public Knowledge, Center for Humane Technology, Electronic Frontier Foundation116 and others have proven track records of protecting digital rights and offer concrete ways to shape our future.
The AI industry’s reluctance to spell out their vision for human flourishing in an AGI world makes the need for thoughtful, proactive governance even more urgent. We cannot afford to stumble blindly into a transformed society without carefully considering and shaping the human experience we want to preserve and enhance .
We face a watershed moment rivaling the Industrial Revolution—but compressed into years, not decades. AI stands at a crossroads: it could accelerate social decay, worker exploitation, and environmental destruction, or catalyze societal renewal and sustainable progress.
This pivotal moment demands immediate action to address root causes:
What emerges from this analysis is the revelation of an interlocking machinery of reality distortion that we must collectively dismantle before it’s too late. It’s not just isolated problems of misinformation, economic exploitation, or environmental damage. Rather, these systems work in concert: AIs feed the internet’s ‘justification machine,’ which is powered by the attention economy, which drives data center expansion, which burdens communities with hidden costs, which erodes trust in institutions, which weakens our collective ability to distinguish truth from fiction and solve problems.
This problem is global, creating a global prisoner’s dilemma where the race for AI supremacy forces safety shortcuts and rushed deployment. This risks not just direct conflict but AI capabilities spreading to non-state actors—while transforming military strategy, enabling unprecedented cyber operations, and accelerating weapons innovation. In the age of artificial general intelligence, the stakes of getting governance right extend far beyond any single nation’s borders.
Understanding this machinery of our potential undoing is crucial to preventing it. While this analysis raises alarming concerns about Big Tech’s consolidated power, many proposed reforms actually dovetail with Silicon Valley’s self-interest: reduced energy costs, improved network infrastructure, and reliable human-curated content all serve both corporate profits and the public good.
The facts are clear: too much power in too few hands has created an unprecedented machinery of reality distortion. The solution must involve redistributing this power—to citizens, communities, and civic oversight mechanisms—while establishing guardrails that ensure technology serves humanity rather than the reverse. The vision is not better digital Caesars with improved intentions, but what some call ‘a world without Caesars’—replacing centralized control with distributed governance, algorithmic transparency, and genuine user agency .
The stakes could not be higher: our response to these challenges will determine whether AI accelerates societal decay or catalyzes sustainable progress. Stay engaged—America needs you now. Hold elected officials accountable by tracking their votes and showing up at town halls. Organize around just one issue you care about—doom scrolling doesn’t count. Join a group taking action or start one yourself. With proper governance frameworks and decisive action, we can preserve human agency, advance shared prosperity, and ensure AI serves rather than subjugates humanity.
These solutions don’t require waiting for Washington. States, businesses, and communities can act:
Reports & Papers
Explainers
Explainers