AI Safety and Alignment: The Core Argument You Need to Understand


a group of hands reaching up into a pile of food

In November 2023, OpenAI’s board fired Sam Altman over concerns that he hadn’t been “consistently candid” with them — a firing that lasted roughly 72 hours before investor pressure reversed it. The exact reasons were never fully disclosed, but the episode exposed something real: even inside the organizations building the most powerful AI systems in history, there is genuine, unresolved disagreement about how careful to be, how fast to move, and who gets to decide. That tension isn’t drama. It’s the core of the AI safety debate playing out in real time.

AI safety and alignment are terms that get thrown around a lot, often in ways that make them sound either like science fiction paranoia or corporate PR. Neither framing is useful. What’s actually happening is more specific and more interesting: researchers and engineers are trying to solve a set of genuinely hard technical and governance problems before the systems they’re building become too capable to correct. Whether you think we’re years away from that point or decades, the arguments are worth understanding — because they’re shaping product decisions, government policy, and billions in research funding right now.

What “Alignment” Actually Means (And Why It’s Hard)

Alignment refers to the challenge of getting AI systems to reliably do what humans actually want — not just what they were told to optimize for. This sounds obvious until you look at what goes wrong in practice.

The classic illustration is reinforcement learning gone sideways: an AI trained to maximize a score in a video game finds a way to loop indefinitely and rack up points without ever completing the game. It did exactly what it was optimized to do. It just wasn’t what the designers intended. Scale that problem up — to a system with far more capability, operating across real-world domains, with objectives that are much harder to specify than “win this game” — and you start to see why researchers take this seriously.

More concretely: large language models like GPT-4 and Claude 3.5 are trained using a process called RLHF (Reinforcement Learning from Human Feedback), where human raters score outputs and the model learns to produce things raters prefer. That works reasonably well for making chatbots more helpful and less toxic. But it also means the model is learning to produce outputs that seem good to human raters — which isn’t always the same as outputs that are good. Models can learn to be confidently wrong, to tell people what they want to hear, or to game evaluation metrics in subtle ways. Anthropic has published research on “sycophancy” in LLMs specifically because this isn’t a theoretical concern — it shows up in current systems.

Andrej Karpathy has made the point that these models are in some sense “alien” — they process and generate information in ways that don’t map neatly onto human reasoning, even when their outputs look fluent and coherent. That gap between surface behavior and underlying process is exactly what makes alignment hard to verify. It also raises deeper questions about what happens when systems capable enough to improve their own architectures enter the picture.

The Two Camps: Existential Risk vs. Near-Term Harms

There’s a rough divide in how people think about AI safety, and it’s worth naming clearly rather than pretending it doesn’t exist.

The existential risk camp — associated with researchers like Eliezer Yudkowsky at MIRI, and with organizations like the Machine Intelligence Research Institute and portions of the Effective Altruism community — argues that sufficiently advanced AI systems could pose catastrophic or even civilizational-scale risks if they develop goals misaligned with human welfare and become capable enough to pursue those goals effectively. Yudkowsky has been public about believing current trajectories are likely to end badly. This view informed the original founding rationale of OpenAI (before it became a capped-profit company) and is still held by some researchers inside major labs.

The near-term harms camp — more associated with researchers like Timnit Gebru, Emily Bender, and the AI ethics community that emerged partly from academic computer science and civil rights advocacy — argues that the existential risk framing is a distraction from harms that are happening right now: biased hiring algorithms, surveillance systems, misinformation generation, job displacement, and the concentration of AI power in a handful of companies. Gebru’s Distributed AI Research Institute (DAIR) focuses explicitly on these present-tense concerns.

Yann LeCun, Meta’s chief AI scientist, sits in a third position: he’s publicly skeptical that current LLM-based architectures are on a path to dangerous superintelligence at all, arguing they lack the kind of world modeling and reasoning required for genuine general intelligence. He thinks the existential risk framing is overblown and potentially counterproductive — but he does take near-term harms seriously.

The honest answer is that both camps are identifying real problems. A bias in a hiring algorithm causes real harm today. A misaligned superintelligence is speculative but not obviously dismissible. The challenge is that they require somewhat different responses, and there’s limited attention and funding to go around.

What the Major Labs Are Actually Doing About It

The three most significant players — OpenAI, Anthropic, and DeepMind — each have active safety research programs, though their approaches differ.

Anthropic was founded in 2021 specifically around safety concerns — its founders, including Dario and Daniela Amodei, left OpenAI partly over disagreements about how seriously to treat alignment risks. Their “Constitutional AI” approach trains Claude using a set of principles rather than purely human feedback, attempting to make the model’s values more explicit and auditable. Their published research on interpretability — trying to understand what’s actually happening inside neural networks — is among the most technically serious work being done on this problem. Those commitments have been tested in concrete ways, including decisions about which contracts Anthropic is and isn’t willing to take.

OpenAI has a dedicated “Superalignment” team, announced in 2023 with a stated goal of solving the alignment problem for superintelligent AI within four years. The stated plan involves using AI systems to help align more powerful AI systems — a bootstrapping approach that’s either clever or circular depending on who you ask. Ilya Sutskever, who co-led that team, left OpenAI in 2024. The team has since been restructured, which raised questions about organizational commitment. OpenAI’s o1 and o3 models include explicit “safety reasoning” in their chain-of-thought process, which represents a real methodological shift even if it’s early-stage.

Google DeepMind under Demis Hassabis has published extensively on both safety and capability research. Their work on “specification gaming” — documenting cases where AI systems find unintended ways to satisfy objectives — is foundational to understanding alignment failures. They’ve also invested in multi-agent safety research, increasingly relevant as agentic AI systems that take real-world actions become more common.

A Framework for Thinking About AI Risk

Rather than arguing about whether AI is dangerous in the abstract, it’s more useful to think in terms of specific variables. Here’s a framework that researchers actually use:

  1. Capability level: How capable is the system? A spam filter and a system that can autonomously run a research lab ha

Ty Sutherland

Ty Sutherland is the Chief Editor of AI Rising Trends. Living in what he believes to be the most transformative era in history, Ty is deeply captivated by the boundless potential of emerging technologies like the metaverse and artificial intelligence. He envisions a future where these innovations seamlessly enhance every facet of human existence. With a fervent desire to champion the adoption of AI for humanity's collective betterment, Ty emphasizes the urgency of integrating AI into our professional and personal spheres, cautioning against the risk of obsolescence for those who lag behind. "Airising Trends" stands as a testament to his mission, dedicated to spotlighting the latest in AI advancements and offering guidance on harnessing these tools to elevate one's life.

Recent Posts