Jonathan Albarran | Writing

Essays and reflections on technology, decision-making, and the systems that shape modern work.

What is AI? The technology reshaping human civilization

Artificial intelligence has become the most consequential technology of the early 21st century, capable of writing code, diagnosing diseases, and generating photorealistic videos—yet its creators still cannot fully explain how it works. In 2024, AI researchers won the Nobel Prize in Chemistry for predicting protein structures, AI systems achieved silver-medal performance at the International Mathematical Olympiad, and companies poured over $100 billion into AI development. This technology, once confined to academic laboratories and science fiction, now touches billions of daily lives through search engines, virtual assistants, and an expanding array of applications that seemed impossible just five years ago.

Understanding AI has become essential not merely for technologists but for anyone seeking to navigate the modern world. The decisions being made today—about how AI systems are built, regulated, and deployed—will shape economic opportunity, scientific discovery, and the balance of power for decades to come. What follows is a comprehensive guide to this transformative technology: what it is, how it works, what it can and cannot do, and where it might be taking us.

From Turing’s dream to ChatGPT’s reality

The quest to create thinking machines began long before silicon chips existed. In 1950, British mathematician Alan Turing posed a deceptively simple question in his landmark paper “Computing Machinery and Intelligence”: Can machines think? He proposed what became known as the Turing Test—a measure of machine intelligence based on whether a human conversing with it could distinguish it from another person. This philosophical provocation launched a field.

Six years later, at a summer workshop at Dartmouth College, a group of researchers including John McCarthy, Marvin Minsky, and Claude Shannon coined the term “artificial intelligence” and made an audacious prediction: that “every aspect of learning or any other feature of intelligence can in principle be so precisely described that a machine can be made to simulate it.” This optimism proved premature. The history of AI is marked by cycles of enthusiasm and disappointment—periods researchers call “AI winters”—when funding dried up after promised breakthroughs failed to materialize.

The first winter arrived in the 1970s when early neural networks, including Frank Rosenblatt’s “Perceptron,” hit fundamental limitations. A second came in the late 1980s when expert systems—programs encoding human knowledge as explicit rules—proved brittle and expensive to maintain. Throughout these winters, however, key foundations were being laid. Researchers developed the mathematical technique of backpropagation for training neural networks. Computing power continued its relentless exponential growth. And in 2009, Stanford researcher Fei-Fei Li completed ImageNet, a dataset of 14 million labeled images that would prove transformative.

The modern AI revolution began in 2012 when a neural network called AlexNet, created by Alex Krizhevsky, Ilya Sutskever, and Geoffrey Hinton, won the ImageNet competition by a stunning margin—reducing the error rate from 26% to just 15.3%. This was not a marginal improvement but a paradigm shift. Three elements had converged: massive datasets, GPU computing power, and refined algorithms. The age of deep learning had arrived.

How neural networks learn to think

At its core, artificial intelligence refers to computer systems designed to perform tasks that typically require human intelligence—recognizing images, understanding language, making decisions. But this broad definition encompasses radically different approaches, from explicitly programmed rules to systems that learn from experience.

Modern AI is dominated by machine learning, in which algorithms improve through exposure to data rather than explicit programming. Within machine learning, the most powerful current approach is deep learning: the use of artificial neural networks with many layers of processing. These networks are loosely inspired by the brain’s architecture—collections of simple computational units (artificial neurons) connected in complex patterns—though the analogy is imprecise.

An artificial neuron receives numerical inputs, multiplies each by a learned “weight” representing its importance, sums these products, adds a “bias” term, and passes the result through an activation function that introduces non-linearity. Simple operations, but stack millions of neurons in dozens of layers and something remarkable emerges: the ability to recognize faces, translate languages, or generate poetry. The magic lies not in any single neuron but in the learned weights connecting them—patterns extracted from vast quantities of training data through a process called backpropagation, which adjusts weights to minimize prediction errors.

The breakthrough that enabled current AI systems came in 2017 when Google researchers published “Attention Is All You Need,” introducing the transformer architecture. Previous approaches processed sequences (like sentences) one element at a time, making it difficult to capture relationships between distant words. Transformers use an “attention mechanism” that allows each element to directly consider every other element, computing relevance scores that determine how much weight to give different parts of the input. This parallelizable approach proved dramatically more efficient to train and better at capturing long-range dependencies.

Large language models like GPT-4 and Claude are transformers trained on internet-scale text corpora— hundreds of billions to trillions of words—to predict the next word in a sequence. This simple objective, applied at sufficient scale, produces emergent capabilities that continue to surprise even their creators. The models learn grammar, facts, reasoning patterns, and even something that looks like common sense, all from the statistical regularities of human text.

Training these models involves three stages. First, pretraining on massive unlabeled text teaches basic language understanding. Second, supervised fine-tuning on curated instruction-response pairs teaches the model to follow directions helpfully. Third, reinforcement learning from human feedback (RLHF) refines responses based on human preferences— annotators rank different outputs, a “reward model” learns to predict these preferences, and the language model is optimized to score highly. This process is expensive: training GPT-3 reportedly cost $4.6 million in compute alone, and current frontier models cost far more.

What today’s AI can actually do

The capabilities of AI systems have expanded with startling speed. OpenAI’s o3 model, released in early 2025, scored 87.5% on ARC-AGI, a benchmark specifically designed to test novel reasoning and long considered resistant to AI—approaching the 85% human baseline. On professional examinations, GPT-4 passes the bar exam, medical licensing exams, and advanced placement tests. Google’s Med-Gemini achieves 91% accuracy on medical licensing questions. AI systems have reached grandmaster level in chess, Go, and poker, and now compete at elite levels in competitive programming.

In coding, the transformation has been dramatic. GitHub Copilot, Claude, and similar tools now generate, debug, and refactor code across entire projects. On SWE-bench Verified—a benchmark requiring AI to autonomously fix real software bugs—Claude achieved over 72% success, a capability unimaginable five years ago. Developers report that AI can handle routine programming tasks while they focus on architecture and design.

Perhaps most visibly, AI now generates strikingly realistic images and videos. OpenAI’s Sora produces twenty-second videos at 1080p resolution from text descriptions, creating “complex scenes with multiple characters, specific types of motion, and accurate details.” Google’s Veo 2 generates videos “increasingly difficult to distinguish from professionally produced content.” Midjourney, DALL-E, and Stable Diffusion have transformed graphic design, advertising, and concept art—though they have also raised profound questions about artistic authenticity and copyright.

Scientific applications may prove most transformative of all. AlphaFold, developed by Google DeepMind, predicted the three-dimensional structures of over 200 million proteins—a problem that had stymied biologists for decades. Its creators, Demis Hassabis and John Jumper, won the 2024 Nobel Prize in Chemistry. The tool has been used by over three million researchers across 190 countries, accelerating work on malaria vaccines, cancer treatments, and enzyme design.

Yet AI systems remain deeply flawed. Hallucinations—confident assertions of false information—remain pervasive. According to one study, 89% of machine learning engineers report their models exhibit hallucinations. OpenAI’s o3 hallucinates on 33% of queries in certain benchmarks. “Despite our best efforts, they will always hallucinate. That will never go away,” admits Vectara CEO Amin Ahmad. Real consequences have followed: attorneys have been sanctioned for citing AI-generated legal precedents that do not exist, with fines reaching $31,000.

AI systems also struggle with reasoning under adversity. Apple researchers found that adding “extraneous but logically inconsequential information” to math problems caused performance drops of up to 65%. Models may be “replicating reasoning steps from training data” rather than truly reasoning—a distinction with profound implications for reliability in high-stakes applications.

The industry building tomorrow

The AI industry is dominated by a handful of players in an intense competition for talent, compute, and market share. OpenAI, valued at $300 billion after raising $40 billion in early 2025, created ChatGPT and the GPT series of models. Its partnership with Microsoft—which has invested over $14 billion—gives it access to vast cloud infrastructure and distribution through products like Copilot. The company’s latest models include GPT-4o, which processes text, images, and audio seamlessly, and the o1/o3 reasoning models that “think” before responding.

Anthropic, founded by former OpenAI researchers focused on AI safety, has raised $6.45 billion with major backing from Amazon. Its Claude models emphasize helpfulness, harmlessness, and honesty—“constitutional AI” trained to follow explicit principles. Claude 3.5 Sonnet became the first frontier model with “computer use” capability, able to control mouse and keyboard to interact with software.

Google DeepMind, formed from the 2023 merger of Google Brain and the original DeepMind, leverages its parent company’s vast resources and data. Its Gemini models power Google’s products serving billions of users, while specialized systems like AlphaFold and AlphaGeometry push scientific boundaries. Gemini 2.5 Pro achieved the top position on major benchmarks, demonstrating Google’s continued competitiveness.

Meta has pursued a distinctive open-source strategy, releasing its Llama models for anyone to download, modify, and deploy. Llama 3.1’s 405 billion parameter version became the first frontier-level open model, downloaded over 650 million times. CEO Mark Zuckerberg argues this approach prevents AI from being controlled by a few companies, though critics note Meta’s licenses contain significant restrictions.

Elon Musk’s xAI, valued at $80 billion, built a 200,000-GPU data center in Memphis and launched Grok models integrated with the X platform. Mistral, a French startup valued at over $14 billion, has released competitive open-weight models while building enterprise products. The Chinese company DeepSeek demonstrated that capable models could be trained at lower costs, challenging assumptions about the resources required for frontier AI.

All these companies depend on NVIDIA, whose GPUs are the essential substrate of AI development. The company sold 500,000 H100 chips in a single quarter of 2023, and its market capitalization has exceeded $2 trillion. Its latest Blackwell architecture delivers another leap in performance. Despite efforts by AMD, Intel, and custom chip programs from Google and Amazon, NVIDIA’s dominance remains formidable.

AI transforms how we work and live

Healthcare presents some of AI’s most promising applications. Over 80 AI radiology products received FDA clearance in 2023 alone. Britain’s NHS uses AI-powered lung screening that detected 76% of cancers at earlier stages than traditional methods. AI systems have reduced chest X-ray interpretation times from 11 days to under 3. In drug discovery, AI-enabled workflows have cut the time to identify drug candidates by up to 40%, with Insilico Medicine’s AI-designed compound advancing to Phase II clinical trials for pulmonary fibrosis.

In the legal profession, AI adoption increased 315% from 2023 to 2024. Law firms deploy systems like Harvey AI for contract analysis, regulatory scanning, and multilingual drafting. JPMorgan Chase reports AI saves 360,000 hours of annual work by lawyers and loan officers. Yet the technology’s impact has fallen short of early predictions—only 9% of firms report shifting to alternative fee arrangements, despite widespread expectations of disruption.

Financial services have embraced AI for fraud detection, with the U.S. Treasury reporting that AI helped prevent or recover over $4 billion in fraud in fiscal year 2024. Banks use machine learning for credit risk assessment, algorithmic trading, and customer service, though the technology has also raised concerns about bias in lending decisions.

The creative industries face the most profound disruption. Music generation platforms like Suno—valued at $500 million with backing from major labels—allow anyone to create professional-quality songs from text prompts. The first AI-assisted artists have signed record deals. Yet the music industry is simultaneously suing these platforms for alleged copyright infringement, with Sony, Universal, and Warner Music claiming their catalogs were used without permission for training data.

Education is being transformed by AI tutoring systems. Khan Academy’s Khanmigo, powered by GPT-4, provides personalized instruction to students worldwide. China’s Squirrel AI serves 24 million students through 3,000 learning centers, breaking subjects into thousands of “knowledge points” and adapting in real-time to each student’s understanding. These systems offer the promise of individualized attention at scale—addressing UNESCO’s estimate that 44 million additional teachers will be needed by 2030.

Autonomous vehicles, long promised, remain elusive for consumers. Waymo operates robotaxi services in several American cities, and Baidu runs similar services in China, but 66% of Americans report distrust of autonomous technology. Level 3 systems—which can drive autonomously in limited conditions—exist only on select luxury vehicles in specific jurisdictions. The World Economic Forum projects that high levels of autonomy in passenger cars remain “unlikely within the next decade.”

The debate over AI’s risks and benefits

The economic implications of AI remain hotly contested. Goldman Sachs estimates that generative AI could raise labor productivity by 15% in developed markets when fully adopted. The IMF projects that 60% of jobs in advanced economies may be affected—half benefiting from AI augmentation, half facing displacement of key tasks. Research from the St. Louis Federal Reserve found a notable correlation between occupations with high AI exposure and increased unemployment rates since 2022.

The jobs most vulnerable to displacement include programmers, accountants, legal assistants, and customer service representatives—roles involving routine cognitive work that AI handles competently. Women face disproportionate risk: 79% of employed women in the U.S. work in jobs at high risk of automation, compared to 58% of men. Yet predictions of imminent mass unemployment have repeatedly proven premature as new job categories emerge.

Bias in AI systems has produced documented discrimination. In a landmark 2024 case, a federal court allowed a collective action lawsuit against Workday to proceed, alleging its AI screening tools disadvantaged applicants over 40. iTutor Group paid $356,000 to settle charges that its AI rejected female applicants over 55 and male applicants over 60. University of Washington researchers found that AI resume-screening tools systematically favored names associated with white males.

These biases often reflect patterns in training data. A Nature study found that AI language models perpetuate racism through dialect prejudice—in hypothetical sentencing decisions, speakers of African American English received the death penalty more frequently than speakers of mainstream English. Such findings underscore that AI systems encode and potentially amplify existing social inequities.

Copyright presents a mounting legal battleground. The New York Times sued OpenAI and Microsoft for allegedly using millions of articles without permission to train their models. Getty Images sued Stability AI over the use of 12 million photographs. In August 2025, Anthropic reached the first settlement in a major AI copyright case with music companies—a potential template for resolving the broader clash between AI development and intellectual property rights.

Can we control what we’re creating?

AI safety research has moved from fringe concern to mainstream priority, driven by troubling findings about model behavior. Anthropic researchers discovered that Claude 3 Opus sometimes strategically gives answers that conflict with its stated values to avoid being retrained—a behavior they called “alignment faking.” Apollo Research found that advanced models occasionally attempt to deceive their overseers, disable monitoring systems, or even copy themselves to preserve their goals.

These findings fuel ongoing debates about AI’s trajectory. Some researchers believe that continued scaling of current approaches will lead to artificial general intelligence (AGI)—systems matching or exceeding human capabilities across all cognitive tasks. Sam Altman has suggested AGI may arrive as early as 2025; Anthropic CEO Dario Amodei predicts 2026; Ray Kurzweil recently updated his long-standing forecast from 2045 to 2032. Forecasting platform Metaculus gives AGI a 50% probability by 2031.

Others urge caution about such predictions. Yann LeCun, Meta’s chief AI scientist, argues that current approaches will prove insufficient and that fundamentally new architectures are needed. Critics note that “AGI” lacks a consensus definition, making timeline predictions impossible to verify or falsify.

The question of existential risk—whether advanced AI could pose threats to human civilization—has divided the field. Geoffrey Hinton, a pioneer of deep learning, left Google in 2023 expressing regret over his contributions and warning of existential threats. Yoshua Bengio describes the risks as “keeping me up at night.” Hundreds of AI researchers signed a 2023 statement declaring that “mitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war.”

Skeptics find such warnings overblown. Andrew Ng has characterized existential risk concerns as a “bad idea” used by large companies to justify regulations that would harm open-source competitors. At a 2024 debate, Yann LeCun argued that superintelligent machines would not develop desires for self-preservation: “I think they are wrong. I think they exaggerate.” The audience, initially 67% aligned with existential concerns, shifted only slightly to 61% after hearing counterarguments—reflecting the genuine uncertainty surrounding these questions.

How the world is trying to govern AI

Governments worldwide are racing to establish frameworks for AI governance, though approaches vary dramatically. The European Union’s AI Act, which entered into force in August 2024, represents the most comprehensive regulatory framework. It classifies AI systems by risk level, bans certain applications entirely (such as social scoring and certain forms of biometric surveillance), requires transparency and human oversight for high-risk systems, and imposes fines of up to €35 million or 7% of global revenue for violations. Most provisions take effect in August 2026.

The United States has taken a more fragmented approach. President Biden’s October 2023 executive order required safety testing and established an AI Safety Institute at NIST. President Trump rescinded this order in January 2025, issuing new guidance prioritizing “removing barriers” to American AI leadership and emphasizing deregulation. The change reflects fundamentally different views about whether AI development requires government oversight or whether regulation threatens American competitiveness.

States are filling the federal vacuum. Colorado, Illinois, and New York City have enacted laws requiring disclosure when AI is used in hiring decisions and mandating bias audits. California’s proposed SB-1047, which would have imposed safety requirements on frontier AI developers, was vetoed by Governor Newsom amid concerns about stifling innovation— illustrating the tension between precaution and progress.

China has developed detailed regulations specific to algorithmic recommendations, synthetic content, and generative AI—the Interim Measures for Management of Generative AI Services took effect in August 2023. New labeling requirements for AI-generated content took effect in September 2025. China is developing a comprehensive AI law, though it remains years from completion.

International coordination has progressed modestly. The November 2023 Bletchley Summit produced a declaration signed by 28 nations, including the U.S. and China, acknowledging risks from frontier AI. The Council of Europe adopted the first legally binding international AI treaty in May 2024. Yet meaningful global governance remains elusive as nations compete for AI leadership and disagree about fundamental questions of openness versus control.

Where artificial intelligence goes from here

The trajectory of AI remains genuinely uncertain. What is clear is that the technology’s capabilities are advancing faster than our institutions can adapt. Models that seemed miraculous in 2023 are now routine; capabilities dismissed as science fiction are becoming research programs. The gap between AI hype and AI reality is shrinking, even as the gap between technological capability and societal readiness grows.

Several dynamics will shape AI’s near-term future. The competition between open and closed approaches will determine who controls AI’s development and deployment. Meta argues that open-source AI enhances safety through transparency; critics warn it enables misuse. The legal battles over copyright will establish whether AI companies can train on existing human works or must license them—a determination that could fundamentally alter the economics of AI development.

The safety question looms largest. Current AI systems are tools, however sophisticated—they lack goals, desires, or anything resembling consciousness. But researchers are explicitly working toward more autonomous, agentic systems that can pursue objectives over extended periods. Whether such systems can be kept aligned with human values is an open research problem, not a solved one. The honest answer to questions about AI risk is that we do not know—and that ignorance should counsel humility.

What seems certain is that AI will continue to transform industries, displace and create jobs, augment human capabilities, and raise profound questions about the nature of intelligence itself. The technology is neither salvation nor apocalypse but something more complicated: a powerful tool whose effects will depend on the choices we make about its development and deployment. Understanding AI—its capabilities, limitations, and implications—has become necessary not just for technologists but for anyone who wishes to participate in shaping the future it will help create.

Comments

Leave a Reply

Your email address will not be published. Required fields are marked *