|
Dear friends,
Another year of rapid AI advances has created more opportunities than ever for anyone — including those just entering the field — to build software. In fact, many companies just can’t find enough skilled AI talent. Every winter holiday, I spend some time learning and building, and I hope you will too. This helps me sharpen old skills and learn new ones, and it can help you grow your career in tech.
To be skilled at building AI systems, I recommend that you:
Let me share why each of these is important.
At the same time, taking courses alone isn’t enough. There are many lessons that you’ll gain only from hands-on practice. Learning the theory behind how an airplane works is very important to becoming a pilot, but no one has ever learned to be a pilot just by taking courses. At some point, jumping into the pilot's seat is critical! The good news is that by learning to use highly agentic coders, the process of building is the easiest it has ever been. And learning about AI building blocks might inspire you with new ideas for things to build. If I’m not feeling inspired about what projects to work on, I will usually either take courses or read research papers, and after doing this for a while, I always end up with many new ideas. Moreover, I find building really fun, and I hope you will too!
Love, Andrew
Top AI Stories of 2025
A New Era Dawns
2025 may be remembered as the dawn of AI’s industrial age. Innovations propelled model performance to new heights, AI-driven applications became indispensable, top companies battled over skilled practitioners, and construction of infrastructure drove the U.S. gross domestic product. As in past winter holiday seasons, this special issue of The Batch traces the major themes of the past 12 months. The coming year promises to consolidate these changes as we weave the technology more tightly into the fabric of daily life.
Thinking Models Solve Bigger Problems
Think step by step. Explain your reasoning. Work backwards from the answer. As 2025 began, models executed these reasoning strategies only when prompted. Now most new large language models do it as a matter of course, improving performance across a wide range of tasks.
What happened: Late last year, OpenAI introduced the first reasoning, or “thinking,” model, o1, which baked in an agentic reasoning workflow. In January, DeepSeek-R1 showed the rest of the world how to build such capabilities. The result: immediate improvements in math and coding performance, more accurate answers to questions, more capable robots, and rapid progress in AI agents.
Driving the story: An early form of reasoning took off with “Large Language Models Are Zero-Shot Reasoners,” the paper that introduced the prompt addendum, “let’s think step by step.” The authors found that manually adding these words to a prompt improved a model’s output. Researchers soon realized they could train this capability into models so they would employ this and other reasoning strategies without explicit prompting. The key: fine-tuning via reinforcement learning (RL). Giving a pretrained LLM a reward for producing correct output trained it to “think” things through before it generated output.
Yes, but: Reasoning models may not be as rational as they seem.
Where things stand: Reasoning dramatically improves LLM performance. However, better output comes at a cost. Gemini 3 Flash with reasoning enabled used 160 million tokens to run the benchmarks in Artificial Analysis’ Intelligence Index (and achieved a score of 71), while Gemini 3 Flash without reasoning used 7.4 million tokens (achieving a much lower score of 55). Moreover, generating reasoning tokens can delay output, adding to pressure on LLM inference providers to serve tokens faster. But researchers are finding ways to make the process more efficient. Claude Opus 4.5 and GPT-5.1 set to high reasoning achieve the same Intelligence Index score, but the former uses 48 million tokens, while the latter uses 81 million.
Big AI Lures Talent With Huge Pay
Leading AI companies fought a ferocious war for talent, luring top talent from competitors with levels of compensation more commonly associated with pro sports.
What happened: In July, Meta launched a hiring spree to staff the new Meta Superintelligence Labs, offering up to hundreds of millions of dollars to researchers from OpenAI, Google, Anthropic, and other top AI companies. The offers included large cash bonuses and compensation for equity forfeited by leaving another company. Meta’s rivals, in turn, poached key employees from Meta and each other, driving up the market value of AI talent to unprecedented levels.
Driving the story: Meta upended traditional pay structures by offering pay packages worth as much as $300 million over four years with liquid compensation that sometimes vastly exceeded the stock options that, at other companies, vest over many years. Having hired Scale AI CEO Alexandr Wang and key members of his team, Meta chief Mark Zuckerberg compiled a wish list, The Wall Street Journal reported.
Behind the news: The trajectory of salaries for AI engineers reflects AI’s evolution from academic curiosity to revolutionary technology.
Where things stand: As 2026 begins, the AI hiring landscape is much changed. To fend off recruiters, OpenAI has offered more stock-based compensation than its competitors, accelerated the vesting schedule for stock options awarded to new employees, and handed out retention bonuses as high as $1.5 million, The Wall Street Journal reported. Despite talk of an AI bubble in 2025, high salaries are rational for companies that plan to spend tens of billions of dollars to build AI data centers: If you’re spending that much on hardware, why not spend a small percentage of the outlay on salaries?
Data-Center Buildout Goes Big
Top AI companies announced plans to build data centers that are projected to burn through trillions of dollars and gigawatts of electricity in the next few years.
What happened: The AI industry’s capital spending topped $300 billion this year alone, much of it allocated to building new data centers to process AI. This was a preliminary budget, as companies mapped out ambitious plans to construct facilities the size of small towns with the energy needs of medium-size cities. The race to build enough processing power to satisfy hoped-for demand for inference and training could cost $5.2 trillion by 2030, the consultancy McKinsey & Company projected.
Driving the story: Top AI companies announced a cascade of data-center projects across the world. Each gigawatt of data-center capacity will cost roughly $50 billion to build.
Yes, but: Can the U.S. economy and infrastructure support such immense investments? There are reasons to wonder.
Where things stand: Despite concerns about an AI bubble, the boom in building infrastructure is generating real jobs and sales in an otherwise tepid economy. Investment in data centers and AI accounted for nearly all the growth of the U.S. gross domestic product in the first half of 2025, according to Harvard economist Jason Furman. At this stage, there is evidence to back the idea that 2025 lifted the curtain on a new industrial age.
Agents Write Code Faster, Cheaper
Coding apps moved beyond autofill-style code completion to agentic systems that manage a wide range of software development tasks.
What happened: Coding emerged as the application of agentic workflows with the most immediate business value. Claude Code, Google Gemini CLI, OpenAI Codex and other apps turned coding agents into one of Big AI’s fiercest competitive battlegrounds. Smaller competitors developed their own agentic models to remain in the game.
Driving the story: When Devin, the pioneering agentic code generator, arrived in 2024, it raised the state of the art on the SWE-Bench benchmark of coding challenges from 1.96 percent to 13.86 percent. In 2025, coding agents that use the latest large language models routinely completed more than 80 percent of the same tasks. Developers embraced increasingly sophisticated agentic frameworks that enable models to work with agentic planners and critics, use tools like web search or terminal emulation, and manipulate entire code bases.
Behind the news: Agentic systems steadily ratcheted up the state of the art on the popular SWE-Bench coding benchmark, and researchers looked for alternate ways to evaluate their performance.
Yes, but: At the beginning of 2025, most observers agreed that agents were good for generating run-of-the-mill code, documentation, and unit tests, but experienced human engineers and product managers performed better on higher-order strategic problems. By the end of the year, companies reported automating senior-level tasks. Microsoft, Google, Amazon, and Anthropic said they were generating increasing quantities of their own code.
Where things stand: In a short time, agentic coding has propelled vibe-coding from puzzling buzzword to burgeoning industry. Startups like Loveable, Replit, and Vercel enable users who have little or no coding experience to build web applications from scratch. While some observers worried that AI would replace junior developers, it turns out that developers who are skilled at using AI can prototype applications better and faster. Soon, AI-assisted coding may be regarded as simply coding, just as spellcheck and auto-complete are part of writing.
China’s AI Chip Industry Takes Root
The United States government’s effort to deprive China of AI computing power backfired as China turned the tables and banned U.S.-designed chips.
What happened: China’s government issued a directive that all new state-funded data centers must be built using chips made by domestic suppliers, Reuters reported in November. The policy emerged shortly after the U.S. reversed its years-long ban on sales to China of advanced chips manufactured using U.S. technology, including products of AI-chip leader Nvidia and its rival AMD. Rather than constrain China, U.S. policies spurred investment and innovation in China’s semiconductor industry.
Driving the story: The U.S. government aimed to block China’s access to AI on the belief that the technology would have as much geopolitical importance as oil. President Trump took a hard line in his first term from 2017 to 2020, limiting China’s access to cutting-edge technology, and he doubled down on this policy throughout 2025. However, the U.S. approach looked increasingly untenable as China’s semiconductor industry made surprising progress, the immense economic value of the AI chip market became clear, and trade restrictions locked Nvidia, now one of the world’s most valuable companies, out of its largest potential market.
Behind the news: U.S. efforts met with some success denying China access to the latest chip-manufacturing equipment. However, with respect to the chips themselves, the U.S. barriers proved to be porous.
Where things stand: China is signaling that it’s willing to do without American hardware. That may be a sign of confidence, given Huawei’s progress. It may also be a bluff, as some authorities evaluate China’s semiconductor industry still to be years behind the frontier of high-volume chip fabrication. Either way, the hard-line U.S. strategy backfired, and its relaxation of trade restrictions is a concession to economic and diplomatic realities.
Work With Andrew Ng
Join the teams that are bringing AI to the world! Check out job openings at DeepLearning.AI, AI Fund, and Landing AI.
Subscribe and view previous issues here.
Thoughts, suggestions, feedback? Please send to thebatch@deeplearning.ai. Avoid our newsletter ending up in your spam folder by adding our email address to your contacts list.
|