|
Dear friends,
Job seekers in the U.S. and many other nations face a tough environment. At the same time, fears of AI-caused job loss have — so far — been overblown. However, the demand for AI skills is starting to cause shifts in the job market. I’d like to share what I’m seeing on the ground.
Granted, this may grow in the future. People who are currently in some professions that are highly exposed to AI automation, such as call-center operators, translators, and voice actors, are likely to struggle to find jobs and/or see declining salaries. But widespread job losses have been overhyped.
At the same time, when companies build new teams that are AI native, sometimes the new teams are smaller than the ones they replace. AI makes individuals more effective, and this makes it possible to shrink team sizes. For example, as AI has made building software easier, the bottleneck is shifting to deciding what to build — this is the Product Management (PM) bottleneck. A project that used to be assigned to 8 engineers and 1 PM might now be assigned to 2 engineers and 1 PM, or perhaps even to a single person with a mix of engineering and product skills.
Andrew
A MESSAGE FROM DEEPLEARNING.AI“Document AI: From OCR to Agentic Doc Extraction” explores how document systems move beyond text extraction. Learn how vision-first, agentic pipelines parse PDFs into structured Markdown and JSON while preserving layouts, tables, and charts. Built in collaboration with LandingAI. Enroll today
News
Agents Unleashed
The OpenClaw open-source AI agent became a sudden sensation, inspiring excitement, worry, and hype about the agentic future.
What’s happened: In November, Developer Peter Steinberger released OpenClaw — formerly named WhatsApp Relay, Clawdbot, and Moltbot — as a personal AI agent to perform tasks like manage calendars, summarize emails, and send reminders. A post on the crowdsourced tech-news site HackerNews noted the project in late January, and it took off, garnering the fastest-growing number of GitHub stars and more Google searches than Claude Code.
How it works: OpenClaw is a configurable agentic framework that runs on a local computer or in a virtual machine in the cloud. Users can build agents to browse and write to their local file systems or operate within predefined sandboxes. They can also give agents permission to use cloud services like email, calendar, productivity applications, speech-to-text and text-to-speech applications, and virtually any service that responds to an API. Agents can use coding tools like Claude Code, interact on social networks, scrape websites, and spend money on users’ behalfs.
Yes, but: OpenClaw and Moltbook initially launched with many security flaws and other issues, some of which have been fixed at the time of this writing. The combination of an open-ended system, insecure design, and inexperienced users resulted in a variety of vulnerabilities. Misconfigured OpenClaw deployments exposed API keys, and Moltbook exposed millions more. Skills designed to perform malicious tasks, such as stealing data, have proliferated. Many users have installed the system on dedicated machines to avoid exposing private data to attackers or well-meaning but accident-prone agents.
Why it matters: OpenClaw made a huge splash and left prominent members of the AI community debating its novelty and importance. For developers, OpenClaw offers a highly customizable and powerful AI assistant that requires careful security precautions. It’s also a glimpse of a future in which autonomous agents go about their business with little input from humans.
We’re thinking: For an imaginative, enterprising open-source project, OpenClaw has inspired more than its share of hype. Press reports have likened Moltbook — which holds messages that are little different than the large language model outputs that have amazed and amused the world since GPT-3 — to the advent of AGI and the Singularity. Let us assure you that agents are not there yet, or anywhere close. Rather, OpenClaw demonstrates that agents can be immensely useful, we are still finding good use cases, and we need to pay careful attention to security. That, and you never know when one of your open-source projects might take off!
Kimi K2.5 Creates Its Own Workforce
An open source vision-language model unleashes minion agents that enable it to perform tasks more quickly and effectively.
What’s new: Moonshot AI released Kimi K2.5, an updated version of its Kimi K2 large language model that adds vision capabilities and the ability to spawn what the authors call subagents — parallel workflows that control their own separate models to execute tasks as AI research, fact checking, and web development — and assign tasks to them.
How it works: Moonshot disclosed little information about how it built Kimi-K2.5. Among the details it revealed:
Results: In the Artificial Analysis Intelligence Index, a weighted average of 10 benchmarks, Kimi K2.5 with thinking mode switched on outperformed all other open-weights models tested. In Moonshot’s tests:
Yes, but: Moonshot didn’t disclose the cost of processing and memory incurred by Kimi K2.5’s use of subagents, so the tradeoff between speed/performance and processing/memory requirements is not clear.
Behind the news: Kimi K2.5 arrives 7 months after Moonshot’s initial vision-language model, the much smaller, 16 billion-parameter Kimi-VL, which also used the MoonViT vision encoder.
Why it matters: Building an agentic workflow can improve a model’s performance on a particular task. Unlike predefined agentic workflows, Kimi K2.5 decides when a new subagent is necessary, what it should do, and when to delegate work to it. This automated agentic orchestration improves performance in tasks that are easy to perform in parallel.
We’re thinking: Kimi K2.5 shifts task execution from chain-of-thought reasoning to agentic teamwork. Instead of responding to prompts sequentially, it acts as a manager of separate workflows/models that execute different parts of the job in parallel.
Learn More About AI With Data Points!
AI is moving faster than ever. Data Points helps you make sense of it just as fast. Data Points arrives in your inbox twice a week with six brief news stories. This week, we covered Google DeepMind’s Project Genie for real-time world creation and OpenAI’s new Codex desktop app for managing coding agents. Subscribe today!
AI Giants Share Wikipedia’s Costs
On its 25th anniversary, Wikipedia celebrated with high-profile deals to make its data easier for AI companies to train their models in exchange for financial support.
What’s new: The Wikimedia foundation announced partnerships with AI companies including Amazon, Meta, Microsoft, Mistral AI, and Perplexity. The partnership program, known as Wikimedia Enterprise, lets these partners access Wikipedia data at higher speeds and volumes than they could by scraping pages on the web. Financial terms were not disclosed.
How it works: Along with donations from users, enterprise partnerships are the Wikimedia Foundation’s chief source of revenue. Wikimedia Enterprise offers APIs that enable developers to directly access encyclopedia articles and other Wikimedia data, including Wikimedia Commons images, Wiktionary’s online dictionary, and Wikidata’s machine-readable knowledge base. Free plans allow for limited data updates and access to a support portal. Paid plans (terms are not public) include daily snapshots of Wikimedia data, potentially unlimited data requests (limits vary depending on how much a subscriber pays), streaming access to real-time revisions, and technical support from human staffers.
Behind the news: Other publishers whose content is widely used to train AI systems have sought payment with varied levels of success. In 2023, Reddit and Stack Overflow announced plans to protect their data from AI crawlers while they sought licensing deals. Reddit was able to reach licensing agreements for Google, OpenAI, and others to use its content to train models. Stack Overflow saw traffic and question volume plummet, dropping from 200,000 questions per month in 2014 to 50,000 questions per month in late 2025. As its audience turned from discussing technical issues on the site to asking AI models for answers, the company pivoted from advertising as its primary revenue source to repackaging its data for AI training.
Why it matters: AI companies want to train their models on Wikipedia, and gathering data by sending API calls is much faster than crawling the web — never mind the rapid pace of crawling required to keep up with the encyclopedia’s never-ending revisions. At the same time, Wikipedia needs revenue to survive. Selling API access offers a helpful service to developers while giving this crucial data source a stronger financial foundation.
We’re thinking: These deals are win-win. People who choose to read the online encyclopedia the old-fashioned way can keep doing so, and people who build AI models can rest easier knowing they won’t kill a key source of training data.
Recipe for Smaller, Capable Models
Mistral compressed Mistral Small 3.1 into much smaller versions, yielding a family of relatively small, open-weights, vision-language models that perform better by some measures than competing models of similar size. The method combines pruning and distillation.
What’s new: Mistral AI released weights for the Ministral 3 family in parameter counts of 14 billion, 8 billion, and 3 billion. Each size comes in base, instruction-tuned, and reasoning variants. The team detailed its recipe for distilling the models in a paper.
How it works: The team built the model using an approach it calls cascade distillation. Starting with a larger parent, they alternately pruned (removed less-important parameters) and distilled (trained a smaller model to mimic the larger model's outputs) it into progressively smaller children.
Performance: Ministral 3 14B (version unspecified) ranks ahead of Mistral Small 3.1 and Mistral Small 3.2 on the Artificial Analysis Intelligence Index, a weighted average of 10 benchmarks. Mistral compared Ministral 3 with Mistral Small 3.1 and open-weights competitors of equal size. Ministral 3 14B base outperformed Mistral Small 3.1 by 1 to 12 percentage points on tests of math and multimodal understanding, and tied on Python coding. It also outperformed its parent on GPQA Diamond. Compared to open-weights competitors:
Why it matters: Cascade distillation offers a way to produce a high-performance model family from a single parent at a fraction of the usual cost. Training the Ministral 3 models required 1 trillion to 3 trillion training tokens compared to 15 trillion to 36 trillion tokens for Qwen 3 and Llama 3 models of similar sizes. Their training runs were also shorter, and their training algorithm is relatively simple. This sort of approach could enable developers to build multiple model sizes without proportionately higher training costs.
We’re thinking: Ministral 3 models can run on generic laptops and smartphones. On-device AI at the edge keeps getting more capable and competitive!
Work With Andrew Ng
Join the teams that are bringing AI to the world! Check out job openings at DeepLearning.AI, AI Fund, and Landing AI.
Subscribe and view previous issues here.
Thoughts, suggestions, feedback? Please send to thebatch@deeplearning.ai. Avoid our newsletter ending up in your spam folder by adding our email address to your contacts list.
|