Dear friends,
This week, Coursera held its annual conference in Las Vegas. A major theme was the shift from knowledge- to skills-based education, which will help many individuals, businesses, and educational institutions. This annual gathering brings together leaders from academia, business, and government to discuss the latest in education and workforce development, and I was delighted to compare notes with others about the latest developments.
While skill-based education applies to many sectors, not just engineering (you can learn skills to perform tasks in human resources, marketing, finance, and much more), it is highly relevant to AI. Skill at steering coding assistants and applying AI building blocks (like prompting, RAG, evals, and so on) lets you build more valuable software. To help learners build these kinds of applied abilities, Coursera is introducing a series of “skill tracks” programs.
Keep building! Andrew
A MESSAGE FROM DEEPLEARNING.AIAI agents often fail when they call APIs out of sequence. In this course, you’ll learn to build a knowledge graph that connects API specifications with business workflows, then build an agent that discovers the right APIs and executes them in proper order. Get started!
News
Meta, OpenAI Reinforce Guardrails
Meta and OpenAI promised to place more controls on their chatbots’ conversations with children and teenagers, as worrisome interactions with minors come under increasing scrutiny.
What’s new: Meta will update chatbots on Facebook, Instagram, and WhatsApp to avoid conversations with minors that simulate sexual attraction and to refer young users to experts rather than discuss self-harm directly. Meanwhile, OpenAI said it would route ChatGPT conversations that show acute distress to reasoning models, which are better equipped to comply with mental-health guidelines, and add parental controls. Both companies have come under intense criticism, Meta for engaging children in flirtatious conversations, OpenAI for allegedly helping a teenager to commit suicide.
How it works: Both companies announced new features intended to protect minors who use their chatbots. The changes will be implemented in coming months.
Behind the news: As users increasingly turn to chatbots as companions and counselors, they sometimes express a sycophantic attitude that may reinforce a user’s subjective perspective or even delusional perceptions. Teens and children have encountered similar behavior, sometimes with dire consequences.
What they’re saying: “One of the things that’s ambiguous about chatbots is whether they’re providing treatment or advice or companionship. . . . Conversations that might start off as somewhat innocuous and benign can evolve in various directions.” — Ryan McBain, co-author of “Evaluation of Alignment Between Large Language Models and Expert Clinicians in Suicide Risk Assessment,” assistant professor at Harvard University medical school, and senior policy researcher at RAND Corp.
Why it matters: Chatbots hold huge value for young people as study aids, information sources, counselors, and so on. Yet they need strong, well designed guardrails that can enable children to explore without exposing them to material that would interfere with their healthy development. Designing adequate guardrails is not a simple task, but it is a necessary aspect of building such applications.
We’re thinking: Suicide is a tragedy whenever it occurs, and the stories of chatbots carrying on sexual conversations with kids are deeply disturbing. Meta and OpenAI lately have strengthened their age verification procedures, and OpenAI said it analyzes conversations for signs that young people may be in crisis so the company can alert guardians and mental-health professionals. We look forward to more features that protect children and empower parents.
Google Must Share Data With AI Rivals
AI companies that aspire to compete with Google in search and other information-retrieval applications got a boost from the United States government.
What’s new: A federal court ruled that Google must turn over its current search index — a database of web links and pages — to U.S.-based AI rivals including OpenAI, Anthropic, and Perplexity as well as search engine competitors. However, the court stopped well short of the U.S. Department of Justice’s request that the company be broken up.
How it works: Last year, the same judge ruled that Google held a monopoly on web search and had acted to maintain it. In the new ruling, the judge ordered remedies to help break that monopoly, but he allowed the company to maintain its competitive position in other businesses — specifically browsers and smartphones — of interest to rival AI companies.
Behind the news: The federal government filed its antitrust case against Google in 2020, well before the 2022 launch of ChatGPT. But the subsequent emergence of generative AI dramatically changed the stakes two ways, as the judge points out in his ruling. First, AI has expanded the field of information retrieval beyond traditional search engines. Second, competitors like OpenAI loosened Google’s grip on the search business in a way Bing or DuckDuckGo had not. The court’s remedies reflect this new order: Google must share its data with competitors in AI as well as search, but more drastic remedies aren’t required, because AI has created robust competition in search. However, Google still faces potential remedies in a separate U.S. antitrust case over its online advertising business, along with a newly levied $3.5 billion fine by European antitrust courts.
Why it matters: The court’s ruling reflects the growing strength of AI companies in the business of retrieving information. However, it provides only limited openings to Google’s AI competitors and stops short of giving them broad opportunities to challenge the company. Had the judge ordered Google to sell off Chrome or Android — browsers and smartphones being major avenues that drive users to a search engine as well as opportunities for broad enhancement by AI — other AI companies would have a better shot at competing with Google Search.
We’re thinking: The judge said predicting the future of AI and search would require a crystal ball. Nonetheless, it’s already clear that large language models are taking over a significant part of the role once played by traditional search engines. Fostering competition could lead to even better products for helping users find information.
Learn More About AI With Data Points!
AI is moving faster than ever. Data Points helps you make sense of it just as fast. Data Points arrives in your inbox twice a week with six brief news stories. This week, we covered Switzerland’s launch of Apertus, a fully open source national LLM, and Anthropic’s $1.5 billion settlement with the Authors’ Guild over copyright claims. Subscribe today!
2 Hours With AI Versus 6 With Teacher
A growing private school system replaces the typical 6-hour school day with 2 hours of personalized, AI-assisted education.
What’s new: Alpha School, which teaches 250 preschool-through-high-school students in Austin, Texas, uses an AI-powered method that presents challenges that are tailored to a student’s level of mastery, doubling the speed of learning, the company claims. Students typically rank in the top 2 percent nationally on standardized tests including AP, MAP, and SAT, and last year, 11 out of 12 members of its first graduating class enrolled at universities that include Howard, Northeastern, Stanford, and Vanderbilt. In the coming year it will open locations in a dozen cities, The New York Times reported.
How it works: Alpha School doesn’t rely on teachers to deliver instruction. Instead, software leads students through 2 hours of academic exercises in math, science, reading, other language skills such as speaking and listening, and academic skills — a method the founders call 2 Hour Learning. The software automatically selects exercises to match students’ current level, and it allows them to progress to a new level only after they have demonstrated mastery of the previous one.
Yes, but: Boards of education in California, Pennsylvania, and Utah rejected charter-school applications submitted by Unbound Academy, an offshoot of Alpha School, on the ground that they failed to meet mandatory standards. Critics argue that the effectiveness of 2-Hour Learning is not supported by rigorous evidence.
Behind the news: MacKenzie Price, who has a degree in psychology, founded Alpha School in 2014 along with her husband Andrew Price, who serves as CFO of the educational software developer Trilogy. The school shifted to AI-assisted education in 2022. It’s one of several U.S. efforts to apply AI to education.
Why it matters: Primary and secondary education are among the great opportunities for AI. Alpha School has built a method and infrastructure for delivering personalized academic education in a way that enables students to learn efficiently, freeing up time for social learning and personal development.
We’re thinking: The press has spilled much ink on how to keep AI from helping students cheat. Instead, let’s focus on how AI can help students learn.
10 Million Tokens of Input Context
An alternative to attention enables large language models to track relationships among words across extraordinarily wide spans of text.
What’s new: Ali Behrouz and colleagues at Google devised a trainable component they call a memory module that stores and retrieves an input’s semantic content. The authors integrated this component into a transformer-like architecture, ATLAS, that can process up to 10 million tokens of input.
Key insight: Given a text token, a recurrent neural network computes a vector that represents it, which it updates when it receives the next token, and so on, so it remembers what it has processed so far. However, the vector may lose relevant information over many input tokens. An alternative is to dedicate a part of the network, or module, to generating a representation of the input and update its weights at inference. The module acts something like a retriever: When it receives sequences of tokens that are similar to those it received previously, it retrieves stored representations of the earlier sequence enriched with the latest context. In this way, it can interpret new input tokens in light of previous ones, like a typical recurrent neural network, without needing to examine all input tokens at once, like a transformer.
How it works: ATLAS replaces a transformer’s attention layers with a trainable memory module. The authors trained a 1.3 billion-parameter model to predict the next token in the FineWeb dataset of text from the web. During training, ATLAS learned good base values for the memory module’s weights, to be further modified at inference.
Results: The authors compared ATLAS to other models of the same size that were trained on the same number of tokens. ATLAS performed best, especially in long-context tasks.
Yes, but: The authors tested ATLAS at relatively small size 1.3 billion parameters. How it would perform at larger scales is unclear.
Why it matters: Keeping track of very long inputs remains a challenge for most LLMs, and processing more than 2 million tokens — the current limit of Google Gemini 2.5 Pro — is a wild frontier. ATLAS updates parameters at inference to maintain context through extraordinarily long inputs, potentially opening up applications that involve data-dense inputs such as video at full resolution and frame rate.
We’re thinking: ATLAS extends context to 10 million tokens — far greater than the vast majority of models. What will such very long context be useful for? How will we evaluate model performance over such long inputs? What tradeoffs come with using more tokens versus better context engineering? ATLAS may push such questions further into the foreground.
Work With Andrew Ng
Join the teams that are bringing AI to the world! Check out job openings at DeepLearning.AI, AI Fund, and Landing AI.
Subscribe and view previous issues here.
Thoughts, suggestions, feedback? Please send to thebatch@deeplearning.ai. Avoid our newsletter ending up in your spam folder by adding our email address to your contacts list.
|