|
Dear friends,
I recently received an email titled “An 18-year-old’s dilemma: Too late to contribute to AI?” Its author, who gave me permission to share this, is preparing for college. He is worried that by the time he graduates, AI will be so good there’s no meaningful work left for him to do to contribute to humanity, and he will just live on Universal Basic Income (UBI). I wrote back to reassure him that there will still be plenty of work he can do for decades hence, and encouraged him to work hard and learn to build with AI. But this conversation struck me as an example of how harmful hype about AI is.
AI is amazing, but it has unfortunately been hyped up to be even more amazing than it is. A pernicious aspect of hype is that it often contains an element of truth, but not to the degree of the hype. This makes it difficult for nontechnical people to discern where the truth really is. Modern AI is a general purpose technology that is enabling many applications, but AI that can do any intellectual tasks that a human can (a popular definition for AGI) is still decades away or longer. This nuanced message that AI is general, but not that general, often is lost in the noise of today's media environment.
Andrew
A MESSAGE FROM DEEPLEARNING.AIIn Design, Develop, and Deploy Multi-Agent Systems with CrewAI, created with João Moura, Co-Founder & CEO of CrewAI, you’ll learn to build production-ready teams of AI agents that plan, reason, and collaborate to automate complex workflows. Enroll now!
News
Toward Safer (and Sexier) Chatbots
Chatbot providers, facing criticism for engaging troubled users in conversations that deepen their distress, are updating their services to provide wholesome interactions to younger users while allowing adults to pursue erotic conversations.
What’s new: Character.AI, which provides chatbots designed for entertainment and companionship, temporarily barred teen users from parts of its offering and announced plans to offer a service for younger users. Meanwhile OpenAI, which faces a lawsuit on allegations that ChatGPT contributed to a teenager’s suicide, updated ChatGPT to better help users in psychological distress and reaffirmed that it would allow adult users to generate erotic content later this year.
Character.AI limits access: The startup imposed limits on young users after it received "reports and feedback from regulators, safety experts, and parents" that expressed concern about its impact on teen users, BBC News reported.
OpenAI detects distress: Around 0.15 percent of ChatGPT users — roughly 1.2 million out of the service’s 800 million weekly active users — show signs of suicidal intent and/or excessive emotional attachment to the chatbot, OpenAI revealed. The company said it has made its models more responsive to such issues, paving the way to provide interactions geared toward adults who don’t suffer from distress.
Behind the news: Both Character.AI and OpenAI were sued by families of underage users who committed suicide after they conversed with their chatbots. In the U.S., California recently passed a state law that outlaws exposing minors to sexual content and requires supporting users who are suicidal and otherwise at risk psychologically. In August, 44 state attorneys general warned xAI, Meta, and OpenAI to restrict sexually explicit material as much as possible. xAI openly embraced adult interactions in July, when it introduced sexually explicit chatbots.
Why it matters: Chatbot companionship is a growing phenomenon, and companies that offer such services — or simply chat — must be ready to manage emotional relationships between users and their software. Managing sexually charged interactions and conversations about mental illness are linked under the umbrella of building guardrails. Sycophancy also plays a role, since models that are prone to agreeing with users can encourage dangerous behavior. A depressed, underage user and a permissive chatbot make a worrisome combination.
We’re thinking: Mental health is a hard problem, in part because it affects so many people. A recent study shows that 5.3 percent of Americans had suicidal thoughts in 2024 — far higher than ChatGPT users’ 0.15 percent. It’s important that chatbot providers do what they can to help troubled users get help.
Better Images Through Reasoning
A new image generator reasons over prompts to produce outstanding pictures.
What’s new: Tencent released HunyuanImage-3.0, which is fine-tuned to apply reasoning via a variety of reinforcement learning methods. The company says this helps it understand users’ intentions and improve its output.
How it works: The authors built a training dataset of paired text and images. They trained the model on image generation via diffusion through several stages and fine-tuned it on text-to-image generation in further stages.
Results: At present, HunyuanImage 3.0 holds first place in the LMArena Text-to-Image leaderboard, ahead of Google Gemini 2.5 Flash Image (Nano Banana), Google Imagen 4.0 Ultra Generate, and ByteDance Seedream 4.0. In addition, 100 people compared 1,000 outputs of 4 competing models to those of HunyuanImage 3.0 in side-by-side contests. The people evaluated which image was better, or whether they were both equally good or equally poor.
Behind the news: Tencent has been on a streak of releasing vision models.
Why it matters: Simplifying training methods can be helpful, since each additional step adds time spent not only training but also debugging, and each additional component can interact with other components in unexpected ways, which adds to the time required to debug the system. Yet Tencent used several stages of pretraining and fine-tuning and produced a superior model.
We’re thinking: One key to this success may be to use different methods for different purposes. For instance, the team used MixGRPO to fine-tune the model for aesthetics and SRPO to better match human preferences.
Learn More About AI With Data Points!
AI is moving faster than ever. Data Points helps you make sense of it just as fast. Data Points arrives in your inbox twice a week with six brief news stories. This week, we covered Meta’s Omnilingual ASR, which transcribes speech in over 1,600 languages, and GEN-0, a new class of robotics foundation models that reveal real-world scaling laws. Subscribe today!
The Year AI Went Industrial
A year-in-review report heralds the dawn of AI’s industrial era.
How it works: The sprawling 300-slide deck highlights the year’s progress in research, industry, politics, and security.
Research: Introduced late last year, reasoning models have redefined the capabilities of large language models. OpenAI’s closed models retained their lead despite strong progress among open-weights competitors, especially China-based developers DeepSeek, Alibaba, and Moonshot. Such models showed significant gains in efficiency, shrinking numbers of trainable parameters by as much as 50 times while maintaining high performance. Models from OpenAI, Google, and Harmonic achieved gold-level performance on problems from the International Mathematical Olympiad, and the medical dialog model AIME outperformed unassisted doctors in diagnostic accuracy.
Industry: Demand for AI services mounted. According to Ramp Business Corporation, which maintains an index of AI adoption by U.S. companies, 44 percent of U.S. companies pay for AI tools, up from 5 percent in 2023. A cohort of 16 companies made nearly $18.5 billion in annualized revenue as of August, demonstrating a business case that gave some confidence to extend their financial commitments into hundreds of billions of dollars. Anticipating further growth, OpenAI and others committed to hundreds of billions of dollars to build data centers, and the availability of electrical power to drive such facilities emerged as a major issue that will shape the path forward. Among providers of closed models, OpenAI led not only in capability but also in price: GPT-5 costs 12 times less than Anthropic Claude Opus for roughly comparable performance.
Security: Cybersecurity concerns rose as one analysis estimated that offensive capabilities are doubling every 5 months. Criminals successfully used Claude Code to create false identities that gained remote employment at Fortune 500 companies, and researchers demonstrated that it’s possible to disable safety guardrails of open-weights models using minimal processing power. Anthropic and OpenAI responded to concerns that their models might be used to develop biological or chemical weapons by adopting preemptive safety measures.
Why it matters: State of AI Report 2025 brings into focus notable trends in AI over the past year and presents them with detailed context and evidence. It’s chock-full of information that weaves diverse threads into coherent lines of progress. Moreover, it provides a consistent perspective on outstanding developments from year to year.
Forecasting Multiple Time Series
Transformers are well suited to predicting future values of time series like energy prices, wages, or weather, but often — as in those examples — multiple time series often influence one another. Researchers built a model that can forecast multiple time series simultaneously.
What’s new: Chronos-2 is a pretrained model that can accept and predict multiple time series in a zero-shot manner to forecast series of a single variable (univariate forecasting), multiple variables (multivariate forecasting), and single variables that depend on other variables (covariate-informed forecasting). Its authors include Abdul Fatir Ansari, Oleksandr Shchur, Jaris Küken, and colleagues at Amazon, University of Freiburg, Johannes Kepler University Linz, Boston College, and Rutgers.
How it works: Given any number of time series, Chronos 2 predicts values at multiple future time steps. Chronos 2 learned to minimize the difference between its predicted future values and ground truth values in subsets of datasets that contain univariate series (including synthetic data generated using methods from earlier work). They supplemented these datasets with synthetic multivariate and covariate data produced using a method devised by the authors: Their method generates multiple independent time series and then produces dependencies between them by applying mathematical transformations at the same time step and across time steps.
Results: Across various benchmarks, Chronos 2 outperformed 14 competing zero-shot models according to their skill score, a measure of how much a model reduces the average difference in predicted values relative to a baseline (higher is better, one is a perfect score).
Behind the news: Most previous works, including the previous versions Chronos and Chronos-Bolt, predict only univariate time series. Later models like Toto-1.0 and COSMIC process multiple inputs or outputs in limited ways. For instance, Toto-1.0 processes multiple inputs and outputs, but the multiple inputs can only represent past information, not future or static information. COSMIC, on the other hand, can handle multiple inputs (past or future) but not multiple outputs.
Why it matters: Chronos 2 can handle past, future, and static inputs as well as multiple outputs, giving developers, researchers, and companies alike the ability to better predict complex time series.
We’re thinking: The author’s attention setup is similar to the way many video transformers apply attention separately across space and time. It saves memory compared to performing attention across both at once, and remains an effective method for understanding data across both.
A MESSAGE FROM RAPIDFIRE AI
Work With Andrew Ng
Join the teams that are bringing AI to the world! Check out job openings at DeepLearning.AI, AI Fund, and Landing AI.
Subscribe and view previous issues here.
Thoughts, suggestions, feedback? Please send to thebatch@deeplearning.ai. Avoid our newsletter ending up in your spam folder by adding our email address to your contacts list.
|