Dear friends,
Last week, I returned home from Asia, where I spoke at Seoul National University in Korea, the National University of Singapore, and the University of Tokyo in Japan and visited many businesses. As I discussed the state of AI with students, technologists, executives, and government officials, something struck me: Around the world, everyone is wrestling with similar AI-related issues.
When the deep learning revolution started about a decade ago, I advised teams to (i) learn about the technology, (ii) start small and build projects quickly to hone intuition about what’s possible, and (iii) use learnings from smaller projects to scale to bigger ones. With the generative AI revolution, my advice remains the same. This time, though, the barrier to entry is lower and thus the time-to-value seems to be shorter. It takes substantial effort to collect data and train and deploy a neural network, but less effort to prompt a large language model and start getting results.
Last Friday, I discussed how businesses can plan for generative AI with Erik Brynjolfsson, Andrew McAfee, James Mili, and Daniel Rock, who co-founded Workhelix (a portfolio company of AI Fund, which I lead). Workhelix helps its customers break down jobs into tasks to see which tasks can be augmented by AI. You can listen to the conversation here.
For instance, a radiologist’s tasks include (i) capturing images, (ii) reading them, (iii) communicating with patients, and so on. Which of these tasks can take advantage of AI to make a radiologist’s job more productive and enjoyable? Can it help optimize image acquisition (perhaps by tuning the X-ray machine controls), speed up interpretation of images, or generate takeaways text for patients?
Although Workhelix is applying this recipe at scale, it’s also useful for teams that are exploring opportunities in AI. Consider not jobs but their component tasks. Are any of them amenable to automation or assistance by AI? This can be a helpful framework for brainstorming interesting project ideas.
Special thanks to Ian Park of the Korean Investment Corporation, Chong Yap Seng of the National University of Singapore, and Yuji Mano of Mitsui, who made my visits much more productive and enjoyable. I also hope to visit other countries soon. Stay tuned!
Keep learning, Andrew
P.S. DeepLearning.AI just launched “Evaluating and Debugging Generative AI,” created in collaboration with Weights & Biases and taught by Carey Phelps. Machine learning development is an iterative process, and we often have to try many things to build a system that works. I used to keep track of all the different models I was training in a text file or spreadsheet. Thankfully better tools are available now. This course will teach you how to use them, focusing on generative AI applications. I hope you enjoy the course!
NewsUkraine’s Homegrown DronesThe war in Ukraine has spurred a new domestic industry. What’s new: Hundreds of drone companies have sprung up in Ukraine since Russian forces invaded the country early last year, The Washington Post reported. How it works: Ukrainian drone startups are developing air- and sea-borne robots, which the country’s military use to monitor enemy positions, guide artillery strikes, and drop bombs, sometimes on Russian territory.
Russia responds: In recent months, Russia has stepped up attacks by Russian-made Lancet fliers that explode upon crashing into their targets. Recent units appear to contain Nvidia Jetson TX2 computers, which could drive AI-powered guidance or targeting, Forbes reported. Russian state news denied that its drones use AI. Behind the news: Other countries are also gearing up for drone warfare.
Why it matters: Drones rapidly have become a battlefield staple, and their offensive capabilities are growing. Governments around the world are paying close attention for lessons to be learned — as are, no doubt, insurgent forces, paramilitary groups, and drug cartels.
Cloud Computing Goes GenerativeAmazon aims to make it easier for its cloud computing customers to build applications that take advantage of generative AI. What’s new: Amazon Web Services’ Bedrock platform is offering new generative models, software agents that enable customers to interact with those models, and a service that generates medical records. The new capabilities are available in what Amazon calls “preview” and are subject to change. How it works: Bedrock launched in April with the Stable Diffusion image generator and large language models including AI21’s Jurassic-2 and Anthropic’s Claude. The new additions extend the platform in a few directions.
Behind the news: Amazon’s major rivals in cloud computing have introduced their own generative-AI-as-a-service offerings.
Why it matters: Access to the latest generative models is likely to be a crucial factor in bringing AI’s benefits to all industries. For Amazon, providing those models and tools to build applications on top of them could help maintain its dominant position in the market for cloud computing. We’re thinking: One challenge to startups that provide an API for generative AI is that the cost of switching from one API to another is low, which makes their businesses less defensible. In contrast, cloud-computing platforms offer many APIs, which creates high switching costs. That is, once you've built an application on a particular cloud platform, migrating to another is impractical. This makes cloud computing highly profitable. It also makes offering APIs for generative AI an obvious move for incumbent platforms.
A MESSAGE FROM DEEPLEARNING.AIJoin our new course “Evaluating and Debugging Generative AI,” and learn to manage and track data sources and volumes, debug your models, and conduct tests and evaluations easily. Sign up for free
K-Pop Sings in Many TonguesA Korean pop star recorded a song in six languages, thanks to deep learning. What’s new: Midnatt (better known as Lee Hyun) sang his latest release, “Masquerade,” in English, Japanese, Mandarin, Spanish, and Vietnamese — none of which he speaks fluently — as well as his native Korean. The entertainment company Hybe used a deep learning system to improve his pronunciation, Reuters reported. You can listen to the results here.
Behind the news: The music industry has been paying close attention to generative audio models lately, as fans have used deep learning systems to mimic the voices of established artists. Reactions from artists and music companies have been mixed.
Why it matters: This application of generated audio suggests that the technology could have tremendous commercial value. K-pop artists frequently release songs in English and Japanese, and popular musicians have recorded their songs in multiple languages since at least the 1930s, when Marlene Dietrich recorded her hits in English as well as her native German. This approach could help singers all over the world to reach listeners who may be more receptive to songs in a familiar language.
Long-Range Weather ForecastsMachine learning models have predicted weather a few days ahead of time. A new approach substantially extends the time horizon. What’s new: Remi Lam and colleagues at Google developed GraphCast, a weather-forecasting system based on graph neural networks (GNNs). Its 10-day forecasts outperformed those of conventional and deep-learning methods. GNN basics: A GNN processes input in the form of a graph made up of nodes connected by edges. It uses a vanilla neural network to update the representation of each node based on those of neighboring nodes. For example, nodes can represent customers and products while edges represent purchases, or — as in this work — nodes can represent local weather while edges represent connections between locations. Key insight: Short-term changes in the weather in a given location depend on conditions in nearby areas. A graph can reflect these relationships using information drawn from a high-resolution weather map, where each node represents an area’s weather and edges connect nearby areas. However, longer-term changes in the weather depend on conditions in both nearby and distant areas. To reflect relationships between more distant areas, the graph can draw on a lower-resolution map, which connects areas at greater distances. Combining edges drawn from higher- and lower-resolution weather maps produces a graph that reflects relationships among both nearby and distant areas, making it suitable for longer-term predictions. How it works: GraphCast produced graphs based on high- and low-resolution weather maps and processed them using three GNNs called the encoder, processor, and decoder. The authors trained the system on global weather data from 1979 to 2017. Given a set of weather conditions and a set of weather conditions measured 6 hours previously for all locations on Earth, GraphCast learned to predict the weather 6 hours in the future and multiples thereof.
Results: Using 2018 data, the authors compared GraphCast’s 10-day forecasts to those of a popular European system that predicts weather based on differential equations that describe atmospheric physics. Compared to actual measurements, GraphCast achieved a lower root mean squared error in 90 percent of predictions. It produced a 10-day forecast at 0.25-degree resolution in under 60 seconds using a single TPU v4 chip, while the European system, which forecasts at 0.1-degree resolution, needed 150 to 240 hours on a supercomputer. GraphCast also outperformed Pangu-Weather, a transformer-based method, in 99.2 percent of predictions. Yes, but: GraphCast’s predictions tended to be closer to average weather conditions, and it performed worse when the weather included extreme temperatures or storms. Why it matters: Given a graph that combines multiple spatial resolutions, GNN can compute the influence of weather over large distances using relatively little memory and computation. This sort of graph structure may benefit other applications that process large inputs such as ultra-high resolution photos, fluid dynamics, and cosmological data. We’re thinking: When it comes to forecasting weather, it looks like deep learning is the raining champ.
A MESSAGE FROM WORKERAThe best defense against disruption is your ability to take advantage of innovation. That’s why enabling employees to learn rapidly is a business imperative. Read Kian Katanforoosh's essential guide to learning velocity for business.
Work With Andrew Ng
Join the teams that are bringing AI to the world! Check out job openings at DeepLearning.AI, AI Fund, and Landing AI.
Subscribe and view previous issues here.
Thoughts, suggestions, feedback? Please send to thebatch@deeplearning.ai. Avoid our newsletter ending up in your spam folder by adding our email address to your contacts list.
|