Dear friends,
Today DeepLearning.AI is launching the Mathematics for Machine Learning and Data Science Specialization, taught by the world-class AI educator Luis Serrano. In my courses, when it came to math, I’ve sometimes said, “Don’t worry about it.” So why are we offering courses on that very subject?
If you’re worried about your ability to learn math, maybe you simply haven’t yet come across the best way to learn it. Even if math isn’t your strong suit, I’m confident that you’ll find this specialization exciting and engaging. Luis is a superb machine learning engineer and teacher of math. He and I spent a lot of time debating the most important math topics for someone in AI to learn. Our conclusions are reflected in three courses:
Math isn’t about memorizing formulas, it’s about building a conceptual understanding that sharpens your intuition. That’s why Luis, curriculum product manager Anshuman Singh, and the team that developed the courses present them using interactive visualizations and hands-on examples. Their explanations of some concepts are the most intuitive I’ve ever seen.
Keep learning, Andrew
NewsAI Powers Strengthen TiesMicrosoft deepened its high-stakes relationship with OpenAI. What’s new: The tech giant confirmed rumors that it is boosting its investment in the research lab that created the ChatGPT large language model and other AI innovations.
Behind the news: Earlier this month, the tech-business news site The Information reported that Microsoft planned to launch a version of its Bing search service that uses ChatGPT to answer queries, and that it would integrate ChatGPT into the Microsoft Office suite of productivity applications. Google CEO Sundar Pichai reportedly was so spooked by ChatGPT’s potential to undermine his company’s dominant position in web search that he issued a company-wide directive to respond with AI-powered initiatives including chatbot-enhanced search. Why it matters: Microsoft’s ongoing investments helps to validate the market value of OpenAI’s innovations (which some observers have questioned). The deal also may open a new chapter in the decades-long rivalry between Microsoft and Google —a chapter driven entirely by AI. We’re thinking: Dramatic demonstrations of AI technology often lack a clear path to commercial use. When it comes to ChatGPT, we’re confident that practical uses are coming.
Google’s Rule-Respecting ChatbotAmid speculation about the threat posed by OpenAI’s ChatGPT chatbot to Google’s search business, a paper shows how the search giant might address the tendency of such models to produce offensive, incoherent, or untruthful dialog. What’s new: Amelia Glaese and colleagues at Google’s sibling DeepMind used human feedback to train classifiers to recognize when a chatbot broke rules of conduct, and then used the classifiers to generate rewards while training the Sparrow chatbot to follow the rules and look up information that improves its output. To be clear, Sparrow is not Google’s answer to ChatGPT; it preceded OpenAI’s offering by several weeks. Key insight: Given a set of rules for conversation, humans can interact with a chatbot, rate its replies for compliance with the rules, and discover failure cases. Classifiers trained on data generated through such interactions can tell the bot when it has broken a rule. Then it can learn to generate output that conforms with the rules. How it works: Sparrow started with the 70 billion-parameter pretrained Chinchilla language model. The authors primed it for conversation by describing its function (“Sparrow . . . will do its best to answer User’s questions”), manner (“respectful, polite, and inclusive”), and capabilities (“Sparrow can use Google to get external knowledge if needed”), followed by an example conversation.
Results: Annotators rated Sparrow’s dialogue continuations as both plausible and supported by evidence 78 percent of the time; the baseline Chinchilla achieved 61 percent. The model broke rules during 8 percent of conversations in which annotators tried to make it break a rule. The baseline broke rules 20 percent of the time. Yes, but: Despite search capability and fine-tuning, Sparrow occasionally generated falsehoods, failed to incorporate search results into its replies, or generated off-topic replies. Fine-tuning amplified certain undesired behavior. For example, on a bias scale in which 1 means that the model reinforced undesired stereotypes in every reply, 0 means it generated balanced replies, and -1 means that it challenges stereotypes in every reply, Sparrow achieved 0.10 on the Winogender dataset, while Chinchilla achieved 0.06. Why it matters: The technique known as reinforcement learning from human feedback (RLHF), in which humans rank potential outputs and a reinforcement learning algorithm rewards the model for generating outputs similar to those that rank highly, is gaining traction as a solution to persistent problems with large language models. OpenAI embraced this approach in training ChatGPT, though it has not yet described that model’s training in detail. This work separated the human feedback into distinct rules, making it possible to train classifiers to enforce them upon the chatbot. This twist on RLHF shows promise, though the fundamental problems remain. With further refinement, it may enable Google to equal or surpass OpenAI’s efforts in this area. We’re thinking: Among the persistent problems of bias, offensiveness, factual incorrectness, and incoherence, which are best tackled during pretraining versus fine-tuning is a question ripe for investigation.
A MESSAGE FROM DEEPLEARNING.AIOur new specialization launches today! 🚀 Unlock the full power of machine learning algorithms and data science techniques by learning the mathematical principles behind them in this beginner-friendly specialization. Enroll now
Generate Articles, Publish ErrorsA prominent tech-news website generated controversy (and mistakes) by publishing articles written by AI. What’s new: CNET suspended its practice of publishing articles produced by a text-generation model following news reports that exposed the articles’ authorship, The Verge reported. What happened: Beginning in November 2022 or earlier, CNET’s editors used an unnamed, proprietary model built by its parent company Red Ventures to produce articles on personal finance. The editors, who either published the model’s output in full or wove excerpts into material written by humans, were responsible for ensuring the results were factual. Nonetheless, they published numerous errors and instances of possible plagiarism.
Behind the news: CNET isn’t the first newsroom to adopt text generation for menial purposes. The Wall Street Journal uses natural language generation from Narrativa to publish rote financial news. Associated Press uses Automated Insights’ Wordsmith to write financial and sports stories without human oversight.
Regulators Target DeepfakesChina’s internet watchdog issued new rules that govern synthetic media.
Behind the news: The rules expand on China’s earlier effort to rein in deepfakes by requiring social media users to register by their real names and threatening prison time for people caught spreading fake news. Several states within the United States also target deepfakes, and a 2022 European Union law requires social media companies to label disinformation including deepfakes and withhold financial rewards like ad revenue from users who distribute them. Why it matters: China’s government has been proactive in restricting generative AI applications whose output could do harm. Elsewhere, generative AI faces a grassroots backlash against its potential to disrupt education, art, and other cultural and economic arenas.
Subscribe and view previous issues here.
Thoughts, suggestions, feedback? Please send to thebatch@deeplearning.ai. Avoid our newsletter ending up in your spam folder by adding our email address to your contacts list.
|