• D.C.
  • BXL
  • Lagos
  • Riyadh
  • Beijing
  • SG
  • D.C.
  • BXL
  • Lagos
Semafor Logo
  • Riyadh
  • Beijing
  • SG


In today’s edition, we have a scoop on the theoretical idea of improving AI models that exhibit bias͏‌  ͏‌  ͏‌  ͏‌  ͏‌  ͏‌ 
 
rotating globe
May 1, 2024
semafor

Technology

technology
Sign up for our free email briefings
 
Reed Albergotti
Reed Albergotti

Hi, and welcome back to Semafor Tech.

Katyanna has an interesting scoop today on Amazon’s efforts to surgically remove data from foundation models. One current problem with Anthropic’s Claude 3 Opus, OpenAI’s GPT-4, and Google’s Gemini Ultra is that they are so vast and complex that they can’t be simply modified like software with a bug.

Instead, the methods used to keep these models in check involve more fine tuning, like reinforcement learning with human feedback, where people rate the model’s responses on certain topics.

Those methods, while effective, are broad-based tools. It doesn’t do much to solve specific problems after a model has been trained. For instance, say you don’t want DALL-E to create anything that resembles Picasso’s art. You could remove all Picasso paintings from the data set and re-train the model. But that would cost millions or potentially tens of millions of dollars, and also make DALL-E less useful.

Enter Amazon’s new idea, which it calls “model disgorgement.” As Katyanna discovered, this is a way to go into an already-trained model and pluck out Picasso, without it affecting the whole system. It sounds simple, but it’s not.

You can read below for more details, but there’s a broader takeaway: The big breakthrough in generative AI happened so recently that it’s going to be a while before we learn how to harness it effectively. Smaller developments like “model disgorgement” are happening every day and will be essential to turning these things into reliable products.

Move Fast/Break Things

➚ MOVE FAST: Asia. Indonesia snagged a $1.7 billion AI infrastructure investment from Microsoft, Thailand will get its first Azure data center, and Elon Musk’s courtship of China is in hyperdrive. His recent trip there and an expanded partnership with Chinese tech firm Baidu were rewarded with a tentative green light for Tesla’s self-driving services to be offered there.

➘ BREAK THINGS: Europe. Ericsson CEO Börje Ekholm said the continent is on its way to becoming “a museum” because of its regulations. His comments come just days after the head of the $1.6 trillion Norwegian sovereign wealth fund noted that “America has a lot of AI and no regulation, Europe has no AI and a lot of regulation.”

Joan Cros Garcia/Corbis/Getty Images
PostEmail
Artificial Flavor

Anthropic, the AI startup that spun out of OpenAI, has always been a bit slower to roll out consumer products. That’s in part because caution runs through its DNA. Its founders left OpenAI partly because of concerns about AI safety.

That’s why today feels a little bit like a consumer coming out party for Anthropic. It announced two new products — an iOS app and a product called “Team Plan” that looks a bit like Google Docs, but for generative AI.

Scott White, a product lead on both projects, told Semafor that the mobile app is a continuation of a consumer trend that began with the web version of its chatbot. “As we launched Claude.ai, we saw a lot of the traffic coming from mobile web, and people using it on the go,” he said. “People were telling us they wanted mobile.”

White said the Team Plan, which costs $30 per month per user, will be useful for a wide range of companies to quickly access Anthropic’s AI tools without much of a learning curve. “Financial services institutions might want to take publicly accessible documents and put them into Claude to help them understand what materials they can use to reach out to private clients about what’s going on in the market,” he said. “We see this as being a very horizontal technology, which is capable of solving a really wide range of problems.”

Anthropic
PostEmail
Katyanna Quach

‘Disgorgement’: A way to get rid of bad AI data

Annegret Hilse/Reuters/File Photo

THE SCOOP

Researchers at Amazon Web Services have come up with new ways to scrub bad data from an AI model that can lead to bias, data privacy violations, or copyright infringement.

The idea, if it works, would be a major development in allowing companies to revise models once they’ve been trained. It could also help them better comply with rules to protect private information or intellectual property, like Europe’s General Data Protection Regulation, which includes the right to be forgotten.

Neural networks, like generative AI models, are trained to perform specific tasks by learning complex patterns from data. In some cases, however, developers may want to remove some data from the model if it exhibits incorrect or harmful behaviors. A company might also want to block AI from copying artists’ work, disclosing sensitive documents, or generating false information for example. But it’s difficult to remove these deleterious effects; they either have to take the model down or retrain it from scratch on better data, which is expensive.

“Even with the most careful controls, when you are training models with trillions of pieces of data, there could be mistakes. So we need to be able to plan ahead to know what to do when these mistakes are revealed,” Stefano Soatto, vice president of AWS AI Labs and a computer science professor at the University of California, Los Angeles, told Semafor in an interview. “Right now, the solution is to throw everything away and learn from scratch, which is quite costly and impacts energy and the environment. It’s not just a matter of masking the result, we have to remove or disgorge the information from the train models.”

Dubbed “model disgorgement,” AWS researchers have been experimenting with different computational methods to try and remove data that might lead to bias, toxicity, data privacy, or copyright infringement. They outlined different techniques in a paper published in the Proceedings of the National Academy of Sciences last month, including splitting the training data into “shards” so it’s easier to delete a specific chunk or use synthetic data.

These methods have yet to be applied internally to any commercial models. Soatto said it’s “still early days” but may eventually be a feasible solution to fix issues after they’ve been deployed in the real world.

Katyanna’s view on how disgorgement could help AI companies avoid legal woes. →

PostEmail
Friends of Semafor

Want a deeper, more critical look at the tech industry? Ed Zitron’s Where’s Your Ed At weekly newsletter offers an unfiltered glance at the tech industry’s growing influence and manipulation of society, interrogating the growth-at-all-costs future that tech’s elite wants to build. Subscribe for free here.

PostEmail
Evidence

As much as cyber attacks and defenses have become automated, and AI is exacerbating fears, Verizon’s latest report on hacks shows how much humans still matter. Its 2024 Data Breach Investigations Report, released today, shows that more than 70% of incidents that involve social engineering are based on pretexting (hackers using existing email chains to convince a person to, say, update a bank account with a deposit) or phishing. A bright spot is that people are getting better at spotting such ruses during internal tests, with more than 20% of users reporting phishing attempts.

PostEmail
What We’re Tracking

The Financial Times is going against the grain by signing a licensing deal with OpenAI. Announced earlier this week, the deal gives the startup rights to FT content for training purposes and allows its ChatGPT service to summarize articles.

Meanwhile, companies like The New York Times and MediaNews Group are suing OpenAI for copyright infringement, highlighting the rift in opinions on how best to deal with the generative AI boom.

This could be a smart move by the FT. It’s fairly likely that news content and other internet data is at its peak value right now, as tech giants spend tens of billions in a high stakes race to improve AI models. As that happens, they’ll begin generating better synthetic data that can be used for training purposes.

Tech companies will also find ways to unlock data sources nobody thought of. Every company of any significant size has so much unstructured data tucked away in disparate storage silos.

By the time media companies settle their lawsuits with OpenAI, Microsoft, Meta and others, their precious data may be less valuable for AI’s purposes.

PostEmail
Obsessions
Rabbit

Launched in January, the appeal of the Rabbit r1, a small voice-activated device, was its simplicity. Users could talk into a microphone, and their speech would be transcribed and fed into a large language model that would then carry out tasks, like ordering an Uber or takeout. The idea was that people wouldn’t have to scroll through their phones and interact with apps; they could just talk to the Rabbit r1.

But the gadget that was designed to replace apps may just be an Android app itself. Android Authority found that it could download the software and run it on a Pixel app, which defeats the whole point of the Rabbit r1. Rabbit’s CEO Jesse Lyu, however, denied that its operating system was just an Android app, and said there were “unofficial rabbit OS app/website emulators out there” copying its product.

— Katyanna Quach

PostEmail
Hot on Semafor
PostEmail