The last few years—even the last few months—have seen artificial intelligence (AI) breakthroughs come at a dizzying pace. AI that can generate paragraphs of text as well as a human, create realistic imagery and video from text, or perform hundreds of different tasks has captured the public’s attention. People see AI’s high level of performance, creative potential and, in some cases, the ability for anyone to use them with little to no technical expertise. This wave of AI is attributable to what are known as foundation models.

What are foundation models?

As the name suggests, foundation models can be the foundation for many kinds of AI systems. Using machine learning techniques, these models apply information learned about one situation to another situation. While the amount of data required is considerably more than the average person needs to transfer understanding from one task to another, the result is relatively similar. For example, once you spend enough time learning how to cook, without too much effort you can figure out how to cook almost any dish, and even invent new ones.

This wave of AI looks to replace the task-specific models that have dominated the landscape. And the potential benefits of foundation models to the economy and society are vast. For example, identifying candidate molecules for novel drugs or identifying suitable materials for new battery technologies requires sophisticated knowledge about chemistry and time-intensive screening and evaluation of different molecules. IBM’s MoLFormer-XL, a foundation model trained on data about 1.1 billion molecules, helps scientists rapidly predict the 3D structure of molecules and infer their physical properties, such as their ability to cross the blood-brain barrier. IBM recently announced a partnership with Moderna to use MoLFormer models to help design better mRNA medicines. IBM also partners with NASA to analyze geospatial satellite data—to better inform efforts to fight climate change—using foundation models.

However, there are also concerns about their potential to cause harm in new or unforeseen ways. Some risks of using foundation models are like those of other kinds of AI, like risks related to bias. But they can also pose new risks and amplify existing risks, such as hallucination, the capability of generation of false yet plausible-seeming content. These concerns are prompting the public and policymakers to question whether existing regulatory frameworks can protect against these potential harms.

What should policymakers do?

Policymakers should take productive steps to address these concerns, recognizing that a risk and context-based approach to AI regulation remains the most effective strategy to minimize the risks of all AI, including those posed by foundation models.

The best way policymakers can meaningfully address concerns related to foundation models is to ensure any AI policy framework is risk-based and appropriately focused on the deployers of AI systems. Read the IBM Policy Lab’s A Policymaker’s Guide to Foundation Models—a new white paper from us, IBM’s Chief Privacy & Trust Officer Christina Montgomery, AI Ethics Global Leader Francesca Rossi, and IBM Policy Lab Senior Fellow Joshua New—to understand why IBM is asking policymakers to:

  1. Promote transparency
  2. Leverage flexible approaches
  3. Differentiate between different kinds of business models
  4. Carefully study emerging risks

Given the incredible benefits of foundation models, effectively protecting the economy and society from its potential risks will help to ensure that the technology is a force for good. Policymakers should swiftly act to better understand and mitigate the risks of foundation models while still ensuring the approach to governing AI remains risk-based and technology neutral. 

Read “A Policymaker’s Guide to Foundation Models”
Was this article helpful?
YesNo

More from Artificial intelligence

Optimize your call center operations with new IBM watsonx assistants features

5 min read - Everyone has had at least one bad experience when dialing into a call center. The robotic audio recording, the limited menu options, the repetitive elevator music in the background, and the general feeling of time wasted are all too familiar. As customers try to get answers, many times they find themselves falling into the infamous spiral of misery, searching desperately to speak to a live agent. While virtual assistants, mobile applications and digital web interfaces have made self-service options in…

IBM, with flagship Granite models, named a strong performer in The Forrester Wave™: AI Foundation Models for Language, Q2 2024

6 min read - As enterprises move from generative artificial intelligence (gen AI) experimentation to production, they are looking for the right choices when it comes to foundation models with an optimal mix of attributes that yield trusted, performant and cost-effective gen AI. Businesses recognize that they cannot scale gen AI with foundation models they cannot trust. We are pleased to announce that IBM, with its flagship Granite family of models, has been named a strong performer in the Forrester Wave™: AI Foundation Models…

Scale enterprise gen AI for code generation with IBM Granite code models, available as NVIDIA NIM inference microservices

3 min read - Many enterprises today are moving from generative AI (gen AI) experimentation to production, deployment and scaling. Code generation and modernization are now among the top enterprise use cases that offer a clear path to value creation, cost reduction and return on investment (ROI). IBM® Granite™ is a family of enterprise-grade models developed by IBM Research® with rigorous data governance and regulatory compliance. Granite currently supports multilingual language and code modalities. And as of the NVIDIA AI Summit in Taiwan this…

IBM Newsletters

Get our newsletters and topic updates that deliver the latest thought leadership and insights on emerging trends.
Subscribe now More newsletters