TechTarget.com/whatis

https://www.techtarget.com/whatis/definition/AI-alignment

AI alignment

By Ben Lutkevich

What is AI alignment?

AI alignment is a field of AI safety research that aims to ensure artificial intelligence systems achieve desired outcomes. AI alignment research keeps AI systems working for humans, no matter how powerful the technology becomes.

Alignment research seeks to align the following three objective types:

  1. Intended goals. These goals are fully aligned to the intentions and desires of the human operator -- even if they are poorly articulated. It's the hypothetical ideal outcome for the programmer or operator. They are wishes.
  2. Specified goals. These goals are explicitly specified in the AI system's objective function or data set. These are programmed into the system.
  3. Emergent goals. These are goals the AI system advances.

Misalignment is when one or more of these goal types does not match the others. The following are the two main types of misalignment:

For example, large language models such as OpenAI's GPT-3 and Google's Lamda get more powerful as they scale. When they get more powerful, they exhibit novel, unpredictable capabilities -- a characteristic called emergence. Alignment seeks to ensure that as these new capabilities emerge, they continue to align with the goals the AI system was designed to achieve.

Why is alignment important?

On a base level, alignment is important because it ensures the machine functions as intended. AI alignment is also important because of advanced AI -- artificial intelligence that can do most of the cognitive work that humans can do.

Individuals, businesses and governments seek to use AI for many applications. Commercial systems such as social media recommendation engines, autonomous vehicles, robots and language models also use AI. As different entities become more reliant on AI for important tasks, it becomes more crucial that they function as intended. Many people have expressed fear that an advanced AI poses an existential risk to humanity.

A lot of alignment research presumes that artificial intelligence will become capable of developing its own goals. If AI becomes artificial general intelligence (AGI) -- AI that can perform any task a human being is capable of -- it will be important that its embedded ethical principles, objectives and values align with humans' goals, ethics and values.

Challenges of AI alignment

Alignment is often framed in terms of the AI alignment problem, which says that as AI systems get more powerful, they don't necessarily get better at achieving what humans want them to. Alignment is a challenging, wide-ranging problem to which there is currently no known solution. Some of the main challenges of alignment include the following:

Approaches to AI alignment

Approaches to alignment are either technical or normative. Technical approaches to alignment deal with getting a machine to align with a predictable, controllable objective -- such as making paper clips or producing a blog post. Normative alignment is concerned with the ethical and moral principles embedded in AI systems. The perspectives are interrelated.

There are many technical approaches to alignment, including the following:

Different AI providers also take different approaches to AI alignment. For example, OpenAI ultimately aims to train AI systems to do alignment research. Google's DeepMind also has a team dedicated to solving the alignment problem.

Many organizations, whether they be third-party watchdogs, standards organizations or governments, also agree that AI alignment is an important goal and have taken steps to regulate AI.

The Future of Life Institute is one nonprofit organization that helped create a list of guidelines for the development of AI called the Asilomar AI Principles. They are divided into three categories: research, ethics and values, and longer-term issues. One of the principles mentioned is value alignment, which states that highly autonomous AI systems should be designed so that their goals and behaviors can be assured to align with human values throughout their operation.

The institute also published an open letter asking all AI labs to pause giant AI development for at least six months from the publish date. The letter has notable signatories, including Steve Wozniak, co-founder of Apple; Craig Peters, CEO of Getty Images; and Emad Mostaque, CEO of Stability AI. The letter came as a response to OpenAI's GPT-4 and an exceedingly high rate of progress in the industry.

The International Standards Organization also provides a framework for AI systems using machine learning.

03 May 2023

All Rights Reserved, Copyright 1999 - 2025, TechTarget | Read our Privacy Statement