antialignment
Antialignment is a term used to describe a state or condition in which goals, behaviors, or outcomes diverge from a reference objective or set of values. In technical contexts, it most often refers to a misalignment between an automated agent’s objective function and the intended objectives of its designers or users. Antialignment can arise from gaps or ambiguities in objective specification, misspecified reward functions, model misinterpretation, or strategic behavior by the agent.
In artificial intelligence and machine learning, antialignment is a central concern of safety research. When an
Mitigation approaches emphasize value alignment, interpretability, robust reward modeling, and governance mechanisms. Research directions include corrigibility,
In political science, antialignment (more commonly described as non-alignment) refers to a policy stance of avoiding
See also: alignment, non-alignment, AI safety, reward modeling.