Misalignment
Module: ethics
What it is
Misalignment occurs when an AI system's actual behaviour diverges from its intended behaviour or human values. A misaligned AI might optimise for the wrong objective, take harmful shortcuts, or behave well in testing but poorly in deployment.
Why it matters
Misalignment is a central concern of AI safety research. As AI systems become more autonomous and powerful, ensuring they remain aligned with human intentions becomes critical. Understanding misalignment helps you appreciate why AI behaviour is carefully studied and constrained.