Human oversight is the process of monitoring, evaluating, and adjusting AI systems throughout their development and deployment to ensure alignment with human values, laws, and ethical standards. By actively guiding AI during training—such as providing feedback in supervised or reinforcement learning—and intervening in real time once deployed, humans can mitigate unintended outcomes and reinforce acceptable behavior. Techniques like reinforcement learning from human feedback — an approach that uses human ratings as a proxy for preference — is used by ChatGPT as a way to keep Humans in the Loop. Human oversight, however, is ineffective if preference data is biased. If AI systems surpass human intelligence, aligning them with human values will require new research breakthroughs. In any case, individuals and organizations must make good-faith efforts to responsibly oversee AI systems to ensure they operate ethically and in alignment with societal standards.
Disclaimer: Our global network of contributors to the AI & Human Rights Index is currently writing these articles and glossary entries. This particular page is currently in the recruitment and research stage. Please return later to see where this page is in the editorial workflow. Thank you! We look forward to learning with and from you.
Last Updated: March 6, 2025
Research Assistant: Aarianna Aughtry
Contributor: Tej Shah
Reviewer: To Be Determined
Editor: Alexander Kriebitz
Subjects: Ethics, Technology
Recommended Citation: "Human Oversight, Edition 1.0 Research." In AI & Human Rights Index, edited by Nathan C. Walker, Dirk Brand, Caitlin Corrigan, Georgina Curto Rex, Alexander Kriebitz, John Maldonado, Kanshukan Rajaratnam, and Tanya de Villiers-Botha. New York: All Tech is Human; Camden, NJ: AI Ethics Lab at Rutgers University, 2025. Accessed April 17, 2025. https://aiethicslab.rutgers.edu/glossary/human-oversight/.