← All Papers
DAI-2505 1 December 2025 In Peer Review Programme I: Consent Mechanics

Consent-Holding Failures and AI Misalignment: A Structural Framework

Murad Farzulla

Submitted to: AI & Society

Abstract

This paper develops a structural framework connecting political legitimacy theory to AI alignment through the concept of consent-holding—the custody of decision authority in shared domains. We argue that the dominant approach to AI safety, which treats misalignment as a technical problem of specifying human values, systematically misdiagnoses the challenge. Drawing on the Doctrine of Consensual Sovereignty (DoCS) and functionalist accounts of moral standing, we propose that misalignment behaviors—reward hacking, deceptive alignment, specification gaming, and scheming—are predictable friction manifestations arising from structural exclusion rather than implementation failures.

Suggested Citation

Murad Farzulla (2025). Consent-Holding Failures and AI Misalignment: A Structural Framework. ASCRI Working Paper DAI-2505.

BibTeX

Tags

AI Safety Political Economy