Author
Abstract
Existing AI governance instruments provide essential foundations for oversight, with strong capabilities for harms that can be detected, bounded, attributed, and substantially remediated after the fact. This paper examines how that architecture can be further specified for three categories of AI-mediated outcome where restorability can fail under identifiable conditions: epistemic contagion through synthetic misinformation, reputational destruction via AI-generated defamation, and structural labor displacement driven by automation at scale. Through comparative regulatory analysis of four core Euro-Atlantic frameworks (the EU AI Act, the NIST AI RMF, the UK White Paper, and the OECD incident-reporting framework), with the G7 Hiroshima Code used as an interoperability reference, the paper identifies an opportunity for additional operational specification: an explicit threshold for the point at which remediation should be complemented by additional ex ante safeguards. It proposes the Systemic Irreversibility Baseline as that complementary layer. The Baseline routes responses along distinct tracks: standard remediation for restorable product-side harms, strict liability for localized irreversible product-side harms, precautionary technical architecture for systemically irreversible productside harms, and macroeconomic transition policy for intended-utility displacement. Four design mandates operationalize the precautionary track: traceable human oversight, auditable uncertainty thresholds for bounded decision tasks, epistemic user-interface friction, and contentprovenance infrastructure. Throughout, "effectively irreversible harm" denotes harm for which restoration to functional equivalence is not feasible within governance-relevant time horizons. The contribution is a governance design framework grounded in existing empirical literatures and comparative legal analysis, intended to complement current regulatory systems rather than replace them.
Suggested Citation
Duseja, Sushil, 2026.
"Beyond Remediation: An Irreversibility Threshold for Preventive AI Governance,"
SocArXiv
g3rnj_v1, Center for Open Science.
Handle:
RePEc:osf:socarx:g3rnj_v1
DOI: 10.31219/osf.io/g3rnj_v1
Download full text from publisher
Corrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:osf:socarx:g3rnj_v1. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
We have no bibliographic references for this item. You can help adding them by using this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: OSF (email available below). General contact details of provider: https://arabixiv.org .
Please note that corrections may take a couple of weeks to filter through
the various RePEc services.