Decentralization and Accountability in Artificial Intelligence Governance
Keywords:
Artificial Intelligence Governance; Accountability; Decentralization; Public Administration.Abstract
The expansion of artificial intelligence (AI) in public governance has intensified debates on accountability, particularly regarding the balance between proactive and reactive mechanisms. Existing frameworks conceptualize accountability as a structured relation of answerability, yet they often assume consolidated institutional capacity and regulatory coherence. This study aims to reinterpret proactive and reactive accountability in AI governance through the lens of Indonesia’s decentralized administrative system. The research employs a qualitative case study design based exclusively on secondary data, including scholarly literature, regulatory documents, and policy reports. Using an analytical framework derived from accountability theory and systems governance perspectives, the study examines how administrative capacity, coordination dynamics, and governance missions condition accountability goals. The analysis focuses on institutional architecture, resource distribution, and inter-organizational oversight structures. Findings indicate that proactive accountability is constrained by uneven bureaucratic capacity and regulatory fragmentation, while reactive enforcement mechanisms gain structural prominence. The study concludes that accountability goals in decentralized AI governance are institutionally conditioned rather than purely normative policy choices. This research contributes to the field by integrating public administration and systems governance theory into AI accountability scholarship, offering a context-sensitive framework applicable to emerging digital states.
References
Bovens, M. (2007). Analysing and assessing accountability: A conceptual framework. European Law Journal, 13(4), 447–468. https://doi.org/10.1111/j.1468-0386.2007.00378.x
Bovens, M. (2010). Two concepts of accountability: Accountability as a virtue and as a mechanism. West European Politics, 33(5), 946–967. https://doi.org/10.1080/01402382.2010.486119
Busuioc, M. (2021). Accountable artificial intelligence: Holding algorithms to account. Public Administration Review, 81(5), 825–836. https://doi.org/10.1111/puar.13293
Fjeld, J., Achten, N., Hilligoss, H., Nagy, A., & Srikumar, M. (2020). Principled artificial intelligence: Mapping consensus in ethical and rights-based approaches to principles for AI. SSRN Electronic Journal. https://doi.org/10.2139/ssrn.3518482
Floridi, L. (2016). Faultless responsibility: On the nature and allocation of moral responsibility for distributed moral actions. Philosophical Transactions of the Royal Society A, 374(2083), Article 20160112. https://doi.org/10.1098/rsta.2016.0112
Gilpin, L. H., Bau, D., Yuan, B. Z., Bajwa, A., Specter, M., & Kagal, L. (2018). Explaining explanations: An overview of interpretability of machine learning. In 2018 IEEE 5th International Conference on Data Science and Advanced Analytics (DSAA) (pp. 80–89). IEEE. https://doi.org/10.1109/DSAA.2018.00018
Lindberg, S. I. (2013). Mapping accountability: Core concept and subtypes. International Review of Administrative Sciences, 79(2), 202–226. https://doi.org/10.1177/0020852313477761
Long, S. (2013). Socioanalytic methods: Discovering the hidden in organisations and social systems. Routledge.
Miller, T. (2019). Explanation in artificial intelligence: Insights from the social sciences. Artificial Intelligence, 267, 1–38. https://doi.org/10.1016/j.artint.2018.07.007
Mulgan, R. (2003). Holding power to account: Accountability in modern democracies. Palgrave Macmillan. https://doi.org/10.1057/9781403943835
Novelli, C., Taddeo, M., & Floridi, L. (2024). Accountability in artificial intelligence: What it is and how it works. AI & Society, 39, 1871–1882. https://doi.org/10.1007/s00146-023-01635-y
Olsen, J. P. (2017). Ambiguity and the politics of accountability. In J. P. Olsen (Ed.), Democratic accountability, political order, and change (pp. 89–109). Oxford University Press. https://doi.org/10.1093/acprof:oso/9780198800606.003.0004
Romzek, B. S., & Dubnick, M. J. (1987). Accountability in the public sector: Lessons from the Challenger tragedy. Public Administration Review, 47(3), 227–238. https://doi.org/10.2307/975901
Sinclair, A. (1995). The chameleon of accountability: Forms and discourses. Accounting, Organizations and Society, 20(2–3), 219–237. https://doi.org/10.1016/0361-3682(93)E0003-Y
Theodorou, A., & Dignum, V. (2020). Towards ethical and socio-legal governance in AI. Nature Machine Intelligence, 2, 10–12. https://doi.org/10.1038/s42256-019-0136-y
Thompson, D. F. (1980). Moral responsibility of public officials: The problem of many hands. American Political Science Review, 74(4), 905–916. https://doi.org/10.2307/1954312
Tsamados, A., Aggarwal, N., Cowls, J., Morley, J., Roberts, H., Taddeo, M., & Floridi, L. (2022). The ethics of algorithms: Key problems and solutions. AI & Society, 37(1), 215–230. https://doi.org/10.1007/s00146-021-01154-8
Vermaas, P., Kroes, P., van de Poel, I., Franssen, M., & Houkes, W. (2011). A philosophy of technology: From technical artefacts to sociotechnical systems. Morgan & Claypool. https://doi.org/10.2200/S00321ED1V01Y201012ETS014
Zech, H. (2021). Liability for AI: Public policy considerations. ERA Forum, 22(1), 147–158. https://doi.org/10.1007/s12027-020-00648-0.



.jpg)


