Transparent Governance in the Algorithmic Age: A Comparative Analysis of Systems and the Application of an Evaluative Framework to BOSCO

Authors

  • José Luis Bustelo Gracia Observatorio ESERP Digital Business and Law School Barcelona. España

DOI:

https://doi.org/10.51915/ret.379

Keywords:

Algorithmic transparency, public governance, artificial intelligence, accountability, ethics

Abstract

In the context of the increasing use of algorithms in public governance, this paper examines the challenges of algorithmic transparency through a dual approach. First, it presents a comparative analysis of three emblematic automated decision-making systems used in public administration—COMPAS, SyRI, and SALER—to identify common patterns of opacity, bias, and lack of oversight. Second, it designs, applies, and validates a comprehensive evaluative framework focused on the principles of explainability, accessibility, human supervision, and fairness, which is empirically tested on BOSCO, an algorithm used in Spain to allocate the social electricity subsidy. Methodologically, the study combines a systematic literature review, a comparative case study analysis, and expert validation through a Delphi panel. The findings reveal structural deficiencies in BOSCO's transparency, particularly regarding technical accessibility and accountability, highlighting the limitations of current regulatory frameworks. The article concludes by proposing a replicable evaluation model that enables public administrations to effectively audit the transparency of algorithmic systems, thereby reinforcing democratic governance in the digital era.

Downloads

Download data is not yet available.

Global Statistics ℹ️

Cumulative totals since publication
0
Views
0
Downloads
0
Total

References

AMONI REVERÓN, G. A. (2021). Derechos ante decisiones judiciales y administrativas algorítmicas reconocidos en los casos CADA, Loomis, Lazio y SyRI. Revista Administración y Ciudadanía, 125–142. https://doi.org/10.36402/ac.v16i2.4803

ANGWIN, J., LARSON, J., MATTU, S., & KIRCHNER, L. (2016). Machine Bias: There´s software used across the country to predict future criminals. And it`s biased agaisnst blacks. ProPublica. https://www.propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing

ANSGAR, K., CHRIS, C., YOHKO, H., HELENA, W., & RASHIDA, R. (2019). A governance framework for algorithmic accountability and transparency. European Parliament Think Tank. https://doi.org/10.2861/59990

ASHISH, S. (2024). AI in governance and policy making. International Journal of Science and Research (USR). https://doi.org/10.21275/sr24519015426

BEKKER, S. (2021). Fundamental rights in digital welfare states: The case of SyRI in the Netherlands. In O.spijkers, W. G. Werner, & R. A. Wessel (Eds), Netherlands Yearbook of International Law 2019 (Vol. 50, pp 289-307). T.M.C. Asser Press. https://doi.org/10.1007/978-94-6265-403-7_24

BOIX PALOP, A. (2020). Los algoritmos son reglamentos: La necesidad de extender las garantías propias de las normas reglamentarias a los programas empleados por la administración para la adopción de decisiones. Revista de Derecho Público: Teoría y Método, (1), 223–270. https://doi.org/10.37417/RPD/vol_1_2020_33

CERRILLO I MARTÍNEZ, A. (2020). ¿Son fiables las decisiones de las Administraciones públicas adoptadas por algoritmos? European Review of Digital Administration & Law, 1 (1-2), 18-36 https://www.erdalreview.eu/free-download/97888255389603.pdf

CHAUDHARY, G. (2024). Unveiling the black box: Bringing algorithmic transparency to AI. Masaryk University Journal of Law and Technology,18 (1), 93-116. https://doi.org/10.5817/mujlt2024-1-4

CHEONG, B. C. (2024). Transparency and accountability in AI systems: Safeguarding wellbeing in the age of algorithmic decision-making. Frontiers in Human Dynamics, 6, Article 1421273. https://doi.org/10.3389/fhumd.2024.1421273

COGLIANESE, C., & LEHR, D. (2019). Transparency and algorithmic governance. Administrative Law Review, Vol 71, P.1, 2019, U of Penn Law School Public Law Research Paper Nº18-38, available at 38, Available at SSRN: https://ssrn.com/abstract=3293008

COMISIÓN EUROPEA. (2021). Propuesta de Reglamento del Parlamento Europeo y del Consejo por el que se establecen normas armonizadas en materia de inteligencia artificial (Ley de Inteligencia Artificial) y se modifican determinados actos legislativos de la Unión. https://eur-lex.europa.eu/legal-content/ES/TXT/?uri=CELEX%3A52021PC0206

CRIADO, J. I., VALERO, J., & VILLODRE, J. (2020). Algorithmic transparency and bureaucratic discretion: The case of SALER early warning system. Information Polity 25, 449–470. https://doi.org/10.3233/IP-200260

DATTA, A., SEN, S., & ZICK, Y. (2017). Algorithmic transparency via quantitative input influence: Theory and experiments with learning systems. En Proceedings of the IEEE Symposium on Security and Privacy.598–617. https://doi.org/10.1109/SP.2016.42

DE MIGUEL BERIAIN, I. (2018). Does the use of risk assessments in sentences respect the right to due process? A critical analysis of the Wisconsin v. Loomis ruling. Law, Probability and Risk, 17(1), 45–53. https://doi.org/10.1093/lpr/mgy001

FREEMAN, K. (2016). Algorithmic injustice: How the Wisconsin Supreme Court failed to protect due process rights in State v. Loomis. North Carolina Journal of Law & Technology, 18(5), 75–106. https://scholarship.law.unc.edu/ncjolt/vol18/iss5/3/

GIL GONZÁLEZ, E. (2017). Aproximación al estudio de las decisiones automatizadas en el seno del Reglamento General Europeo de Protección de Datos a la luz de las tecnologías big data y aprendizaje computacional. Revista Española de la Transparencia, (5), 165–179. https://dialnet.unirioja.es/servlet/articulo?codigo=6312525

GOBIERNO DE ESPAÑA. (2017). Real Decreto 897/2017, de 6 de octubre, por el que se regula la figura del consumidor vulnerable, el bono social y otras medidas de protección para los consumidores domésticos de energía eléctrica. https://www.boe.es/buscar/act.php?id=BOE-A-2017-11505

GOVERNMENT OF CANADA. (2021). Algorithmic Impact Assessment: A tool for analyzing and mitigating the risks associated with automated decision systems. https://canada-ca.github.io/aia-eia-js/

GUTIÉRREZ DAVID, M. E. (2021). Administraciones inteligentes y acceso al código fuente y algoritmos públicos: Conjurando riesgos de cajas negras decisionales. Derecom: Revista Internacional de Derecho de la Comunicación y de las Nuevas Tecnologías, (30), 143-228. Recuperado de https://dialnet.unirioja.es/servlet/articulo?codigo=7980499

HUERGO LORA, A. (2024). ¿Por qué aciertan las sentencias sobre el algoritmo del bono social eléctrico? Almacén de Derecho. https://almacendederecho.org/por-que-aciertan-las-sentencias-sobre-el-algoritmo-del-bono-social-electrico

JACKSON, E., & MENDOZA, C. (2020). Setting the record straight: What the COMPAS core risk and need assessment is and is not. Harvard Data Science Review, 2(1),3 https://doi.org/10.1162/99608f92.1b3dadaa

JIMÉNEZ SERRANÍA, V., CASTILLA BAREA, M., & DELGADO MARTÍN, J. (2024). El Reglamento Europeo de Inteligencia Artificial. (M. BARRIO ANDRÉS, Dir). Valencia, España: Editorial: Tirant lo Blanch.

KANELLOPOULOS, Y. (2018). A model for evaluating algorithmic systems accountability. arXiv.org. https://doi.org/10.48550/arXiv.1807.06083

LAZCOZ MORATINOS, G., & CASTILLO PARRILLA, J. A. (2020). Valoración algorítmica ante los derechos humanos y el Reglamento General de Protección de Datos. El caso SyRI. Revista Chilena de Derecho y Tecnología, 9(1), 207–225. https://doi.org/10.5354/0719-2584.2020.56843

MIRI, Z., SARGEANT, H., & WELLER, A. (2022). Transparency, governance and regulation of algorithmic tools deployed in the criminal justice system. A UK case study. In Proccedings of the 2022 AAAI/ACM Conference on AI, Ethnics, and Society (AIES`22) Association for Computing Machinery. https://doi.org/10.1145/3514094.3534200

PAGE, M. J., MCKENZIE, J. E., BOSSUYT, P. M., ... & MOHER, D. (2021). The PRISMA 2020 statement: An updated guideline for reporting systematic reviews. BMJ, 372, n71. https://doi.org/10.1136/bmj.n71

PALMA ORTIGOSA, A. (2019). Decisiones automatizadas en el RGPD. El uso de algoritmos en el contexto de la protección de datos. Revista General de Derecho Administrativo, (50). https://dialnet.unirioja.es/servlet/articulo?codigo=6823524

PASQUALE, F. (2015). The Black Box Society: The secret algorithms that control money and information. Harvard University Press. https://doi.org/10.4159/harvard.9780674736061

PONCE SOLÉ, J. (2024). ¿Por qué se equivocan las sentencias sobre el algoritmo del bono social eléctrico? Almacén de Derecho. https://almacendederecho.org/por-que-se-equivocan-las-sentencias-sobre-el-algoritmo-del-bono-social-electrico

PUNCEL CHORNET, A. (2019). Inteligencia artificial para la transparencia pública: el sistema de alertas tempranas (SALER) de la Generalitat Valenciana. Boletín Económico de ICE, (3116), 41–61. https://doi.org/10.32796/bice.2019.3116.6914

RACHOVITSA, A., & JOHANN, N. (2022). The human rights implications of the Use of AI in the digital welfare state: Lessons learned from the Dutch SyRI Case. Human Rights Law Review. https://doi.org/10.1093/hrlr/ngac010

SHEA, B. J., GRIMSHAW, J. M., WELLS, G. A., ... & BOUTER, L. M. (2007). Development of AMSTAR: A measurement tool to assess the methodological quality of systematic reviews. BMC Medical Research Methodology, 7(1), 10. https://doi.org/10.1186/1471-2288-7-10

SINGH, D., KUMAR, I., NIJHAWAN, G., & GEETHA, B. T. (2024). Ensuring transparency and fairness in AI decision-making processes influenced by large language models. En Proceedings of the 2024 IEEE International Conference on Communication Systems and Network Technologies (CSNT) (pp. 987–996). IEEE. https://doi.org/10.1109/CSNT60213.2024.10545998

SORIANO ARNANZ, A. (2021). Decisiones automatizadas: problemas y soluciones jurídicas. Más allá de la protección de datos. Revista de Derecho Público: Teoría y Método, (3), 85–127. https://doi.org/10.37417/RPD/vol_3_2021_535

TOBIAS, P., BITZER, M., WIENER, M., & CRAM, A. (2023). Algorithmic transparency: Concepts, antecedents and consequences a review and research framework. Communications of the Association for Information Systems, 52 293-331 331. https://doi.org/10.17705/1CAIS.05214

VALERO TORRIJOS, J. (2019). Las garantías jurídicas de la inteligencia artificial en la actividad administrativa desde la perspectiva de la buena administración. Revista Catalana de Dret Públic, (58), 82–96. https://doi.org/10.2436/rcdp.i58.2019.3307

VELASCO RICO, C. I. (2024). Implementación de sistemas algorítmicos en el sector público, preferencias ciudadanas y obligaciones de transparencia a la luz del Reglamento de IA de la Unión Europea. En A. Cerrillo i Martínez, F. Di Lascio, I. Martín Delgado, & C.I Velasco Rico (Dirs,), Inteligencia artificial y administraciones públicas: una tiple visión en clave comparada (pp.47-70). Iustel. https://dialnet.unirioja.es/servlet/articulo?codigo=9830938

Published

2025-11-27

How to Cite

Bustelo Gracia, J. L. (2025). Transparent Governance in the Algorithmic Age: A Comparative Analysis of Systems and the Application of an Evaluative Framework to BOSCO. Spanish Journal of Transparency, (22), 145–179. https://doi.org/10.51915/ret.379

Similar Articles

1 2 3 4 5 6 7 8 9 10 > >> 

You may also start an advanced similarity search for this article.