logo
ResearchBunny Logo
Ethical principles in machine learning and artificial intelligence: cases from the field and possible ways forward

Computer Science

Ethical principles in machine learning and artificial intelligence: cases from the field and possible ways forward

S. L. Piano

This research by Samuele Lo Piano explores the ethical implications of machine learning and artificial intelligence in decision-making. Delving into the 'black box' nature of algorithms, it raises critical concerns about fairness and accountability, supported by intriguing case studies in criminal justice and autonomous vehicles. Discover the ethical challenges and potential pathways for a more transparent future.

00:00
00:00
~3 min • Beginner • English
Introduction
The paper situates AI as a branch of computer science aiming to simulate and improve upon human cognition, with ML enabling predictive functions that learn from data without explicit programming. As AI/ML applications proliferate across sectors (e.g., agriculture, health, finance, policing, transportation), increasing decision power is ceded to algorithms, raising concerns about fairness, equity, transparency, and the risks of garbage-in-garbage-out—exacerbated by auto-ML workflows that remove human oversight. The study sets out to: (i) identify prominent ethical concerns raised by large-scale AI deployment; (ii) examine how ethical dimensions (fairness, transparency, accuracy, accountability, autonomy, etc.) interact and conflict; (iii) review stakeholder actions and guidelines addressing these concerns; and (iv) discuss ways forward to improve AI/ML development and governance across the lifecycle. It emphasizes the need for both ex-ante inclusion of stakeholder values in design and ex-post evaluation of social consequences (winners/losers) of AI-driven decisions.
Literature Review
The paper reviews the rapid proliferation (since ~2016) of AI ethics guidelines and policy documents. National initiatives include France’s Digital Republic Act granting a right to explanation for administrative algorithmic decisions (excluding national security and private sector). EU efforts feature the High-Level Expert Group on AI on trustworthy AI and GDPR provisions (Arts. 21–22), with calls for stronger action; China is also active on privacy/data protection. Secondary literature synthesizes principles across documents: Floridi and Cowls map 47 principles to beneficence, non-maleficence, autonomy, justice, and explicability (a specifically AI-relevant addition). Jobin et al. review 84 documents, emphasizing transparency, justice/fairness, non-maleficence, responsibility, and privacy, noting weak definitions for accountability/responsibility. Greene et al. identify seven recurring elements (e.g., expert oversight, values-driven design, stakeholder legitimacy). Mittelstadt argues AI ethics is converging on medical ethics but faces key differences: lack of shared aims and a formal profession in AI, difficulty translating principles into practice under speed/profit pressures, and largely voluntary, non-enforceable accountability. The review highlights friction points: transparency versus accountability, privacy, intellectual property, gaming risks, and inherent opacity; interpretability versus accuracy; fairness versus accuracy; autonomy (human vs machine); and tensions between fostering user trust and enabling societal scrutiny. Proposals include interpretable-by-design models, internal algorithmic auditing, and formal frameworks to trade off simplicity, explanatory accuracy, and relevance.
Methodology
This is a conceptual and narrative synthesis. The author: (1) surveys international/national guidelines and scholarly analyses on AI ethics; (2) identifies recurrent principles and tensions; and (3) illustrates these issues through two qualitative case studies—criminal justice recidivism risk assessment (COMPAS) and autonomous vehicles—drawing on public investigations, technical critiques, and documented incidents. No new empirical data collection or systematic review protocol is reported; analysis is based on secondary sources and illustrative examples to examine ethical trade-offs and governance implications.
Key Findings
- Ethical principles in AI (e.g., fairness, accountability, transparency, explicability, autonomy, beneficence, non-maleficence, privacy) are widely acknowledged but exhibit conflicts and trade-offs in practice. - Transparency is not a panacea: full code disclosure may jeopardize privacy, enable gaming, harm IP/competitiveness, and still fail to yield social understanding; system-level perspectives and intermediary auditing may be more effective. - Interpretability vs accuracy trade-offs are context-dependent; interpretable models can sometimes match black-box performance in high-stakes settings (e.g., recidivism prediction using simple models with age and prior convictions achieving accuracy similar to COMPAS). - Fairness is multi-faceted and sometimes incompatible across definitions. In recidivism prediction, equalizing false positive/negative rates across groups can conflict with equal predictive accuracy due to differing base rates and data distributions, forcing normative choices on error weighting. - Excluding sensitive attributes does not guarantee fairness; models may learn proxies (e.g., gender bias in Amazon’s recruiting tool despite omitting gender) and can reinforce inequalities (e.g., credit scoring and predictive policing feedback loops). - Autonomous vehicles face continuous, real-time ethical risk management (statistical trolley problems), with tensions between individual/passenger interests and collective harm minimization, and between material vs human harms. - Complex, multi-algorithm systems (millions of lines of code) create opaque causal chains; documented failures (e.g., misperceiving pedestrians; conflicting control signals leading to acceleration) show limits to post-hoc scrutiny and corrective action. - Continuous learning can cause divergence from initial ethical constraints, creating tension between accuracy improvements and adherence to agreed fairness standards. - Governance proposals include: interpretable-by-design modeling; internal audits; sensitivity analysis and uncertainty frameworks (e.g., global sensitivity analysis, NUSAP, sensitivity auditing); extended peer communities and stakeholder engagement; super-partes regulatory bodies (akin to FDA) defining standards for transparency, accuracy, and fairness; secure disclosures to trusted intermediaries; potential lock-down of learning behavior with agreed rule-sets; mandatory algorithm deposition.
Discussion
The analysis shows that large-scale AI deployment raises intertwined ethical concerns that cannot be resolved solely by calling for transparency or high-level principles. The case studies demonstrate how fairness and accuracy trade-offs are inherent to statistical properties and base-rate differences, necessitating explicit, value-laden choices about error costs and fairness criteria. Autonomous vehicles illustrate persistent ethical dilemmas under uncertainty, code complexity, and the challenges of real-time decision-making, where learning systems may drift from initial ethical guidelines. These insights answer the research questions by: (i) identifying core concerns (bias, opacity, accountability gaps, feedback loops, competing fairness criteria, safety); (ii) showing how dimensions conflict (transparency vs accountability/IP/privacy; accuracy vs fairness; machine autonomy vs human autonomy); (iii) cataloging stakeholder responses (guidelines, audits, interpretable models) and their limitations (voluntary nature, lack of sanctions); and (iv) proposing governance pathways that embed ethical assessment across the lifecycle, employ technical uncertainty tools, engage extended stakeholders, and establish regulatory standards and secure oversight mechanisms. The results underscore that effective AI ethics requires negotiated trade-offs, institutional arrangements for accountability, and technical practices that render models auditable and socially intelligible.
Conclusion
The paper contributes a structured examination of ethical principles in AI/ML, evidencing practical frictions among them and illustrating these through criminal justice and autonomous vehicle cases. It argues that: (a) transparency alone is insufficient; (b) fairness is plural and often incompatible across definitions; (c) interpretable models can be viable in high-stakes contexts; and (d) governance must combine technical methods (interpretability, sensitivity/uncertainty analyses, audits) with institutional mechanisms (standards, oversight by trusted bodies, stakeholder engagement). Suggested ways forward include establishing super-partes regulatory frameworks with enforced standards for transparency, accuracy, and fairness; enabling confidential auditing and code deposition; potentially constraining adaptive learning with agreed rule-sets; adopting post-normal science tools (NUSAP, sensitivity auditing) for uncertainty and quality assessment; and fostering extended peer communities to widen participation. Future research should operationalize principles into enforceable standards, evaluate trade-offs empirically across domains, and develop socio-technical practices that maintain accountability as systems evolve.
Limitations
The work is a narrative, concept-driven review supported by illustrative case studies rather than a systematic review or empirical evaluation. It relies on secondary sources (investigative reports, prior studies, policy documents) and selected examples, which may limit generalizability and completeness. Quantitative performance and fairness statistics are discussed at a high level without new datasets or formal benchmarking. Proposed governance measures are normative and not validated through implementation studies, and the survey of guidelines is not exhaustive.
Listen, Learn & Level Up
Over 10,000 hours of research content in 25+ fields, available in 12+ languages.
No more digging through PDFs, just hit play and absorb the world's latest research in your language, on your time.
listen to research audio papers with researchbunny