Industry Standards as vehicle to address socio-technical AI challenges
1. Industry Standards as vehicle to
address socio-technical challenges
from AI – the case of
Algorithmic Bias Considerations
ANSGAR KOENE,
HORIZON DIGITAL ECONOMY RESEARCH INSTITUTE, UNIVERSITY OF NOTTINGHAM
ANSGAR.KOENE@NOTTINGHAM.AC.UK
4TH FEBRUARY 2019
1
6. Case study: Recidivism risk prediction
6
Is the algorithm fair
to all groups?
When base rates differ, no non-trivial solution can achieve similar FPR,
FNR, FDR, FOR!
Machine Bias: There’s
software used across the
country to predict future
criminals. Propublica
Angwin, J., Larson, J., Mattu, S., & Kirchner, L. (2016, May 23). Machine bias. ProPublica.
https://www.propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing.
7. Arkansas algorithmic Medicaid
assessment instrument
When introduced in 2016, many people with cerebral palsy had their care dramatically
reduced – they sued the state resulting in a court case.
Investigation in court revealed:
The algorithm relies on 60 answer scores to questions about descriptions, symptoms and
ailments. A small number of variables could matter enormously: a difference between a
three instead of a four on a handful of items meant a cut of dozens of care hours a month.
One variable was foot problems. When an assessor visited a certain person, they wrote that
the person didn’t have any foot problems — because they were an amputee and didn’t
have feet.
Third-party software vendor implementing the system, mistakenly used a version of the
algorithm that didn’t account for diabetes issues.
Cerebral palsy, wasn’t properly coded in the algorithm, causing incorrect calculations for
hundreds of people, mostly lowering their hours.
https://www.theverge.com/2018/3/21/17144260/healthcare-medicaid-algorithm-arkansas-cerebral-palsy
8. Fairness is fundamentally a societally defined construct (e.g. equality of
outcomes vs equality of treatment)
◦ Cultural differences between nations/jurisdictions
◦ Cultural changes in time
“Code is Law”: Algorithms, like laws, both operationalize and entrench spatio-
temporal values
Algorithms, like the law, must be:
◦ transparent
◦ adaptable to change (by a balanced process)
The ‘messy’ problem of fair algorithms
8
9. Algorithmic systems are socio-technical
Algorithmic systems do not exist in a vacuum
They are built, deployed and used:
◦ by people,
◦ within organizations,
◦ within a social, political, legal and cultural context.
The outcomes of algorithmic decisions can have significant impacts
on real, and possibly vulnerable, people.
11. 11
EU response (in addition to GDPR)
EU Parliament Science and Technology Options Assessment
(STOA) panel request for study on “Algorithmic Opportunities
and Accountability”
12. Governance options
12
Florian Saurwein, Natascha Just, Michael Latzer, (2015) "Governance of algorithms: options and limitations",
info, Vol. 17 Issue: 6, pp.35-49, doi: 10.1108/info-05-2015-0025
22. ACM Principles on Algorithmic
Transparency and Accountability
Awareness
Access and Redress
Accountability
Explanation
Data Provenance
Auditability
Validation and Testing
22
23. FAT/ML: Principles for Accountable Algorithms
and a Social Impact Statement for Algorithms
Responsibility: Externally visible avenues of redress for adverse effects, and
designate an internal role responsible for timely remedy of such issues.
Explainability: Ensure algorithmic decisions and data driving those decisions can
be explained to end-users/stakeholders in non-technical terms.
Accuracy: Identify, log, and articulate sources of error and uncertainty so that
expected/worst case implications can inform mitigation procedures.
Auditability: Enable third parties to probe, understand, and review algorithm
behavior through disclosure of information that enables monitoring, checking,
or criticism, including detailed documentation, technically suitable APIs, and
permissive terms of use.
Fairness: Ensure algorithmic decisions do not create discriminatory or unjust
impacts when comparing across different demographics.
https://www.fatml.org/resources/principles-for-accountable-algorithms 23
24. Artificial Intelligence at Google
Our Principles
24https://dzone.com/articles/ethical-ai-lessons-from-google-ai-principles
25. Are ethics principles the unenforcable
‘self-regulation’?
Wagner, B. (2018). Ethics as an Escape from Regulation: From
ethics-washing to ethics-shopping? In M. Hildebrandt (Ed.), Being
Profiling. Cogitas ergo sum. Amsterdam University Press.
25
28. 28
IEEE P70xx Standards Projects
IEEE P7000: Model Process for Addressing Ethical Concerns During System Design
IEEE P7001: Transparency of Autonomous Systems
IEEE P7002: Data Privacy Process
IEEE P7003: Algorithmic Bias Considerations
IEEE P7004: Child and Student Data Governance
IEEE P7005: Employer Data Governance
IEEE P7006: Personal Data AI Agent Working Group
IEEE P7007: Ontological Standard for Ethically Driven Robotics and Automation Systems
IEEE P7008: Ethically Driven Nudging for Robotic, Intelligent and Autonomous Systems
IEEE P7009: Fail-Safe Design of Autonomous and Semi-Autonomous Systems
IEEE P7010: Wellbeing Metrics Standard for Ethical AI and Autonomous Systems
IEEE P7011: Process of Identifying and Rating the Trustworthiness of News Sources
IEEE P7012: Standard for Machines Readable Personal Privacy Terms
29. Related standards activities
British Standards Institute (BSI) – BS 8611 Ethics design and application of robots
ISO/IEC JTC1 SC42
◦ Artificial Intelligence Concepts and Terminology
◦ Framework for Artificial Intelligence Systems Using Machine Learning
◦ SG 2 on Trustworthiness
◦ transparency, verifiability, explainability, controllability, etc.
◦ robustness, resiliency, reliability, accuracy, safety, security, privacy, etc.
Jan 2018 China published “Artificial Intelligence Standardization White Paper.”
29
31. P7003 - Algorithmic Bias Considerations
All non-trivial* decisions are biased
We seek to minimize bias that is:
◦ Unintended
◦ Unjustified
◦ Unacceptable
as defined by the context where the system is used.
*Non-trivial means the decision space has more than one possible outcome and the choice is not
uniformly random.
32. Causes of algorithmic bias
Insufficient understanding of the context of use.
Failure to rigorously map decision criteria.
Failure to have explicit justifications for the chosen criteria.
33. Key question when developing or
deploying an algorithmic system
33
Who will be affected?
What are the decision/optimization criteria?
How are these criteria justified?
Are these justifications acceptable in the context where the
system is used?
34. 34
P7003 foundational sections
Taxonomy of Algorithmic Bias
Legal frameworks related to Bias
Psychology of Bias
Cultural aspects
P7003 algorithm development sections
Algorithmic system design stages
Person categorization and identifying affected population groups
Assurance of representativeness of testing/training/validation data
Evaluation of system outcomes
Evaluation of algorithmic processing
Assessment of resilience against external manipulation to Bias
Documentation of criteria, scope and justifications of choices
37. Thank you – Questions?
37
https://reentrust.org/
ansgar.koene@nottingham.ac.uk
Notas del editor
Industry Standards as vehicle to address socio-technical challenges from AI – the case of
Algorithmic Bias Considerations. Ansgar Koene
Abstract: Algorithmic decision-making technologies (colloquially referred to as “AI”) in industry,
commerce and public service provision are giving rise to concerns about potential negative impacts on individuals
(e.g. algorithmic discrimination bias) and the wider socio-economic fabric of society (e.g. displacement of jobs).
As a response to public concerns and government inquiries a number of industry initiatives have been launched
in an effort to stave off government intervention. Many of these initiatives are focusing on establishing “ethical
principles” or formulating “best practices” that lack clear compliance specifications. Industry standards by
contrast are well established self-regulation tools which do include compliance metrics and can be directly linked
to compliance certification. This talk will discuss issues of algorithmic bias and outline ways in which a standard
for Algorithmic Bias Considerations can help to minimizing unjustified, unintended and inappropriate bias in
algorithmic decision making.
The scholar Danielle Keats Citron cites the example of Colorado, where coders placed more than 900 incorrect rules into its public benefits system in the mid-2000s, resulting in problems like pregnant women being denied Medicaid. Similar issues in California, Citron writes in a paper, led to “overpayments, underpayments, and improper terminations of public benefits,” as foster children were incorrectly denied Medicaid. Citron writes about the need for “technological due process” — the importance of both understanding what’s happening in automated systems and being given meaningful ways to challenge them.
Automated decisions are not defined by algorithms alone. Rather, they emerge from automated systems that mix human judgment, conventional software, and statistical models, all designed to serve human goals and purposes. Discerning and debating the
social impact of these systems requires a holistic approach that considers:
Computational and statistical aspects of the algorithmic processing;
Power dynamics between the service provider and the customer;
The social-political-legal-cultural context within which the system is used;
All non-trivial decisions are biased. For example, a good results from a search engine should be biased to match the interests of the user as expressed by the search-term, and possibly refined based on personalization data.
When we say we want ‘no Bias’ we mean we want to minimize unintended, unjustified and unacceptable bias, as defined by the context within which the algorithmic system is being used.
In the absence of malicious intent, bias in algorithmic system is generally caused by:
Insufficient understanding of the context that the system is part of. This includes lack of understanding who will be affected by the algorithmic decision outcomes, resulting in a failure to test how the system performs for specific groups, who are often minorities. Diversity in the development team can partially help to address this.
Failure to rigorously map decision criteria. When people think of algorithmic decisions as being more ‘objectively trustworthy’ than human decisions, more often than not they are referring to the idea that algorithmic systems follow a clearly defined set of criteria with no ‘hidden agenda’. The complexity of system development challenges, however, can easily introduce ‘hidden decision criteria’ introduced as a quick fix during debugging or embedded within Machine Learning training data.
Failure to explicitly define and examine the justifications for the decision criteria. Given the context within which the system is used, are these justifications acceptable? For example, in a given context is it OK to treat high correlation as evidence of causation?