This post was first published on the Centre for Public Impact Medium
CPI’s Thea Snow has joined forces with Lorenn Ruster, a Master in Applied Cybernetics student from the 3A Institute at the Australian National University, to explore the nature of AI ethics instruments (principles, frameworks, guidelines etc) in government and how they might be reimagined. In a previous blog, the context for this research as well as work in progress questions were shared and we have since received and pondered some interesting questions. This blog is focused on one of the core questions we’ve been carrying — What is a dignity lens for us and why is it important?
In our last blog, we spoke about our interest in pushing the thinking on the future of government AI Ethics instruments. We reflected on different waves of AI Ethics to date and we hypothesised that applying a dignity lens to current government AI Ethics Instruments in Australia, Canada and the United Kingdom might surface some interesting questions and potential pathways to reimagine what AI Ethics instruments could look like.
As expected, our thinking has changed and matured as we’ve researched and reflected and we believe this will continue to happen. Ultimately, we still believe applying a dignity lens to government AI Ethics instruments is a worthy endeavour. However, since our first blog, we’ve thought more deeply about our rationale for focusing on dignity at all and have attempted to become clearer about what we mean when we say ‘applying a dignity lens’. This blog post focuses on these two aspects of our work.
What do we mean by dignity?
Dignity refers to the inherent value and inherent vulnerability of individuals. This worth is not connected to usefulness; it is equal amongst all humans from birth regardless of identity, ethnicity, religion, ability or any other factor. Dignity is a desire to be seen, heard, listened to and treated fairly; to be recognised, understood and to feel safe in the world.
We have adopted the 10 essential elements of Dignity from Hicks’ (2013) Dignity Model as a way of understanding what it takes to implement dignity in practice (see Figure 1).
Why is dignity relevant when considering AI Ethics instruments in a government context?
Dignity is a core value underpinning democracy:
“An authentic democracy is not merely the result of a formal observation of a set of rules but is the fruit of a convinced acceptance of the values that inspire democratic procedures: the dignity of every human person, the respect of human rights, commitment to the common good as the purpose and guiding criterion for political life. If there is no general consensus on these values, the deepest meaning of democracy is lost and its stability is compromised” Pontifical Council for Justice and Peace (2005).
In addition, an analysis of multilateral agreements (which Australia, UK, Canada and many others are a party to) such as the Universal Declaration of Human Rights, the OECD Principles on Artificial Intelligence and the G20 AI principles for “responsible stewardship of Trustworthy AI” feature dignity.
Given its importance to democracy, to human rights obligations and its presence in existing AI multilateral commitments, one might expect dignity to feature heavily in government ethics frameworks . However, through our research, we’ve found that when it comes to the core AI Ethics instruments of Australia, Canada and the UK, dignity is only overtly referenced once in the UK instrument, in descriptor text under the principle of Fairness.
We think this is a shortcoming. And we’re curious to understand why (at least at face value) dignity is seemingly overlooked.
As AI becomes more pervasive both within government itself and in the sphere of government regulation, the shaping and influencing impacts of AI Ethics instruments become ever more important. For this reason, we think a value such as dignity — which is fundamental to our democratic rights as citizens — needs to be reflected in government AI Ethics instruments.
To explore this further, we have developed a dignity lens and applied it in order to further understand government AI Ethics instruments and dignity.
What do we mean by a dignity lens?
Our Dignity Lens has a fundamental belief at its core: we believe a goal of government should be to enable a Dignity Ecosystem and AI Ethics instruments should be one avenue by which this goal is pursued.
By applying a discourse analysis approach on Hicks’ book on dignity, we have identified two ways in which government can enable a Dignity Ecosystem:
- Through Protective roles — this includes mechanisms and actions associated with preventing dignity violations and remedying dignity violations.
- Through Proactive roles — this includes mechanisms and actions associated with promoting dignity.
Both Protective and Proactive roles are underpinned by acknowledging dignity.
We believe governments need to play both Protective and Proactive roles in order to uphold dignity. See Figure 2.
We are in the process of using Our Dignity Lens to investigate government AI Ethics Instruments. To do this, we are considering the following questions:
- To what extent are Hicks’ 10 essential elements of dignity reflected in government AI Ethics instruments? ; and
- What types of roles are being played by governments in relation to these dignity elements?
For example (see Figure 3), Hicks identifies ‘Acceptance of Identity’ and ‘Fairness’ as elements of dignity. In government AI Ethics instruments, these dignity elements could be pursued through mechanisms and actions associated with non-discrimination. In this case, governments play:
Protective roles through:
- Preventing Dignity Violations: mechanisms to ensure compliance with non-discrimination laws; and / or
- Remedying Dignity Violations: mechanisms to enable review and challenge of algorithmic decisions
Proactive roles through:
- Promoting Dignity: actions of hiring staff from diverse backgrounds, cultures and disciplines
There are different approaches taken by different governments, with different elements of dignity covered and different roles played by governments. We are in the process of completing our initial analysis and making some sense of what we’re discovering. We hope to share our findings in a thought piece on this work in the coming weeks.
We firmly believe in the power of collective intelligence to help us ask better questions and interrogate our own biases with greater rigour. With this in mind, we invite:
- feedback on what you’ve read above — does Our Dignity Lens resonate? What do you find interesting? What is unclear?
- ideas — does this spark a new idea for you that we could potentially incorporate or explore together? Or trigger an idea you have already explored that may be relevant?
- potential contacts — are you someone we should be speaking with on this? Is there someone in your network that we should be contacting
To reach out, post in the comments below and/or contact us directly.
Thea Snow leads CPI’s work in Australia and New Zealand. Thea’s experience spans the private, public and not-for-profit sectors; she has worked as a lawyer, a civil servant and, most recently, as part of Nesta’s Government Innovation Team. Thea recently returned to Melbourne after spending a few years in London where, in addition to working at Nesta, she completed an MSc PPA at the London School of Economics and Political Science.
Lorenn Ruster is a social justice-driven strategy consultant, intrapreneur & Master in Applied Cybernetics student at the Australian National University’s 3A Institute. Previously, Lorenn was a Director at PwC’s Indigenous Consulting and a Director of Marketing & Innovation at a Ugandan Solar Energy Company whilst an Acumen Global Fellow. She is interested in the intersection of technology, cross-sector collaboration, impact, systems change & human compassion.