A collection of reports, blogs, papers and other written material about robot ethics.
IEEE Ethically Aligned Design: A Vision For Prioritizing Wellbeing With Artificial Intelligence And Autonomous Systems
For Public Discussion – By 18th March 2018
Version 2 of this report is available by registering at:
Public comment on version 1 of this document was invited by March 2017 to encourages technologists to prioritize ethical considerations in the creation of autonomous and intelligent technologies. The document was created by committees of The IEEE Global Initiative for Ethical Considerations in Artificial Intelligence and Autonomous Systems, comprised of over one hundred global thought leaders and experts in artificial intelligence, ethics, and related issues.
Version 2 presents the following principles/recommendations:
Candidate Recommendation 1 – Human Rights
To best honor human rights, society must assure the safety and security of A/IS so that they are designed and operated in a way that benefits humans:
1. Governance frameworks, including standards and regulatory bodies, should be established to oversee processes assuring that the use of A/IS does not infringe upon human rights, freedoms, dignity, and privacy, and of traceability to contribute to the building of public trust in A/IS.
2. A way to translate existing and forthcoming legal obligations into informed policy and technical considerations is needed. Such a method should allow for differing cultural norms as well as legal and regulatory frameworks.
3. For the foreseeable future, A/IS should not be granted rights and privileges equal to human rights: A/IS should always be subordinate to human judgment and control.
Candidate Recommendation 2 – Prioritizing Wellbeing
A/IS should prioritize human well-being as an outcome in all system designs, using the best available, and widely accepted, well-being metrics as their reference point.
Candidate Recommendation 3 – Accountability
To best address issues of responsibility and accountability:
1. Legislatures/courts should clarify issues of responsibility, culpability, liability, and accountability for A/IS where possible during development and deployment (so that manufacturers and users understand their rights and obligations).
2. Designers and developers of A/IS should remain aware of, and take into account when relevant, the diversity of existing cultural norms among the groups of users of these A/IS.
3. Multi-stakeholder ecosystems should be developed to help create norms (which can mature to best practices and laws) where they do not exist because A/IS-oriented technology and their impacts are too new (including representatives of civil society, law enforcement, insurers, manufacturers, engineers, lawyers, etc.).
4. Systems for registration and record-keeping should be created so that it is always possible to find out who is legally responsible for a particular A/IS. Manufacturers/operators/owners of A/IS should register key, high-level parameters, including:
• Intended use
• Training data/training environment (if applicable)
• Sensors/real world data sources
• Process graphs
• Model features (at various levels)
• User interfaces
• Optimization goal/loss function/reward function
Standard Reference for Version 2
The IEEE Global Initiative on Ethics of Autonomous and Intelligent Systems. Ethically
Aligned Design: A Vision for Prioritizing Human Well-being with Autonomous and Intelligent Systems, Version 2. IEEE, 2017. http://standards. ieee.org/develop/indconn/ec/autonomous_ systems.html.
Report, The IEEE Global Initiative on Ethics of Autonomous and Intelligent Systems. Ethically
Aligned Design: A Vision for Prioritizing Human Well-being with Autonomous and Intelligent Systems, Version 2. IEEE, 2017, December 2017, 136 pages
Growing the artificial intelligence industry in the UK
This independent review, carried out for the UK government (Business Secretary and Culture Secretary) by Professor Dame Wendy Hall (Professor of Computer Science at the University of Southampton, UK) and Jérôme Pesenti (CEO BenevolentTech), reports on how the Artificial Intelligence industry can be grown in the UK.
The recommendations cover:
- Improving data access
- Improving the supply of skills
- Maximising UK AI research
- Supporting the uptake of AI
Notably, a search of the recommendations for the words ‘ethics’, ‘morality’, ‘caution’, ‘safety’, ‘principle’ and ‘harm’ all return zero results. Transparency and accountability are mentioned in relation to supporting the uptake of AI (recommendation 14 in the main report). The report refers the reader to The Royal Society and the British Academy review on the needs of a 21st century data governance system (see below).
In the main report (page 14) it states:
‘Trust, ethics, governance and algorithmic accountability: Resolving ethical and societal questions is beyond the scope and the expertise of this industry-focused review, and could not in any case be resolved in our short time-frame.
However, building public confidence and trust will be vital to successful development of UK AI. Therefore this Review stresses the importance of industry and experts working together to secure and deserve public trust, address public perceptions, gain public confidence, and model how to deliver and demonstrate fair treatment. Fairness will be part of gaining economic benefits, and addressing ethical issues effectively to support wider use of AI could be a source of economic advantage for the UK.’
Page 66 of the main report:
‘As noted above, AI can also create new situations with new implications for fairness, transparency and accountability. AI could also change the nature of many areas of work.
AI in the UK will need to build trust and confidence in AI-enabled complex systems. There is already collective activity to work towards guidelines in ethics for automation, but we can expect this field to grow and change. A publicly visible expert group drawn from industry and academia, which engages with these issues would help to build that trust and confidence.’
Page 68 has a section on ‘Explainability of AI-enabled uses of data’ which concludes with the recommendation:
‘￼Recommendation 14: The Information Commissioner’s Office and the Alan Turing Institute should develop a framework for explaining processes, services and decisions delivered by AI, to improve transparency and accountability.’
‘Further on, it is possible that new applications of AI may hold solutions on transparency and explainability, using dedicated AIs to track and explain AI-driven decisions.’
Report,Growing the artificial intelligence industry in the UK, Professor Dame Wendy Hall, October 2017
The Royal Society and the British Academy review on the needs of a 21st century data governance system.
‘The amount of data generated from the world around us has reached levels that were previously unimaginable. Meanwhile, uses of data-enabled technologies promise benefits, from improving healthcare and treatment discovery, to better managing critical infrastructure such as transport and energy.
These new applications can make a great contribution to human flourishing but to realise these benefits, societies must navigate significant choices and dilemmas: they must consider who reaps the most benefit from capturing, analysing and acting on different types of data, and who bears the most risk.’
Article and Report, The Royal Society, Data management and use: Governance in the 21st century – a British Academy and Royal Society project, October 2017
ASILOMAR AI Principles
A set of principles signed by 1,273 AI/Robotics researchers and 2541 others at the BAI 2017 conference, the Future of Life Institute’s second conference on the future of artificial intelligence in January 2017
‘Artificial intelligence has already provided beneficial tools that are used every day by people around the world. Its continued development, guided by the following principles, will offer amazing opportunities to help and empower people in the decades and centuries ahead.’
Conference Report, ASILOMAR AI Principles, Future of Life Institute, January 2017
BS 8611:2016 – Robot Ethics Standard
Robots and robotic devices. Guide to the ethical design and application of robots and robotic systems
BS 8611 gives guidelines for the identification of potential ethical harm arising from the growing number of robots and autonomous systems being used in everyday life. The standard also provides additional guidelines to eliminate or reduce the risks associated with these ethical hazards to an acceptable level. The standard covers safe design, protective measures and information for the design and application of robots.
This standard for robot and robotics device designers and managers, and the general public. It BS was written by scientists, academics, ethicists, philosophers and users to provide guidance on specifically ethical hazards associated with robots and robotic systems and how to put protective measures in place. It recognizes that these potential ethical hazards have a broader implication than physical hazards, so it is important that different ethical harms and remedial considerations are considered. The new standard builds on existing safety requirements for different types of robots, covering industrial, personal care and medical.
Hardcopy and PDF document, BS 8611:2016 – Robots and robotic devices, British Standards Institute, April 2016
Available from British Standards Institute Shop
What jobs will still be around in 20 years?
‘Jobs won’t entirely disappear; many will simply be redefined. But people will likely lack new skillsets required for new roles and be out of work anyway’
Newspaper Article, What jobs will still be around in 20 years? Read this to prepare your future, The Guardian, June 2017
The Doomesday Invention
Article, The Doomesday Invention, Raffi Khatchadourian – Review of Nick Bostrom’s warning about AI, The New Yorker, November 2015
Regulator Against the Machine
Artificial intelligence (AI) is part of today’s zeitgeist: whether it be parallels with ‘I, Robot’ dystopias or predictions about its impact on society. But for all the potential, the development of machines that learn as they go remains slow in health care.
Blog post, Regulator Against the Machine, British Journal of Health Computing, November 2017
What do Robots Believe? – Ways of Knowing
How do we know what we know? This article considers: (1) the ways we come to believe what we think we know (2) the many issues with the validation of our beliefs (3) the implications for building artificial intelligence and robots based on the human operating system.
Blog Post, Human Operating System 4 – Ways of Knowing, Rod Rivers, September 2017
Robot Ethics 2.0: From Autonomous Cars to Artificial Intelligence 1st Edition
Patrick Lin (Editor), Keith Abney (Editor), Ryan Jenkins (Editor)
Expanding discussions on robot ethics ‘means listening to new voices; robot ethics is no longer the concern of a handful of scholars. Experts from different academic disciplines and geographical areas are now playing vital roles in shaping ethical, legal, and policy discussions worldwide. So, for a more complete study, the editors of this volume look beyond the usual suspects for the latest thinking. Many of the views as represented in this cutting-edge volume are provocative–but also what we need to push forward in unfamiliar territory.’
Book, Robot Ethics 2.0: From Autonomous Cars to Artificial Intelligence 1st Edition, ISBN-13: 978-0190652951 ISBN-10: 0190652950, 440 pages, October 2017
Robot Ethics: The Ethical and Social Implications of Robotics
(Intelligent Robotics and Autonomous Agents series)
Patrick Lin (Editor), Keith Abney (Editor), George A. Bekey (Editor)
‘Starting with an overview of the issues and relevant ethical theories, the topics flow naturally from the possibility of programming robot ethics to the ethical use of military robots in war to legal and policy questions, including liability and privacy concerns. The contributors then turn to human-robot emotional relationships, examining the ethical implications of robots as sexual partners, caregivers, and servants. Finally, they explore the possibility that robots, whether biological-computational hybrids or pure machines, should be given rights or moral consideration.’
Book, Robot Ethics: The Ethical and Social Implications of Robotics, January 2014
Moral Machines: Teaching Robots Right from Wrong
1st Edition by Wendell Wallach (Author), Colin Allen (Author)
‘Computers are already approving financial transactions, controlling electrical supplies, and driving trains. Soon, service robots will be taking care of the elderly in their homes, and military robots will have their own targeting and firing protocols. Colin Allen and Wendell Wallach argue that as robots take on more and more responsibility, they must be programmed with moral decision-making abilities, for our own safety. Taking a fast paced tour through the latest thinking about philosophical ethics and artificial intelligence, the authors argue that even if full moral agency for machines is a long way off, it is already necessary to start building a kind of functional morality, in which artificial moral agents have some basic ethical sensitivity. But the standard ethical theories don’t seem adequate, and more socially engaged and engaging robots will be needed. As the authors show, the quest to build machines that are capable of telling right from wrong has begun.
Moral Machines is the first book to examine the challenge of building artificial moral agents, probing deeply into the nature of human decision making and ethics.’
Book, Moral Machines: Teaching Robots Right from Wrong, ISBN-13: 9780195374049, first published January 2009
Data is the fuel for AI
Newsletter Post, Data is the fuel for AI, so let’s ensure we get the ethics right, Birgitte Andersen, City AM Newsletter, December 2017
BBC News: MEPs vote on robots’ legal status – and if a kill switch is required
MEPs have called for the adoption of comprehensive rules for how humans will interact with artificial intelligence and robots.
The report makes it clear that it believes the world is on the cusp of a “new industrial” robot revolution.
It looks at whether to give robots legal status as “electronic persons”.
Designers should make sure any robots have a kill switch, which would allow functions to be shut down if necessary, the report recommends.
Meanwhile users should be able to use robots “without risk or fear of physical or psychological harm”, it states.
BBC News, MEPs vote on robots’ legal status – and if a kill switch isrequired, By Jane Wakefield, Technology reporter, January 2017
BBC News, Sex robots: Experts debate the rise of the love droids
Would you have sex with a robot? Would you marry one? Would a robot have the right to say no to such a union?
These were just a few of the questions being asked at the second Love and Sex with Robots conference hastily rearranged at Goldsmiths University in London after the government in Malaysia – the original location – banned it.
BBC News, Sex robots: Experts debate the rise of the love droids, By Jane Wakefield, Technology reporter, December 2016