LADYBIRD, the animal-friendly vacuum cleaning robot, was conceived in 2014 by Oliver Bendel and introduced at Stanford University (AAAI Spring Symposia) in 2017 and then implemented as a prototype at the School of Business FHNW. In the context of the project, a menu was proposed with which the user can set the morale of the vacuum cleaner robot. As the name implies, it spares ladybirds. It should also let spiders live. But if you want to have certain insects sucked in, you could define this via a menu. It is important that LADYBIRD remains animal-friendly overall. The idea was to develop a proxy morality in detail via the menu. The vacuum cleaner robot as a proxy machine does what the owner would do. In 2018 the moral menu (MOME) for LADYBIRD was born as a design study. It can be combined with other approaches and technologies. In this way, the user could learn how others have decided and how the personal morality that he has transferred to the machine is assessed. He could also be warned and enlightened if he wants to suck in not only vermin but also spiders.
Fig.: Moral menu for LADYBIRD
The LADYBIRD project starts in March 2017. More and more autonomous and semi-autonomous machines make decisions that have moral implications. Machine ethics as a discipline examines the possibilities and limits of moral machines. In this context, Prof. Dr. Oliver Bendel developed various design studies and thus submitted proposals for their appearance and functions. He focused on animal-friendly machines which make morally sound decisions, and chatbots with specific skills. The project is about a service robot, which shall spare beneficial insects – a vacuum cleaner called LADYBIRD. An annotated decision tree modelled for this objective and a set of sensors will be used. Three students work from March to October in the practice project at the School of Business FHNW. Since 2013, the principal Oliver Bendel published several articles on LADYBIRD and other animal-friendly machines, e.g., “Ich bremse auch für Tiere (I also brake for animals)” in the Swiss magazine inside.it. The robot will be the third prototype in the context of machine ethics at the School of Business FHNW. The first one was the GOODBOT (2013), the second one the LIEBOT (2016). All these machines can be described as simple moral machines.
Fig.: The robot should spare the ladybird
The LIEBOT project is based on preparatory works by Prof. Dr. Oliver Bendel who already initiated the GOODBOT, a chatbot which can be considered a simple moral machine. Since 2013 he has published several articles on this subject and presented automatic strategies with a view to lying. A business informatics student of the School of Business FHNW, Kevin Schwegler, was contracted early in 2016 to implement the LIEBOT (aka LÜGENBOT) as a prototype in the scope of his graduation thesis under consideration and continuance of the preparatory works. The objective of the LIEBOT project is to give practical evidence of the potential of lies and risks of natural language systems. Online media and websites create or aggregate more and more texts automatically (robo-content) and robo-journalism is growing. Natural language dialog systems are becoming very popular. The chatbot to be programmed – the LIEBOT – shall be able to produce untruths and to respond in a morally inadequate manner. This makes it a reversion of the premise applied to the development of the GOODBOT and a continuance of the corresponding work under new auspices. A simple immoral machine is born, a kind of a Munchausen machine. Further information as from autumn 2016 via luegenbot.ch and liebot.org.
Fig.: The adventures of Baron Munchausen
The proceedings of the AAAI conference 2016 have been published in March 2016 (“The 2016 AAAI Spring Symposium Series: Technical Reports”). The symposium “Ethical and Moral Considerations in Non-Human Agents” was dedicated to the discipline of machine ethics. Ron Arkin (Georgia Institute of Technology), Luís Moniz Pereira (Universidade Nova de Lisboa), Peter Asaro (New School for Public Engagement, New York) and Oliver Bendel (School of Business FHNW) spoke about moral and immoral machines. The contribution “Annotated Decision Trees for Simple Moral Machines” (Oliver Bendel) can be found on the pages 195 – 201. In the abstract it is said: “Autonomization often follows after the automization on which it is based. More and more machines have to make decisions with moral implications. Machine ethics, which can be seen as an equivalent of human ethics, analyses the chances and limits of moral machines. So far, decision trees have not been commonly used for modelling moral machines. This article proposes an approach for creating annotated decision trees, and specifies their central components. The focus is on simple moral machines. The chances of such models are illustrated with the example of a self-driving car that is friendly to humans and animals. Finally the advantages and disadvantages are discussed and conclusions are drawn.” The proceedings can be ordered via www.aaai.org.
Fig.: Oliver Bendel, Cindy Mason, Luís Moniz Pereira and others in Stanford
Auf der Plattform KMU Business World ist am 30. August 2013 der Artikel “Einfache moralische Maschinen” von Oliver Bendel erschienen. Eine Vorabveröffentlichung brachte der Blog des Instituts für Wirtschaftsinformatik der Hochschule für Wirtschaft FHNW. Moralische Maschinen sind nach der Meinung des Philosophen und Wirtschaftsinformatikers bereits heute möglich. Man dürfe sie nur nicht zu kompliziert denken und zu komplex gestalten. Wenn es sich um spezialisierte autonome oder teilautonome Systeme handele, könne man bei ihren Kernfunktionen ansetzen und sie mit geringfügigen Änderungen, mit wenigen zusätzlichen Komponenten in einfache moralische Maschinen verwandeln. Wie das geschehen kann, wird an vier Beispielen gezeigt, an Servicerobotern, privaten Drohnen, Chatbots und selbstständig fahrenden Autos.
Abb.: SpaceJustin vom DLR German Aerospace Center (Quelle: http://commons.wikimedia.org)