The AI car problem: to kill the child or the elderly person?

The AI car problem: to kill the child or the elderly person?

Would you kill one to save many? For centuries philosophers have been discussing this question for which there seems to be no objective answer at all. Or is there?

Researchers from MIT, Harvard University, University of British Columbia and Université Toulouse Capitole tested millions of people from 200 countries from all over the world in an experiment called “The Moral Machine” on their moral-driven decisions. Participants were confronted with a scenario based on the famous trolley problem which allows the person control over a lever which a trolley is moving towards. On each track a person, an animal or a group is blocking the route, hence the participant has to decide: Do I switch? Whom do I kill?

According to The Hacker News, the research is meant to help in the development of algorithms for AI-driven cars.

Read more about if AI cars are going to decide for or against you in future here. 

Copyright Lyonsdown Limited 2020

Top Articles

How the human immune system inspired a new approach to cyber-security

Artificial intelligence is being used to understand what’s ‘normal’ inside digital systems and autonomously fight back against cyber-threats

Solarwinds CEO blames former intern for hilarious password fiasco

SolarWinds has accused a former intern of creating a very weak password for its update server and storing it on a GitHub server for months.

Hackers accessed Npower customer accounts via a credential-stuffing attack

Npower suffered a major data breach that involved hackers using stolen passwords to gain access to a large number of customer accounts.

Related Articles