Morality algorithm proves AI can also be friendly


We’re used to seeing headlines of AI beating us at our own games in adversarial roles, but a new study has proven they can also excel when it comes to cooperation and compromise.

The study, from an international team of computer scientists, found AI can be programmed with a higher degree of morality than humans. The researchers set out to build a new type of algorithm for playing games which requires working together rather than simply winning at all cost.

Jacob Crandall, BYU Computer Science Professor and lead author of the study, comments:

“The end goal is that we understand the mathematics behind cooperation with people and what attributes artificial intelligence needs to develop social skills. AI needs to be able to respond to us and articulate what it’s doing. It has to be able to interact with other people.”

In many real world applications, AI will have to compromise and cooperate with both humans and other machines. The ability to program AIs with friendly traits is unlikely to come as much surprise, but evidence they can express them better than humans opens up new possibilities.

Crandall and his team created an algorithm called S# and tested its performance across a variety of two-player games. In most cases, the machine outperformed humans in the games.

“Two humans, if they were honest with each other and loyal, would have done as well as two machines,” says Crandall. “As it is, about half of the humans lied at some point. So essentially, this particular algorithm is learning that moral characteristics are good. It’s programmed to not lie, and it also learns to maintain cooperation once it emerges.”

Take the current negotiations between Britain and the EU as the former exits the bloc. Both sides claim to want an ‘orderly exit’, but it’s clear human emotions are involved which keeps leading to deadlocks in the talks despite time running out. AI negotiators could run through all the scenarios and find the best areas to compromise without any feelings of mistrust.

“In society, relationships break down all the time,” he said. “People that were friends for years all of a sudden become enemies. Because the machine is often actually better at reaching these compromises than we are, it can potentially teach us how to do this better.”

What are your thoughts on AI morality? Let us know in the comments.

 Interested in hearing industry leaders discuss subjects like this and sharing their use-cases? Attend the co-located AI & Big Data Expo events with upcoming shows in Silicon Valley, London and Amsterdam to learn more. Co-located with the  IoT Tech Expo, Blockchain Expo and Cyber Security & Cloud Expo so you can explore the future of enterprise technology in one place.

Click to comment

You must be logged in to post a comment Login

Leave a Reply

To Top

We are using cookies on our website

We use cookies to personalise content and ads, to provide social media features, and to analyse our traffic. Please confirm if you accept our tracking cookies. You are free to decline the tracking so you can continue to visit our website without any data sent to third-party services. All personal data can be deleted by visiting the Contact Us > Privacy Tools area of the website.