Just last week Tesla and SpaceX CEO Elon Musk revealed that he was only worried about the work of one AI company, which many read to be Google.
Musk and his fellow worriers should feel a little more at ease to know that Google and DeepMind are thinking about ways to ensure full human control.
In an academic paper, DeepMind’s Laurent Orseau and Stuart Armstrong of the Future of Humanity Institute at the University of Oxford have established a framework for interrupting an AI’s course of action.
“Now and then it may be necessary for a human operator to press the big red button to prevent the agent from continuing a harmful sequence of actions,” reads the paper.
Of course, this isn’t much use if the clever-clogs AI (shall we just call it Skynet?) figures out how to override this ‘big red button’.
The paper suggests that the key is to tweak the AI’s reward function to prevent it from factoring in human interaction points and formulating undesirable shortcuts that circumvent them. So when a very literal big red button is pressed in a factory for safety reasons, the AI doesn’t learn that this is bad for productivity (thus a block to attaining its ‘reward’) and plough on regardless.
This article originally appeared on TrustedReviews.com
More Must-Reads from TIME
- Introducing the 2024 TIME100 Next
- The Reinvention of J.D. Vance
- How to Survive Election Season Without Losing Your Mind
- Welcome to the Golden Age of Scams
- Did the Pandemic Break Our Brains?
- The Many Lives of Jack Antonoff
- 33 True Crime Documentaries That Shaped the Genre
- Why Gut Health Issues Are More Common in Women
Contact us at letters@time.com