Professor Stephen Hawking gives the inaugural speech for Cambridge University’s newly-launched Centre for the Future of Intelligence.
The Leverhulme Centre for the Future of Intelligence (LCFI) kicked off yesterday at Cambridge University with opening remarks from professor Stephen Hawking and Artificial Intelligence (AI) pioneer professor Maggie Boden.
The LCFI’s mission is “to build a new interdisciplinary community of researchers, with strong links to technologists and the policy world, and a clear practical goal – to work together to ensure that we humans make the best of the opportunities of artificial intelligence, as it develops over coming decades.”
“Success in creating AI could the biggest event in the history of our civilization,” remarked Hawking, adding, “Perhaps with the tools of this new technological revolution, we will be able to undo some of the damage done to the natural world by the last one—industrialization.”
Professor Hawking, who has warned that AI will either be the best or worst thing for humanity, will also be making the opening remarks next month at Microsoft’s “Future Decoded” conference in London.
New projects on the nature and impact of AI being initiated and tackled by the LCFI include:
- Science, Value, and the Future of Intelligence (examines conceptions of value in the science of [artificial] intelligence)
- Policy and Responsible Innovation (focuses on the collaborations between technologists, policymakers, and other stakeholders needed for the responsible development of AI)
- The Value Alignment Problem (seeks to design methods for preventing AI systems from inadvertently acting in ways inimical to human values)
- Kinds of Intelligence (draws on current work in psychology, neuroscience, philosophy, computer science, and cognitive robotics in order to further develop and critically assess notions of general intelligence used in AI)
- Autonomous Weapons – Prospects for Regulation (aims to bring an interdisciplinary approach to the question of regulating autonomous weapons systems)
- AI: Agents and Persons (explores the nature and future of AI agency and personhood, and its impact on our human sense of what it means to be a person)
- Politics and Policy (examines the challenges that the future of AI poses for democratic politics, including questions of political agency, accountability and representation)
- Trust and Transparency (developing processes to ensure that AI systems are transparent, reliable and trustworthy)
- Horizon Scanning and Road-Mapping (aims to map the landscape of potential AI breakthroughs and their social consequences)
With the inauguration of the LCFI, the research center is the latest organization to tackle the global challenge of humans vs machines and best AI practices.
Research into the future of AI is exploding! OpenAI, the Partnership on AI, and the White House’s new report have all independently begun to explore the potentials of AI with very similar aims.
Read More: Partnership on AI vs OpenAI: Consolidation of Power vs Open Source
With so many different organizations looking to not only research the science of AI, most are also looking towards philosophy, politics, and policy making.
One of the key variations to look for in the future is the question of Open Source vs Privately-held research. Whose research will be the one policy makers go to? Whose projects will governments wish to implement? And who will become the ethical and moral authority between private corporations, government think tanks, and academic institutions?