Two of the world’s leading AI experts – Yann LeCun, Yoshua Bengio, and Geoffrey Hinton – have raised the alarm on the development of artificial general intelligence (AGI). In a recent interview with Global News, Bengio expressed concern about the hypothetical AI model that could reason through any task a human could. Both Bengio and Hinton have spent their careers pushing the boundaries of AI, but they are both acutely aware of the potential dangers of developing AI beyond human intelligence.
Concerns about the development of artificial general intelligence (AGI)
While AI has advanced significantly over the past decade, it is still far from being able to reason like a human. However, with the development of AGI, a hypothetical AI model capable of reasoning through any task that a human can, the situation could change rapidly. Bengio and Hinton have noted the risk that AGI (Artificial General Intelligence) could represent a fundamental turning point in human history, compelling us to consider the possibility of non-human entities with superior intelligence. “If humans lose our edge as the most intelligent beings on Earth, how do we survive that?” Hinton once asked in an interview with MIT Technology Review.
Counterarguments to AI doomsayers
Some critics of AI doomsayers believe that concerns about the rapid progress of AI are unfounded. They argue that the development of AGI will take much longer than people think, if it is even feasible at all. However, Bengio and Hinton maintain that the consequences of AGI development are too significant to ignore. They assert that AI development requires more investment in research to understand and prevent the risks associated with the technology.
Possibility of international cooperation on AI regulation
The hope among some AI doomsayers is that something similar to the international cooperation on nuclear disarmament and human cloning could play out again with a consensus on AI. While AI development has primarily occurred in Western liberal democracies, China and Russia have also shown significant interest in developing AI. However, international cooperation on AI regulation has been slow to develop so far. The European Union is taking the lead in addressing the AI wild west with an act that could become the most comprehensive AI regulatory framework yet.
The EU’s comprehensive regulatory framework for AI
Under the EU’s new regulations, some AI applications are promised to be outright banned, such as real-time facial recognition systems in public and predictive policing. There will also be strict regulations on high-risk AI applications, including medical diagnoses, crime detection, and public safety. The EU’s AI regulatory framework is a significant first step, but the international community needs to create a more comprehensive dialogue on AI regulation.
There is a global interest in AI regulation
It’s clear that AI regulation is on the agenda of the world’s powers, not just Western liberal democracies. Currently, the US government is focused on supporting the advancement of AI while also mitigating its potential risks. Meanwhile, China’s strategic focus on becoming a world leader in AI makes it one of the biggest players globally. Like the EU, China is developing AI regulations that provide a framework for the safe development and deployment of AI.
In conclusion, the development of AGI is not a matter of if, but when. Bengio’s and Hinton’s concerns over AGI should be heeded, and investment in understanding and mitigating AI risks is essential. Regulations like the EU’s AI regulatory framework need to be implemented worldwide to ensure the safe deployment of AI technology. At the same time, global efforts to address the root causes of social and economic inequality could alleviate some of the potential risks of AI development.