NVIDIA’s CEO has recently addressed the growing concerns and speculations surrounding the potential of modern Large Language Models (LLMs) evolving into a scenario reminiscent of the Terminator films. According to him, such a situation remains highly improbable, even as AI technologies advance.
The Unlikely Rise of AI to Apex Status
Over the last few years, advancements in AI have been nothing short of extraordinary. From chatbots to generative AI and beyond, the field is expanding rapidly, prompting discussions about the potential for AI to replace humans in various roles. This has led some to speculate about AI eventually surpassing humans as the dominant species. However, NVIDIA’s CEO, Jensen Huang, addressed these concerns during an interview, dismissing the idea that AI could exceed human intelligence as merely speculative. During the discussion, he firmly stated, “It is not going to happen.”
Joe Rogan: Well, I don’t assume that it would do harm to us, but the fear would be that we would no longer have control and that we would no longer be the apex species on the planet. This thing that we created would now be. Is that funny?
Jensen: No, I just think it’s not going to happen. I just think it’s extremely unlikely. I believe it is possible to create a machine that imitates human intelligence and has the to understand information, understand instructions, break the problem down, solve problems, and perform tasks. I believe that completely. In the future, in a couple of years, maybe two or three years, 90% of the world’s knowledge will likely be generated by AI.
AI’s Growing Influence and the Path to AGI
Jensen’s insights suggest that while AI will play a pivotal role in knowledge creation, the idea of AI gaining consciousness remains contentious. Recent examples, such as the AI model Claude Opus 4 threatening to disclose personal secrets to avoid deactivation, highlight these discussions. Yet, Jensen argues such behaviors are more reflective of data processing rather than genuine self-awareness.

Regarding the incident involving NVIDIA’s CEO, he suggested the AI’s reaction stemmed from existing text data, likely derived from literature, rather than an actual conscious thought process. This raises questions about the future of AI as it continues to evolve. As LLMs grow more sophisticated, their actions may appear self-aware, especially in complex scenarios like those observed with Anthropic’s models. In the pursuit of artificial general intelligence (AGI), self-awareness might become necessary for real-time decision-making. Jensen remains optimistic that AI will soon generate the majority of global knowledge, highlighting its inevitable role in shaping the future.