LOGIN | Register
Cooperation
CES 2026 英伟达CEO黄仁勋发表演讲
文章语言:
EN
Share
Minutes
原文
会议摘要
Nvidia is spearheading the shift from traditional computing to AI-centric paradigms, highlighted by advancements in AI supercomputing, open-source models, and AI integration across sectors. The unveiling of Nvidia's Vera Rubin AI supercomputer showcases its commitment to pioneering AI technology, emphasizing energy efficiency, security, and scalability for future industries.
会议速览
Revolutionizing Computing: Nvidia's Vision of AI's Impact on the Industry
The dialogue outlines Nvidia's perspective on the transformative shift in computing driven by AI, highlighting the emergence of agentic systems and the significant investment flowing into AI-driven innovations, signaling a fundamental reinvention of the computing industry.
Revolutionizing AI: Nvidia's Open Models and Physical AI Breakthroughs
Nvidia pioneers open-source AI models, enhancing global innovation, and excels in physical AI, impacting sectors like biology, robotics, and weather prediction, setting new industry standards.
Pride in World-Class AI Models for Multimodal Content Interpretation and Semantic Search
The dialogue highlights the company's world-class AI models, including PDF retrievers, speech recognition, and semantic search capabilities, which are essential for understanding multimodal documents and building AI agents, positioning them at the top of industry leaderboards.
Revolutionizing AI: Customizable, Multimodal, and Multi-Model Agents for Enhanced Reasoning and Application
The dialogue explores the advancements in AI models, focusing on their ability to reason, adapt, and integrate multiple models for enhanced problem-solving. It highlights the customizable nature of AI applications, their multimodal capabilities, and the use of intent-based model routing for optimized performance. Examples include building a personal assistant that can manage emails, interact with the physical world, and utilize specialized models for specific tasks, all while ensuring privacy and efficiency.
Daily Tasks and Script Update for Collaborative Completion
Tasks include grocery shopping for essentials and sending a script update by day's end, emphasizing teamwork and deadlines.
Transforming Sketches to Architectural Renderings and Virtual Tours with Open Source Tools
A sketch is transformed into an architectural rendering, followed by a virtual tour of the room. Access is shared for collaboration, highlighting the progress in open source tools for creative projects.
Revolutionizing AI Application Development with Pre-Trained Language Models and Agentic Frameworks
The dialogue highlights the transformative impact of integrating pre-trained language models into agentic frameworks for rapid AI application creation. It emphasizes the capability to reason through unseen data and solve problems efficiently, showcasing the evolution from the unimaginable to the trivial in just a few years.
Revolutionizing Enterprise AI: Integrating Advanced Platforms for Multimodal Interaction
Leading enterprise platforms like Palantir, ServiceNow, and Snowflake are being transformed by AI integration, particularly with Nvidia's technology, enabling a more human-like, multimodal user interface that simplifies interaction and revolutionizes application development and platform usability.
Revolutionizing AI with Physical Understanding: From Simulation to Real-World Interaction
The dialogue explores the development of physical AI, emphasizing the creation of a system that understands the physical world through simulation, data generation, and foundation models, enabling AI to learn and interact with reality effectively.
Revolutionizing Physical AI: Nvidia Cosmos Unleashes Synthetic Data Power for Diverse and Predictive Real-World Applications
Nvidia Cosmos, a pioneering foundation model, harnesses synthetic data to train physical AI, enabling realistic video generation, coherent motion simulation, and predictive analysis. This open frontier world model learns from internet-scale data, real driving, robotics, and 3D simulations, offering developers interactive closed-loop environments to explore edge cases and reason through complex scenarios.
Revolutionizing Autonomous Driving: Cosmos Model Trains AI to Reason and Adapt for Every Scenario
Cosmos, a groundbreaking foundation model, has been instrumental in developing Alpha Mayo, the world's first reasoning autonomous vehicle AI. Trained using human demonstrations and Cosmos-generated data, Alpha Mayo not only drives naturally but also reasons about its actions, addressing the long tail of driving scenarios by decomposing them into manageable parts. This advancement significantly enhances the capability of autonomous vehicles to handle diverse driving situations.
Arrival and Precautionary Measure
Upon arrival, a warning to secure safety restraints is issued, emphasizing preparedness for what lies ahead.
NVIDIA's Vision for Autonomous Driving: A Comprehensive AI-Powered Stack from Chips to Cars
Exploring NVIDIA's pioneering efforts in developing a full-stack solution for autonomous vehicles, featuring advanced AI models, dedicated processors, and strategic partnerships with automotive giants. The initiative aims to revolutionize the transportation sector, ensuring safety and efficiency in future mobility solutions.
Revolutionizing Autonomous Vehicles: Dual Safety-Certified Stacks and Open Ecosystem
The dialogue highlights a pioneering approach to autonomous vehicle technology, featuring dual safety-certified software stacks designed for redundancy and safety. The first stack, Alpha Mile, is trained end-to-end, while the second, classical Av stack, serves as a fail-safe. This vertically integrated system, open for ecosystem collaboration, aims to ensure safety and reliability in autonomous driving, with a vision for every car and truck to be autonomous in the future. The technology is already a significant business, facilitating training data processing, synthetic data generation, and chip development, with opportunities for companies to engage at various levels of integration.
Autonomous Vehicles & Robots: The Future of AI-Powered Systems
The dialogue highlights the imminent shift towards autonomous vehicles and robotic systems, emphasizing the use of synthetic data and simulation techniques. It envisions a future where robots, in various forms and sizes, will play a significant role, suggesting an upcoming era dominated by AI-powered robotic advancements.
Revolutionizing Robotics and Chip Design with AI and Simulation
A discussion on integrating AI and simulation technologies, like Isaac Sim and Nvidia's tools, into robotics and semiconductor design. Highlights include advancements in robot creation, chip design with Cadence and Synopsis, and Siemens' integration of these technologies. The future envisions agentic designers aiding in the creation of robots and systems, emphasizing computational design and testing before physical production.
Revolutionizing Industries: Nvidia and Siemens Unite for AI-Powered Automation
Nvidia and Siemens collaborate to integrate AI technologies into physical industries, addressing the global labor shortage and advancing automation through physical AI and robotics, marking a significant shift in industrial innovation.
Revolutionizing Industry: Nvidia and Siemens Unite for Physical AI
Nvidia and Siemens are pioneering a new industrial era by integrating physical AI throughout the industrial lifecycle, from design to operations, heralding a transformative age for industries.
OpenAI's Dominance in Token Generation vs. Future of Open Source Models
OpenAI currently leads in token generation, but open source models, with their diverse research and domains, are expected to surpass in scale over time.
Revolutionizing AI Computation: Introducing Vera Rubin, the Next-Gen Supercomputer
A groundbreaking supercomputer, Vera Rubin, has been developed to tackle the escalating computational demands of AI, coinciding with the unveiling of its production amidst the AI advancement race.
Revolutionizing AI: The Genesis of the Vera Rubin Supercomputer and Its Unprecedented Computational Power
The narrative details the creation of the Vera Rubin AI supercomputer, highlighting its core components including Vera CPU, Rubin GPU, Bluefield 4 DPU, and Mv link switch. It underscores the system's advanced co-design, achieving 100 petaflops of AI performance, enhanced data connectivity, and scalability to thousands of racks, marking a monumental advancement in AI technology.
Revolutionizing Chip Design: Extreme Co-Design for Future-Proof Technology
A company has designed six new chips, emphasizing extreme co-design to keep up with the rapid advancement of technology and the slowing of Moore's Law. The Vera CPU, featuring 88 multi-threaded cores, offers twice the performance per watt of the world's most advanced CPUs, while the Ruben GPU, connected to the CPU, boasts incredible single-threaded performance and memory capacity. This approach involves redesigning every chip to handle the increasing demands of larger models and token generation, ensuring the industry can continue to advance.
Revolutionizing GPU Performance: Extreme Co-Design and Dynamic Precision Adjustment
Discusses advancements in GPU technology, emphasizing extreme co-design and the introduction of MVFP4 Tensor Core, enabling dynamic precision adjustment for enhanced performance with limited transistor increases.
Revolutionizing Data Centers: Nvidia's Spectrum X, Bluefield 4, and NVLink 6 Switch
Nvidia's Spectrum X Ethernet, Bluefield 4 processor, and NVLink 6 switch are transforming data center operations by enhancing networking efficiency, enabling virtualization, and accelerating data processing. These innovations, including programmable RM DMA and data path accelerators, significantly improve throughput and security in large-scale AI data centers, offering unparalleled performance and flexibility.
Revolutionizing Supercomputing: High-Speed GPU Networking and Energy-Efficient Cooling
The dialogue highlights a groundbreaking supercomputing system featuring 400 Gb/s switches enabling GPUs to communicate at peak performance. It discusses the system's components, including two miles of copper cables and an innovative cooling mechanism using hot water, achieving energy efficiency and standardization across major computer manufacturers.
Revolutionizing Data Centers: NVIDIA's Spectrum X with Integrated Silicon Photonics
NVIDIA introduces the Spectrum X Ethernet switch, leveraging TSMC's new Coupe process for integrated silicon photonics, enabling 512 ports at 200 Gb/s. This innovation connects lasers directly to the chip, enhancing data center efficiency. Building on the success of Spectrum XX, Spectrum X aims to redefine networking by combining low-latency performance with Ethernet's manageability, positioning NVIDIA as a leader in the networking industry amid AI-driven computing advancements.
Revolutionizing AI Context Memory Storage with Blue Field Fort and Grace Blackwell
AI's growing need for context memory storage is addressed by NVIDIA's Blue Field Fort and Grace Blackwell, integrating fast context memory within racks to handle increasing data demands efficiently, reducing network congestion and enhancing energy efficiency in data centers.
Revolutionizing Confidential Computing & AI Efficiency: Advancements in Power, Security, and Performance
A confidential computing system, ensuring data safety during transit, rest, and compute, has been developed with every bus encrypted for unparalleled security. This system also introduces power smoothing technology, drastically reducing energy waste and enabling efficient use of the entire power budget. Performance-wise, it showcases significant improvements in training AI models, factory throughput, and cost-effectiveness in token generation, pushing the boundaries of AI efficiency and technology leadership.
NVIDIA's Comprehensive Approach to AI Innovation
Discusses NVIDIA's role in creating the entire AI stack from chips to applications, emphasizing their commitment to enabling others to build impactful AI applications.
Balancing AI Complexity and Practicality in Everyday Solutions
Discusses the importance of selecting appropriate AI models for tasks, emphasizing that not every problem requires the largest or smartest model. Highlights the development of an AI supercomputer platform and addresses practical issues, such as dealing with unexpected distractions like squirrels, while maintaining focus on effective AI deployment.
要点回答
Q:What are the capabilities of the AI system described in the speech?
A:The AI system is completely customizable, allowing the user to teach it specific skills relevant to their company and domain. It is also always at the forefront of technology, combining deep domain expertise with modern AI advancements.
Q:What framework is used to build applications, and how is it integrated into enterprise platforms?
A:The framework used to build applications is called a blueprint, and it is integrated into enterprise SaaS platforms all over the world, allowing users to construct and utilize AI systems easily.
Q:How is the AI system designed to manage different tasks?
A:The AI system uses an intent-based model router to determine the best model for each application or problem, ensuring that tasks like emails remain private and are processed locally, while other jobs can call upon a broader range of models.
Q:What is the significance of the physical AI advancements mentioned?
A:Physical AI advancements are significant because they involve taking intelligence that interacts with screens and speakers and applying it to the physical world, understanding common sense, object permanence, causality, and physical laws. It enables AI to interact with the real world more naturally and intelligently.
Q:What are the challenges in developing physical AI and how are they addressed?
A:Challenges in developing physical AI include the scarcity of data, the need for simulation to evaluate AI actions, and the diversity and unpredictability of the real world. These challenges are addressed by using synthetic data generation grounded in physical laws, the use of foundation models like Nvidia's Cosmos, and creating AI that can reason about its actions and their consequences.
Q:What is the role of the Nvidia Foundation model, Cosmos, in physical AI?
A:The role of the Nvidia Foundation model, Cosmos, in physical AI is to serve as a unified representation of the world, capable of aligning language, images, 3D, and action. It is trained on internet-scale video, real driving, and robotics data, as well as 3D simulation, enabling it to perform physical AI skills and generate realistic scenarios for AI training.
Q:What is the new capability announced for autonomous vehicles, and how does it differ from previous approaches?
A:The new capability announced for autonomous vehicles is Alpha Mayo, the world's first thinking, reasoning, autonomous vehicle AI. It differs from previous approaches by being trained end-to-end from camera input to actuation output, reasoning about its actions, and providing explanations for those actions, all coupled directly with specific training data.
Q:How does Alpha Mayo contribute to handling complex driving scenarios?
A:Alpha Mayo contributes to handling complex driving scenarios by breaking them down into smaller, more normal circumstances that the car is programmed to understand. It reasons about each situation, determines the appropriate action, and communicates its rationale, which helps in managing the long tail of driving scenarios that are difficult to cover with traditional methods.
Q:What are the five layers of AI technology mentioned in the speech?
A:The five layers of AI technology include land, power, and shell as the lowest layer, followed by chips such as GPUs and networking chips like CPUs. The third layer is the infrastructure, with specific mention of physical AI environments like Omniverse and Cosmos. The fourth layer is the models, and the fifth layer is applications like the Mercedes Benz vehicle.
Q:What does the speaker predict about the future of the robotics industry and autonomous vehicles?
A:The speaker predicts that the future will see a proliferation of autonomous vehicles, with the potential for a billion cars on the road to be autonomous within the next few years. The speaker believes that these vehicles could be robo taxis, owned and driven by individuals, or driven by themselves, but all will have the capability to operate autonomously and be powered by AI. This is anticipated to be one of the largest robotics industries.
Q:What is the significance of the first Nvidia AV car mentioned in the speech?
A:The significance of the first Nvidia AV car is that it is a milestone in the development of autonomous vehicles, scheduled to go on the road in Q1, then to Europe in Q2, and to the United States in Q3 and Q4. It signifies the integration of AI into everyday vehicles and is part of Nvidia's ongoing efforts to advance autonomous driving technology.
Q:What are the implications of the recent NCAP rating for the Mercedes Benz CLA?
A:The recent NCAP rating for the Mercedes Benz CLA as the world's safest car implies that the vehicle, and by extension, Nvidia's technology, has achieved a high standard of safety. The rating is significant as it attests to the reliability and robustness of the model system, which is based on diverse and redundant sensors and a self-driving car stack certified with every line of code.
Q:What is the function of the second software stack mentioned, and why is it important?
A:The second software stack is designed to be fully traceable and plays a crucial role in ensuring the safety of autonomous vehicles. It mirrors the primary software stack and is built to provide a safety net in uncertain situations, where the system can switch to a simpler, safer mode if needed. This second stack is important because it helps mitigate risks and contributes to the overall safety of the autonomous vehicle system.
Q:How is the entire Nvidia stack described in terms of vertical integration?
A:The entire Nvidia stack is described as vertically integrated, meaning that Nvidia built the stack together with its partners, such as Mercedes Benz in the case of the CLA model. Nvidia plans to deploy the car, operate the stack, and maintain it for as long as necessary. However, the stack is open for the ecosystem to use, allowing for collaboration in developing Level 4 autonomous vehicles and robo-taxis.
Q:What advancements in AI and robotics are being highlighted by the speaker?
A:The speaker highlights several advancements in AI and robotics, including the use of NVIDIA's Isaac simulation software to train robots, the development of various types of robots by different companies, and the integration of AI into robotic systems. The speaker emphasizes that these developments are part of the next era for robotic systems and are expected to revolutionize industries such as chip design and manufacturing.
Q:What partnerships are mentioned as part of the ecosystem for physical AI and robotics?
A:Partnerships mentioned include collaborations with companies like Ola, Mercedes, Siemens, and Cadence. These partnerships involve integrating Nvidia technology into various applications, such as cars and chip design tools, to revolutionize industries by bringing physical AI to the full industrial life cycle, from design to production and operations.
Q:How has the demand for Nvidia GPUs changed over time?
A:The demand for Nvidia GPUs has been increasing exponentially, with models increasing by an order of magnitude every year. This demand is a result of the rising computational needs for AI and the advent of reinforcement learning.
Q:What is the significance of the Vera Rubin's production status for AI advancement?
A:The significance of Vera Rubin being in full production is that it allows for timely advancement to the next frontier of AI. The rapid pace of technological progress in AI requires constant improvement in computational capability, and Vera Rubin's timely production is crucial for staying at the forefront of this technological race.
Q:What technological innovations were made in the architecture of Vera Rubin?
A:The architectural innovations of Vera Rubin include a system of six chips engineered to work as one, featuring a custom designed CPU called Vera that is double the performance of the previous generation and a Rubin GPU. Vera and Rubin are co-designed to bidirectionally and coherently share data faster with lower latency. The compute tray is completely redesigned with no cables, hoses, or fans, integrating a Bluefield 4 DPU, Vera CPU, and Rubin GPUs. Additionally, the MVL 72 rack, with its 6 breakthrough chips, 18 compute trays, and other components, represents a major leap to the next frontier of AI.
Q:What is the importance of the Vera Rubin's MVF P 4 Tensor Core?
A:The Vera Rubin's MVF P 4 Tensor Core is important because it is an entire processing unit within the chip that dynamically adapts its precision and structure to deal with different levels of the transformer. This allows for higher throughput while intelligently managing precision needs, which is crucial for achieving performance levels that would otherwise be impossible given the number of transistors available.
Q:How has the assembly process of the Vera Rubin nodes been simplified?
A:The assembly process of the Vera Rubin nodes has been simplified by eliminating cables, reducing the complexity and time required for assembly. The nodes consist of only two tubes and take two hours to assemble, a stark contrast to the previous design. This innovation drastically reduces the potential for errors and the time needed to set up the system.
Q:What makes the Spectrum X networking technology revolutionary for AI?
A:The Spectrum X networking technology is revolutionary for AI because it is specifically designed to handle the intense traffic and low latency requirements of AI. By providing exceptional performance, especially in large data centers, Spectrum X enables significant increases in throughput without the need for additional expensive hardware. This makes it highly sought after and an industry standard for high-performance computing.
Q:What are the characteristics of the nvlink spines in the computing system?
A:The nvlink spines consist of two miles of copper cables, which are the best conductor known to mankind and are structured to be the most extensive use of copper cables in computing systems worldwide. These cables are shielded and allow for data transmission at 400 Gb per second.
Q:What revolutionized the computing industry and why is standardization on components crucial?
A:The development of an industry standard system revolutionized the computing industry. It is crucial for standardization on components because there are around 80,000 different components that make up the Mx systems, and frequent changes would lead to significant waste and inefficiency.
Q:How do the new computing systems compare to previous ones in terms of performance, energy efficiency, and water usage?
A:The new computing systems are more energy efficient, with double the performance and power of previous systems like Grace Blackwell while maintaining the same air flow and water temperature (45 degrees C) without requiring water chillers. This makes the new system highly efficient and eco-friendly.
Q:What is Vera Rubin, and what are its significant features and performance benefits?
A:Vera Rubin is a highly efficient computing system that is 1.7 times more transistors, with 5 times more peak inference performance and 3.5 times more peak training performance. It features integrated silicon photonic technology and has energy efficiency benefits, being twice as energy efficient and having the ability to save 6% of the world's data center power.
Q:What are the benefits of the new chip introduced using TSMC's process called N5P?
A:The new chip using TSMC's N5P process integrates silicon photonic technology directly onto the chip, enabling 512 ports at 200 Gb per second and significantly enhancing performance and efficiency.
Q:What is the purpose of the new Ethernet AI switch, and how does it integrate with other data center components?
A:The purpose of the new Ethernet AI switch is to facilitate high-speed connectivity within data centers. It integrates with the rest of the data center through large chips that have silicon photonics directly connected to them, allowing for efficient laser connections to the rest of the network.
Q:What was the motivation behind introducing Spectrum X, and how has it affected the company's market position?
A:Spectrum X was introduced to reinvent networking and leverage the simplicity and manageability of Ethernet, which is universally understood and used in data centers. This move made the company the largest networking company in the world.
Q:How is artificial intelligence changing the way storage is managed?
A:Artificial intelligence is changing storage management by moving away from traditional SQL to using semantics information and temporary memory called Kv cache. This creates challenges in managing and storing the increasing amount of data associated with AI models, leading to the need for more efficient memory solutions.
Q:What challenges does Grace Blackwell aim to address, and what is its connection to Hopper?
A:Grace Blackwell aims to address the limitations of HBM memory in storing the rapidly growing context memory needed for AI applications. To expand context memory, Grace Blackwell is connected directly to Hopper, creating a fast and efficient solution for AI-related data handling.
Q:What is bluefield fort, and how does it help with managing the Kv cache in AI operations?
A:Bluefield fort is a new category of storage systems designed to provide a fast Kv cache context memory store right in the rack. This solution addresses the issue of network congestion caused by the movement of data in AI operations, enabling more efficient handling of temporary AI memories.
Q:What are the key performance metrics highlighted for Vera Rubin?
A:Key performance metrics for Vera Rubin include training AI models faster, enabling quicker delivery of technology leadership and pricing power. It also boasts higher factory throughput, directly impacting data center revenues, and cost-effective token generation.
Q:What is Nvidia's approach to AI across its offerings?
A:Nvidia's approach to AI spans across chips, infrastructure, models, and applications, aiming to reinvent AI to create the entire stack for developers to build incredible applications. The company's job is to provide the full range of tools and resources needed to advance AI innovation.
play
English
English
进入会议
1.0
0.5
0.75
1.0
1.5
2.0