NVIDIA's Platform Strategy NVIDIA's strategy is built on three platforms: CUDA X, Systems, and AI Factories. The company emphasizes its ecosystem and its role as a platform provider. The five-layer cake of AI is covered, from infrastructure and chips to platforms, models, and applications. CUDA: The Foundation of NVIDIA's Ecosystem Celebrating 20 years of CUDA, a platform for accelerated computing. CUDA's architecture (SIMD) and its evolution with tensor cores are highlighted. The massive installed base of hundreds of millions of CUDA-enabled GPUs is a key advantage, driving a flywheel effect of developer attraction and innovation. GeForce's role in introducing CUDA to the world and nurturing future developers is acknowledged. RTX, introduced about 8-10 years ago, represents a redesign for modern computer graphics and fused with AI. Neural Rendering and Structured Data Introducing Neural Rendering, a fusion of 3D graphics and AI, exemplified by DLSS 5. This fusion combines controllable 3D graphics (structured data) with generative AI (probabilistic computing). Structured data is presented as the foundation of trustworthy AI. AI's Impact on Data Processing AI will increasingly utilize structured data (SQL, Spark, Pandas) and unstructured data (vector databases, PDFs, videos). NVIDIA introduces two foundational libraries: QDF for data frames (structured data) and QVS for vector stores (unstructured data). Partnerships with IBM (Watson X.Data), Dell (AI Data Platform), and Google Cloud (Vertex AI, BigQuery) are showcased to accelerate data processing. Accelerated computing is presented as the successor to Moore's Law, enabling significant leaps in performance and cost reduction. Cloud Partnerships and Confidential Computing NVIDIA integrates with major cloud providers: Google Cloud, AWS, and Microsoft Azure. Key collaborations include bringing OpenAI to AWS, accelerating EMR and SageMaker. Confidential computing with vGPUs is highlighted for secure deployment of AI models. Partnerships extend to enterprise solutions with Oracle, CoreWeave, and Palantir/Dell for on-premise AI platforms. NVIDIA's Vertical Integration and Horizontal Openness NVIDIA operates as a vertically integrated but horizontally open company. This approach is necessary for application acceleration across various domains. The company focuses on understanding applications, domains, and algorithms to deploy solutions across data centers, cloud, edge, and robotics. Industry Verticals and CUDA X Libraries NVIDIA's presence spans numerous industries: automotive, financial services, healthcare, industrial, media & entertainment, quantum, retail, robotics, and telecommunications. CUDA X libraries are presented as NVIDIA's "crown jewels," activating computing platforms for specific problems. Examples of CUDA X libraries include QOpt, QLitho, QDSS, QEquivariance, Arial, Warp, and Parabricks. The Rise of AI Natives and the Compute Demand The last two years have seen a surge in investment in AI startups ("AI natives"). This boom is driven by the immense compute demand from these companies, either creating or integrating with tokens. The shift from retrieval-based to generative computing, powered by LLMs, reasoning AI, and agentic models, has fundamentally changed computing. The Inference Inflection and AI Factories The "inference inflection" has arrived, as AI is now performing productive work. Computing demand has increased exponentially, with estimates suggesting a million-fold increase in the last two years. NVIDIA's AI infrastructure is positioned as the lowest-cost option for AI, offering confidence in long-term utilization and cost-effectiveness. The demand for NVIDIA GPUs is "off the charts," leading to a projected $1 trillion in demand through 2027. Grace Blackwell and Vera Rubin Systems NVIDIA has re-architected systems for inference, exemplified by Grace Blackwell and the upcoming Vera Rubin. These systems feature NVLink 72, FP4 precision for training and inference, and optimized algorithms like Dynamo and TensorRT LLM. Performance gains are significant, with Grace Blackwell showing up to 35x (or even 50x) better perf/watt compared to Hopper. Token cost is a critical metric, and NVIDIA claims the lowest cost per token through extreme co-design. Vera Rubin and Grok Integration The Vera Rubin platform is designed for agentic AI, featuring a new CPU optimized for single-threaded performance and data processing. It incorporates hot-water liquid cooling and a novel 6th generation NVLink switching system. The Grok system, a deterministic data flow processor optimized for inference, is being integrated with Vera Rubin. This integration, facilitated by Dynamo software, aims to combine high throughput (Vera Rubin) with low latency (Grok) for enhanced token generation performance. NVIDIA's Product Roadmap and AI Factory Design NVIDIA's roadmap includes Blackwell, Rubin, Rubin Ultra, and Feynman generations, with continuous architectural advancements. The company is evolving from a chip company to an AI infrastructure and AI factory company. The NVIDIA DSX platform, using Omniverse, is introduced for designing and operating AI factories virtually, optimizing for throughput, resilience, and energy efficiency. OpenClaw and the Agent Revolution OpenClaw is presented as a revolutionary open-source framework for agentic computing, comparable in significance to Linux and HTML. It enables the creation of AI agents that can perceive, reason, and act, managing resources, accessing tools, and interacting with LLMs. Enterprise readiness is addressed with "NemoClaw" (NVIDIA OpenClaw reference), incorporating security and privacy guardrails. The concept of "agentic as a service" is introduced, transforming SaaS companies into "gas" companies. Open Models and Sovereign AI NVIDIA offers a diverse ecosystem of open-source AI models (NemoTron, Cosmos, AlphaMaya, Groot, BioNemo) for specialized domains. These models aim to enable customization for domain-specific and sovereign AI needs. A NemoTron coalition is announced, bringing together companies to advance AI model development. Physical AI, Robotics, and Autonomous Vehicles NVIDIA is also driving advancements in physical AI and robotics. The company provides training computers, simulation platforms (Isaac Lab, Cosmos World Models), and foundation models (Groot) for robots. Partnerships for robo-taxi readiness are highlighted, including BYD, Hyundai, Nissan, and Uber. The AlphaMaio platform enables autonomous vehicles to reason, explain their actions, and follow instructions.