InferenceX™ In the Media
Coverage and mentions of InferenceX™ (formerly InferenceMAX) across industry publications, blogs, and media outlets.
March 20, 2026·EE Times·Article
March 20, 2026
EE TimesArticle
“GTC 2026 Keynote: Long Live the Inference King”
March 20, 2026·Sina Tech·Article
March 20, 2026
Sina TechArticle
“英伟达热点小时报: 数据中心就是Token工厂 — Jensen Huang GTC 2026”
March 19, 2026·AMD·Video
March 19, 2026
AMDVideo
“AMD Head of AI Product, Ramine Roane, shows his Inference King Champion Belt”
March 18, 2026·Morgan Stanley·Video
March 18, 2026
Morgan StanleyVideo
“Morgan Stanley TMT Conference 2026 (timestamp: 21:41)”
March 18, 2026·Spheron Network·Article
March 18, 2026
Spheron NetworkArticle
“NVIDIA H200 vs B200 vs GB200: Which GPU to Rent for AI in 2026?”
March 18, 2026·GMI Cloud (Juejin)·Article
March 18, 2026
GMI Cloud (Juejin)Article
“GMI Cloud 成为英伟达 Dynamo 1.0 及 OpenShell 首发合作伙伴”
March 17, 2026·PYMNTS·Article
March 17, 2026
PYMNTSArticle
“Nvidia’s Jensen Huang Says AI Compute Could Near $1 Trillion by 2027”
March 17, 2026·Sebastian Barros·Article
March 17, 2026
Sebastian BarrosArticle
“Telcos are the best channel to Democratize AI”
March 16, 2026·NVIDIA·Video
March 16, 2026
NVIDIAVideo
“GTC 2026: Jensen announces that he is one of the Inference King”
March 16, 2026·NVIDIA Developer Blog·Article
March 16, 2026
NVIDIA Developer BlogArticle
“How NVIDIA Dynamo 1.0 Powers Multi-Node Inference at Production Scale”
March 10, 2026·TensorWave·Article
March 10, 2026
TensorWaveArticle
“MI355X Just Flipped the Script on B200 for FP8 DeepSeek Disagg”
March 9, 2026·Forbes·Article
March 9, 2026
ForbesArticle
“Meta’s Most Efficient AI Infrastructure for Digital Intelligence”
March 7, 2026·The Register·Article
March 7, 2026
The RegisterArticle
“Unpacking the deceptively simple science of tokenomics”
March 5, 2026·SAIL Media·Video
March 5, 2026
SAIL MediaVideo
“The Future of AI Infrastructure: Why Inference Max Matters”
March 4, 2026·GPU Mode By Mark Saroufim, Meta Pytorch Engineer·Video
March 4, 2026
GPU Mode By Mark Saroufim, Meta Pytorch EngineerVideo
“Lecture 100: InferenceX Continuous OSS Inference Benchmarking”
February 26, 2026·Chip Briefing·Article
February 26, 2026
Chip BriefingArticle
“Weekly: Micron Leans Into Memory Supercycle”
February 24, 2026·SambaNova·Article
February 24, 2026
SambaNovaArticle
“Introducing the SN50 RDU: Purpose-Built for Agentic Inference”
February 19, 2026·SDxCentral·Article
February 19, 2026
SDxCentralArticle
“Nvidia pulls ahead as AMD’s software stack falls short: report”
February 17, 2026·AMD Developer Blog·Article
February 17, 2026
AMD Developer BlogArticle
“Speed is the Moat: Inference Performance on AMD GPUs”
February 16, 2026·LMSys SGLang·Article
February 16, 2026
LMSys SGLangArticle
“SGLang is also officially crowned "InferenceMax King" by SemiAnalysis”
February 16, 2026·NVIDIA Blog·Article
February 16, 2026
NVIDIA BlogArticle
“New SemiAnalysis InferenceX Data Shows NVIDIA Blackwell Ultra Delivers up to 50x Better Performance and 35x Lower Costs for Agentic AI”
February 16, 2026·SemiAnalysis·Article
February 16, 2026
SemiAnalysisArticle
“InferenceX v2: NVIDIA Blackwell Vs AMD vs Hopper”
February 16, 2026·WCCFTech·Article
February 16, 2026
WCCFTechArticle
“NVIDIA’s Blackwell Ultra Pushes "Agentic AI" Performance to New Heights, Delivering Up to 50× Higher Tokens/Watt & Stronger Long-Context Workloads”
February 6, 2026·Google Cloud Blog·Article
February 6, 2026
Google Cloud BlogArticle
“Scaling MoE Inference with NVIDIA Dynamo on Google Cloud”
February 5, 2026·Baseten·Article
February 5, 2026
BasetenArticle
“How to run LLM performance benchmarks (and why you should)”
January 25, 2026·arXiv·Article
January 25, 2026
arXivArticle
“Kareus: Joint Reduction of Dynamic and Static Energy in Large Model Training”
November 19, 2025·NVIDIA·Article
November 19, 2025
NVIDIAArticle
“InferenceMAX Mention During NVIDIA Q3 FY2026 Report”
November 13, 2025·Red Hat - Co-Maintainers of vLLM·Video
November 13, 2025
Red Hat - Co-Maintainers of vLLMVideo
“[vLLM Office Hours #37] InferenceMAX & vLLM”
November 11, 2025·AMD SVP of AI, Vamsi Boppana·Video
November 11, 2025
AMD SVP of AI, Vamsi BoppanaVideo
“AMD Instinct MI350™: Generational Efficiency gains - Up to 10x on InferenceMAX”
October 28, 2025·NVIDIA·Video
October 28, 2025
NVIDIAVideo
“NVIDIA GTC DC 2025 - Jensen Huang talks about InferenceMAX”
October 24, 2025·Microsoft Azure Engineering Blog·Article
October 24, 2025
Microsoft Azure Engineering BlogArticle
“Scaling Multi-Node LLM Inference with NVIDIA Dynamo and ND GB200 NVL72 GPUs on AKS”
October 16, 2025·Crusoe·Article
October 16, 2025
CrusoeArticle
“The New AI Benchmark: Unlocking Real-World Performance with InferenceMAX by SemiAnalysis”
October 14, 2025·SGLang LMSYS Org·Article
October 14, 2025
SGLang LMSYS OrgArticle
“SGLang and NVIDIA Accelerating SemiAnalysis InferenceMAX and GB200 Together”
October 13, 2025·NVIDIA Developer Blog·Article
October 13, 2025
NVIDIA Developer BlogArticle
“NVIDIA Blackwell Leads on SemiAnalysis InferenceMAX v1 Benchmarks”
October 10, 2025·TensorWave·Article
October 10, 2025
TensorWaveArticle
“Benchmark Breakdown: How AMD’s MI300X, MI325X, and MI355X Are Redefining AI Inference Economics”
October 10, 2025·Tom's Hardware·Article
October 10, 2025
Tom's HardwareArticle
“InferenceMax AI Benchmark Tests Software Stacks, Efficiency, and TCO — Vendor-Neutral Suite Runs Nightly and Tracks Performance Changes Over Time”
October 9, 2025·SemiAnalysis·Article
October 9, 2025
SemiAnalysisArticle
“InferenceMAX™: Open Source Inference Benchmarking”
October 9, 2025·Barron’s·Article
October 9, 2025
Barron’sArticle
“Nvidia, AMD Chips Compared in New Benchmarks. See Who Comes Out on Top.”
October 9, 2025·NVIDIA Blog·Article
October 9, 2025
NVIDIA BlogArticle
“NVIDIA Blackwell Raises Bar in New InferenceMAX Benchmarks, Delivering Unmatched Performance and Efficiency”
October 9, 2025·vLLM Blog·Article
October 9, 2025
vLLM BlogArticle
“SemiAnalysis InferenceMAX: vLLM and NVIDIA Accelerate Blackwell Inference”
October 9, 2025·AMD Developer Blog·Article
October 9, 2025
AMD Developer BlogArticle
“InferenceMAX: Benchmarking Progress in Real Time”
October 9, 2025·Open Compute Project·Video
October 9, 2025
Open Compute ProjectVideo
