InferenceX™ In the Media

Coverage and mentions of InferenceX™ (formerly InferenceMAX) across industry publications, blogs, and media outlets.

March 4, 2026
GPU Mode By Mark Saroufim, Meta Pytorch EngineerVideo

Lecture 100: InferenceX Continuous OSS Inference Benchmarking

February 26, 2026
Chip BriefingArticle

Weekly: Micron Leans Into Memory Supercycle

February 24, 2026
SambaNovaArticle

Introducing the SN50 RDU: Purpose-Built for Agentic Inference

February 20, 2026
LMSys SGLangArticle

Unlocking 25x Inference Performance with SGLang on NVIDIA GB300 NVL72

February 20, 2026
arXivArticle

RPU -- A Reasoning Processing Unit

February 19, 2026
SDxCentralArticle

Nvidia pulls ahead as AMD’s software stack falls short: report

February 17, 2026
AMD Developer BlogArticle

Speed is the Moat: Inference Performance on AMD GPUs

February 16, 2026
LMSys SGLangArticle

SGLang is also officially crowned "InferenceMax King" by SemiAnalysis

February 16, 2026
SemiAnalysisArticle

InferenceX v2: NVIDIA Blackwell Vs AMD vs Hopper

February 6, 2026
Google Cloud BlogArticle

Scaling MoE Inference with NVIDIA Dynamo on Google Cloud

February 5, 2026
BasetenArticle

How to run LLM performance benchmarks (and why you should)

January 25, 2026
arXivArticle

Kareus: Joint Reduction of Dynamic and Static Energy in Large Model Training

January 5, 2026
NVIDIAVideo

CES Analyst Q&A [timestamp - 3:37]

November 19, 2025
NVIDIAArticle

InferenceMAX Mention During NVIDIA Q3 FY2026 Report

November 13, 2025
Red Hat - Co-Maintainers of vLLMVideo

[vLLM Office Hours #37] InferenceMAX & vLLM

November 11, 2025
AMD SVP of AI, Vamsi BoppanaVideo

AMD Instinct MI350™: Generational Efficiency gains - Up to 10x on InferenceMAX

October 28, 2025
NVIDIAVideo

NVIDIA GTC DC 2025 - Jensen Huang talks about InferenceMAX

October 24, 2025
Microsoft Azure Engineering BlogArticle

Scaling Multi-Node LLM Inference with NVIDIA Dynamo and ND GB200 NVL72 GPUs on AKS

October 14, 2025
SGLang LMSYS OrgArticle

SGLang and NVIDIA Accelerating SemiAnalysis InferenceMAX and GB200 Together

October 13, 2025
NVIDIA Developer BlogArticle

NVIDIA Blackwell Leads on SemiAnalysis InferenceMAX v1 Benchmarks

October 10, 2025
TensorWaveArticle

Benchmark Breakdown: How AMD’s MI300X, MI325X, and MI355X Are Redefining AI Inference Economics

October 9, 2025
SemiAnalysisArticle

InferenceMAX™: Open Source Inference Benchmarking

October 9, 2025
Barron’sArticle

Nvidia, AMD Chips Compared in New Benchmarks. See Who Comes Out on Top.

October 9, 2025
vLLM BlogArticle

SemiAnalysis InferenceMAX: vLLM and NVIDIA Accelerate Blackwell Inference

October 9, 2025
AMD Developer BlogArticle

InferenceMAX: Benchmarking Progress in Real Time

October 9, 2025
Open Compute ProjectVideo

SemiAnalysis InferenceMAX Benchmarking the AI Frontier

October 9, 2025
NVIDIAVideo

InferenceMAX Instagram Explainer