MLCommons
banner
mlcommons.org
MLCommons
@mlcommons.org
MLCommons is an AI engineering consortium, built on a philosophy of open collaboration to improve AI systems. Through our collective engineering efforts, we continually measure and improve AI technologies' accuracy, safety, speed, and efficiency.
MLPerf Training v5.1 results are live!
Record participation: 20 organizations submitted 65 unique systems featuring 12 different accelerators. Multi-node submissions increased 86% over last year, showing the industry's focus on scale.
Results: mlcommons.org/2025/11/trai...
#MLPerf
1/3
November 12, 2025 at 4:06 PM
Don’t miss #MLCommons Endpoints in San Diego, Dec 1–2!
Learn, connect, and shape the future of AI with top experts at Qualcomm Hall.
🗓 Dec 1–2 | 🎟 Free tickets available now!

www.eventbrite.com/e/mlcommons-...

#AI #MachineLearning #SanDiego
November 10, 2025 at 10:17 PM
🚨 NEW: We tested 39 AI models for security vulnerabilities.

Not a single one was as secure as it was "safe."

Today, we're releasing the industry's first standardized jailbreak benchmark. Here's what we found 🧵1/6

mlcommons.org/2025/10/ailu...
October 15, 2025 at 7:33 PM
MLPerf Inference v5.1 results are live!
Record 27 organizations submitted 1,472 performance results across new and established AI workloads.
Three new benchmarks debut:

Reasoning with Deepseek R1
Speech to text with Whisper
Small LLM with Llama 3.1 8B

Read More: mlcommons.org/2025/09/mlpe...
September 9, 2025 at 6:15 PM
MLCommons just launched MLPerf Mobile on the Google Play Store! 📱
Benchmark your Android device’s AI performance on real-world ML tasks with this free, open-source app.
Try it now: play.google.com/store/apps/d...
July 10, 2025 at 7:01 PM
Call for Submissions!

#MLCommons & @AVCConsortium are accepting submissions for the #MLPerf Automotive Benchmark Suite! Help drive fair comparisons & optimize AI systems in vehicles. Focus is on camera sensor perception.

📅 Submissions close June 13th, 2025

Join: mlcommons.org/community/su...
June 5, 2025 at 6:12 PM
MLCommons is partnering with Nasscom to develop globally recognized AI reliability benchmarks, including India-specific, Hindi-language evaluations. Together, we are advancing trustworthy AI.
🔗 mlcommons.org/2025/05/nass...

#AIForAll #IndiaAI #ResponsibleAI #Nasscom #MLCommons
May 29, 2025 at 3:07 PM
1/ MLCommons announces the release of MLPerf Client v0.6, the first open benchmark to support NPU and GPU acceleration on consumer AI PCs.
Read more: mlcommons.org/2025/04/mlpe...
April 28, 2025 at 3:12 PM
MLCommons introduced a new Automotive PointPainting benchmark in its recent MLPerf Inference v5.0 release. The benchmark is based on established industry methods such as PointPainting, DeepLab3+, and the Waymo Open Dataset. For more details, visit mlcommons.org/2025/03/auto....
April 7, 2025 at 9:32 PM
1/ MLPerf Inference benchmark v5.0 includes a new GNN benchmark providing a standardized way to assess the performance of GNN implementations and hardware accelerators.
April 4, 2025 at 8:57 AM
1/ We are excited to announce the latest MLCommons MLPerf Inference v5.0 benchmark suite results. This round featured robust participation from 23 submitting organizations delivering over 17,000 performance results!
mlcommons.org/2025/04/mlperf-inference-v5-0-results/
April 2, 2025 at 4:47 PM
MedPerf integrates smart contracts & private data objects to improve accountability & integrity.
#MedPerf is an open framework for benchmarking medical AI using real-world private datasets to ensure transparency and privacy.
mlcommons.org/2025/03/medp...

#MedicalAI #smartcontracts
March 10, 2025 at 9:11 PM
Submissions are open for the MLPerf Training v5.0 Benchmark.

We are excited to add a new pretraining benchmark, llama3.1 405B, to showcase the latest innovations in AI.

To participate, join the Working Group
mlcommons.org/working-grou...
March 7, 2025 at 9:18 PM
Croissant, a metadata format for ML-ready datasets, has gained traction since its launch in March 2024. Major platforms like Kaggle and HuggingFace now support it.

The MLCommons Croissant working group co-chairs shared insights on its rapid adoption and future plans.
mlcommons.org/2025/02/croi...
February 12, 2025 at 4:33 PM
MLCommons, in partnership with the AI Verify Foundation, released the AILuminate v1.1, incorporating new French language capabilities into its first-of-its-kind AI safety benchmark.

Learn more: mlcommons.org/2025/02/ailu...

#ailuminate #parisaiactionsummit #aiverifyfoundation
February 11, 2025 at 4:05 PM
@mlcommons.org releases Unsupervised People's Speech dataset:
1M+ hours of multilingual audio
821K+ hours of detected speech
89 languages
48+ TB of data

Empowering research in:
✅ Speech recognition
✅ Language ID
✅ Global communication tech

Learn more: mlcommons.org/2025/01/new-...

#nlp #datasets
January 30, 2025 at 8:40 PM
Announcing the release of AILuminate, a first-of-its kind benchmark to measure the safety of LLMs. The AILuminate v1.0 benchmark offers a comprehensive set of safety grades for today's most prevalent #LLMs.
mlcommons.org/2024/12/mlco...
(1/4)
December 4, 2024 at 3:29 PM