🚀 E-book
Learn how to master the modern AI infrastructural challenges.
  • Press Release by Clarifai

    Artificial Analysis Ranks Clarifai’s gpt-oss-120b Model at Top for Performance and Cost-Efficiency

Washington, D.C. – September 11, 2025 – Artificial Analysis, a leading third-party benchmarking platform today released its API Provider Benchmarking & Analysis report, finding Clarifai’s, full-stack AI platform delivered record-setting performance for its hosted gpt-oss-120b model in independent evaluations. With an impressive output speed of 313 tokens per second and a Time to First Token (TTFT) latency of just 0.27 seconds, the groundbreaking results underscore Clarifai as one of the top hardware- and GPU-agnostic engines for AI workloads where speed, flexibility, efficiency and reliability matter most. 

Data from the third-party benchmarking firm, Artificial Analysis, shows Clarifai as the best software-first vendor in terms of output speed versus time-to-first-token for gpt-oss-120B

 

These results highlight Clarifai’s decade-long track record of technical leadership in AI infrastructure. Paired with its industry-leading compute orchestration technology, these capabilities make Clarifai the prime choice for customers and cloud partners that need best-in-class AI performance without being locked into specific hardware vendors.

 

Clarifai is one of the top vendors among the GPU providers to offer gpt-oss-120b with unbeatable speed and  efficiency. The Clarifai-hosted model shows impressive performance across:

  • High Throughput: The gpt-oss-120b delivers a median output speed that surpasses that of every hyperscaler. 
  • Ultra-Low Latency: It boasts an ultra-low Time to First Token (TTFT) of 0.26 seconds, crucial for real-time and responsive AI applications.
  • Unrivaled Cost-Efficiency: With a blended price of just $0.16 per million tokens, it stands as Clarifai's most cost-efficient option.

 

Data from the third-party benchmarking firm, Artificial Analysis, shows that Clarifai stands out as the most cost-efficient option among software-first vendors with a blended price of just $0.16 per million tokens.

 

These metrics are a testament to Clarifai’s decade-plus of experience in serving production AI workloads for customers of its API. With regular uptimes of 99.99%, the optimized stack Clarifai provides is designed to deliver a high-speed, low-latency end-to-end experience in any compute environment from cloud to on-premise, without sacrificing security or reliability.

 

“Our team has been relentless in optimizing every layer of the stack, from the model architecture to the end-to-end user experience,” said Matthew Zeiler, Founder & Chief Executive Officer at Clarifai. “These independent benchmarks validate what our customers have already experienced—that our platform is engineered to deliver superior speed while providing the flexibility and efficiency required for modern AI workloads.”

 

The flexibility of the Clarifai platform is a key differentiator. The company’s compute orchestration capabilities support a variety of deployment environments, including serverless, dedicated instances, and multi-cloud setups, ensuring customers can deploy and scale models with ease. This is augmented by Local Runners, which allow developers to connect models running on their local machines or private servers directly to Clarifai's platform via a seamless, publicly accessible API.

 

Pricing and availability:

The gpt-oss-120b model is available as a hosted offering on the Clarifai platform, noted for its cost efficiency at $0.16 per million tokens. With Local Runners, you can also deploy powerful models on your own dedicated compute, not just in a hosted offering. Clarifai’s model-agnostic platform, which supports a diverse portfolio of models from various creators, enables you to select the optimal model for each task based on your specific needs without vendor lock-in.

Read the entire benchmarking report here.

 

About Clarifai

Clarifai is a global leader in AI and the pioneer of the full-stack AI platform that helps organizations, teams, and developers build, deploy, and operationalize AI at scale. Clarifai's cutting-edge AI platform supports today's modern AI technologies like Large Language Models (LLMs), Large Vision Models (LVMs), Retrieval Augmented Generation (RAG), automated data labeling, high-volume production inference, and more. Founded in 2013, Clarifai is available in cloud, on-premises, or hybrid environments and has been used to build more than 1.5 million AI models with more than 500,000 users in 170 countries. Learn more at www.clarifai.com. 

For more information or media requests, contact: pr@clarifai.com.