Groq and HUMAIN Launch OpenAI Models with Real-Time Access on GroqCloud

HUMAIN, an AI services firm backed by Saudi Arabia’s Public Investment Fund, and Groq, a leader in ultra-fast inference technology, have jointly announced the deployment of two cutting-edge OpenAI open-source models—gpt-oss-120B and gpt-oss-20B—on the GroqCloud platform. These models, now available globally, come with full 128,000-token context support, real-time response capabilities, and integrated backend tools, all running on Groq’s optimized inference infrastructure from day one. This partnership is aimed at broadening access to advanced AI capabilities in Saudi Arabia and beyond, empowering developers with immediate production-level tools.

Groq has a strong history of enabling OpenAI’s open-source projects at scale, including earlier work with the Whisper model. This latest integration furthers that commitment by offering top-tier models ready for real-world use with robust performance and global scalability. Local support provided by HUMAIN ensures that developers within the Kingdom can utilize these tools effectively and efficiently, fueling innovation and growth in the regional AI landscape.

Groq CEO Jonathan Ross emphasized the company’s mission to accelerate access to state-of-the-art open-source models, noting that Groq’s infrastructure is specifically designed to make them both fast and affordable from the moment they launch. The collaboration with HUMAIN is seen as a strategic step to extend these benefits to developers in Saudi Arabia, giving them tools to work more intelligently and efficiently from the start.

Tareq Amin, CEO of HUMAIN, highlighted the performance and cost advantages Groq delivers, making it possible to deploy high-performance AI at scale within the Kingdom. Supported by Groq’s industry-leading compute platform and OpenAI’s sophisticated models, this endeavor is viewed as a catalyst for a new generation of AI-driven projects in the area.

The GroqCloud environment is designed to fully unlock the capabilities of OpenAI’s new models, with built-in tools such as web search and code execution to enhance contextual accuracy and enable complex automation. The infrastructure allows developers to run gpt-oss-120B at over 500 tokens per second and gpt-oss-20B at over 1,000 tokens per second, offering exceptional performance without compromising affordability. Groq provides these models at industry-leading prices, setting a new standard in open-source AI deployment.