
AMD Empowers Zyphra to Achieve Breakthrough in Frontier-Scale AI Training with ZAYA1 Foundation Model
AMD announced a major milestone in the field of artificial intelligence development and large-scale model training, as Zyphra successfully completed the training of ZAYA1, the first large-scale Mixture-of-Experts (MoE) foundation model to be trained using an end-to-end AMD GPU and networking platform. ZAYA1 represents an important step forward for production-ready AI infrastructure, demonstrating that advanced model development can be achieved using AMD Instinct™ MI300X GPUs and AMD Pensando™ networking technology, all enabled through the AMD ROCm™ open software stack. Zyphra has published a detailed technical report today outlining the performance results and architectural advantages of the new system.
According to Zyphra’s results, the ZAYA1 model delivers competitive—or in many cases superior—performance compared to leading open-source models across tasks including reasoning, mathematics, and coding. These benchmarks illustrate the real-world capability, scalability, and efficiency of AMD’s Instinct GPU platform for high-intensity, production-scale AI workloads that traditionally required extremely large system footprints and complex compute orchestration.
A Significant Milestone for Advanced AI Development
The completion of ZAYA1 marks a transformative accomplishment in the AI ecosystem, particularly for organizations pursuing frontier-scale intelligence without relying solely on proprietary or closed-platform hardware environments. ZAYA1 demonstrates that AMD’s ecosystem—spanning silicon, networking, software and compute orchestration—is capable of powering complex modular architectures like MoE at scale.
In recent years, Mixture-of-Experts models have gained significant traction because they enable large models to scale efficiently by activating only a subset of parameters during inference, reducing compute cost without sacrificing performance. However, training MoE architectures at full scale traditionally requires overcoming major challenges related to memory efficiency, multi-GPU parallelism, distributed I/O performance and network reliability. Through this collaboration, Zyphra proved that AMD’s platform can successfully support such workloads with high efficiency and competitive price-performance ratios.
Reflecting on the accomplishment, Emad Barsoum, Corporate Vice President of AI and Engineering within AMD’s Artificial Intelligence Group, emphasized the importance of this achievement for the broader ecosystem:
AMD leadership in accelerated computing is empowering innovators like Zyphra to push the boundaries of what’s possible in AI. This milestone showcases the power and flexibility of AMD Instinct GPUs and Pensando networking for training complex, large-scale models.
This collaboration signals AMD’s strengthening position as a viable alternative platform for large-scale AI and computationally intensive environments—especially as industries seek greater flexibility, supply diversity and open-source aligned development.
Zyphra’s Vision for Efficient and Scalable Intelligence
At the core of Zyphra’s mission is a commitment to efficiency—not simply in energy or cost, but in model architecture design, training methodology and customer-focused deployment strategies. Zyphra’s CEO, Krithik Puthalath, highlighted how efficiency drives every aspect of the company’s engineering approach, shaping both internal innovation and external partnerships:
Efficiency has always been a core guiding principle at Zyphra. It shapes how we design model architectures, develop algorithms for training and inference, and choose the hardware with the best price-performance to deliver frontier intelligence to our customers.
He continued by acknowledging the strategic importance of the collaboration:
ZAYA1 reflects this philosophy and we are thrilled to be the first company to demonstrate large-scale training on an AMD platform. Our results highlight the power of co-designing model architectures with silicon and systems, and we’re excited to deepen our collaboration with AMD and IBM as we build the next generation of advanced multimodal foundation models.
Zyphra’s approach exemplifies a growing trend in the AI industry: system-level design where model architectures and hardware evolve together, rather than in isolation. As compute requirements accelerate, co-optimization becomes critical for both performance and sustainability.
Efficient Training at Scale, Powered by AMD Instinct GPUs
A key enabler of ZAYA1’s training success was the AMD Instinct MI300X GPU, which provides 192 GB of high-bandwidth memory (HBM). This capacity played a crucial role by eliminating the need for expert or tensor sharding—two complex and resource-intensive approaches commonly required to distribute model components across GPU clusters with lower memory capacity. By avoiding these techniques, Zyphra dramatically reduced system complexity and improved throughput across layers of the training stack.
Zyphra also reported more than a 10x improvement in model save times through AMD-optimized distributed I/O pipelines, allowing training checkpoints to be captured faster and more reliably. Faster checkpointing not only increases productivity but enhances resilience to runtime interruptions in large-scale environments.
Performance benchmarks reinforce the practical success of the AMD approach: ZAYA1-Base, with 8.3 billion total parameters and 760 million active parameters, matches or surpasses the performance of leading open models including:
- Qwen3-4B (Alibaba)
- Gemma3-12B (Google)
- Llama-3-8B (Meta)
- OLMoE
The ability to achieve results comparable to or better than substantially larger models highlights the advantages of MoE optimization and high-bandwidth memory utilization.
Collaborative Engineering to Enable Next-Generation Compute
Zyphra’s achievement builds upon close collaboration with both AMD and IBM. Together, the organizations architected and deployed a large-scale training cluster powered by AMD Instinct GPUs and connected through AMD Pensando networking interconnect, providing high-speed, low-latency system communication crucial for MoE efficiency.
Earlier this quarter, AMD and IBM announced the jointly engineered environment used for training ZAYA1, which combines AMD Instinct MI300X GPUs with IBM Cloud’s high-performance networking fabric and distributed storage infrastructure. This deployment provided the foundational compute footprint necessary for ZAYA1’s large-scale pretraining.
The project illustrates a powerful example of multi-vendor ecosystem alignment innovating toward shared goals: scalable, open, efficient AI development.
Redefining the Future of Frontier AI Computing
The successful training of ZAYA1 is more than a technical achievement—it signals broader change across the AI landscape. As demand for frontier-level intelligence accelerates, organizations are increasingly evaluating new pathways to scale compute without relying exclusively on closed or constrained hardware supply channels. Zyphra’s results demonstrate that AMD’s GPU and networking technology is capable of enabling strong performance competitive with industry leaders, providing more choice in the global compute market.
At the same time, the success highlights the growing importance of:
- Open software ecosystems such as AMD ROCm
- High-bandwidth memory advances
- Efficient distributed compute strategies
- Collaborative hardware-software co-design
As Zyphra continues developing its next generation of advanced multimodal foundation models, the AMD and IBM ecosystem will play an important role in powering the next phase of its innovation roadmap.
Source Link:https://www.amd.com/en/



