Zyphra today announced Zyphra Cloud, a full-stack AI platform that brings advanced innovations from Zyphra Research into production for developers, enterprises, and frontier AI hyperscalers. Powered by AMD Instinct™ MI355X GPUs on TensorWave's purpose-built infrastructure, Zyphra Cloud unifies model serving, agent infrastructure, and scalable compute into a single platform for building and deploying advanced AI systems.
Zyphra Cloud launches with Zyphra Inference, a serverless inference service providing access to frontier open-weight models including DeepSeek V3.2, Kimi K2.6, and GLM 5.1. Zyphra Inference combines custom kernels, novel long-context inference algorithms, and advanced parallelism schemes to deliver high-throughput, low-latency performance for production-grade long-horizon use cases such as agentic coding, deep research, and long-horizon workflow automation.
Login to comment