Penguin Solutions announced the launch of its OriginAI Factory Platform and the MemoryAI KV Cache Server, and it revealed a partnership with Deepgram and Dell to deliver a tailored infrastructure for Deepgram’s enterprise voice‑AI services. The product announcements were made on March 16, 2026, and the partnership was disclosed on March 17, 2026.
The OriginAI Factory Platform is designed to accelerate AI inference by providing pre‑defined GPU architectures and integrated software. It bundles NVIDIA DGX B300 platforms based on the Blackwell generation, enabling customers to deploy high‑performance AI workloads with reduced complexity and faster time‑to‑market. By standardizing the hardware‑software stack, Penguin Solutions aims to capture a growing demand for turnkey AI infrastructure in data‑center and edge environments.
The MemoryAI KV Cache Server is the first production‑ready CXL‑based key‑value cache that delivers up to 11 TB of memory. It combines 3 TB of DDR5 main memory with up to eight 1 TB CXL add‑in cards, increasing GPU memory bandwidth and reducing latency for enterprise‑scale inference. Phil Pokorny, Chief Technology Officer, said, “CXL‑enabled KV cache technology delivers faster time‑to‑first‑token, reduced time per output token, and increased overall end‑to‑end token throughput.” This capability addresses the “memory wall” that limits many AI models, positioning Penguin as a leader in high‑capacity, low‑latency inference.
The partnership with Deepgram and Dell leverages Penguin’s OriginAI platform and Dell’s AI‑optimized PowerEdge XE7745 servers, NVIDIA RTX PRO 6000 Blackwell Server Edition GPUs, and Dell PowerScale storage. Joe Castillo, Vice President of Sales, noted, “By partnering with Deepgram and utilizing proven Dell AI infrastructure, Penguin Solutions is delivering a validated, scalable, end‑to‑end architecture.” The collaboration is intended to meet Deepgram’s low‑latency, high‑throughput requirements for enterprise voice‑AI applications, a market that is expanding as businesses adopt real‑time speech‑to‑text and conversational AI solutions.
Collectively, the product launches and partnership signal Penguin’s strategy to deepen its presence in the AI infrastructure market. The OriginAI Factory Platform targets customers seeking rapid deployment of inference workloads, while the MemoryAI KV Cache Server addresses the critical memory bottleneck that limits model size and speed. The Deepgram–Dell alliance expands Penguin’s reach into the voice‑AI niche, a segment that is projected to grow as enterprises seek real‑time, conversational interfaces. By combining its hardware expertise with Dell’s scale and Deepgram’s software, Penguin positions itself as a one‑stop solution for high‑performance, low‑latency AI services.
Abe Pursell, Vice President of Partnerships and Business Development at Deepgram, said, “The infrastructure behind our platform has to be equally robust to support that level of innovation. Penguin Solutions demonstrated a deep understanding of our technical requirements, translating them into a sophisticated infrastructure environment that meets and exceeds expectations.”
The content on EveryTicker is for informational purposes only and should not be construed as financial or investment advice. We are not financial advisors. Consult with a qualified professional before making any investment decisions. Any actions you take based on information from this site are solely at your own risk.