Monday, April 29, 2024

Edgecortix Announces Sakura AI Co-processor Delivering Industry Leading Low-Latency and Energy-Efficiency

EdgeCortix Inc., the innovative fabless semiconductor design company with a software first approach, focused on delivering class-leading compute efficiency and latency for edge artificial intelligence (AI) inference; unveiled the architecture, performance metrics and delivery timing for its new industry leading, energy-efficient, AI inference co-processor.

Also Read: Edgecortix Announces Sakura AI Co-processor Delivering Industry Leading Low-Latency and Energy-Efficiency

Today, at the prestigious TechInsights’ Linley Spring Processor Conference, EdgeCortix officially introduced its flagship energy-efficient AI co-processor (accelerator), branded SAKURA™.  Designed entirely in their Tokyo based development center, the company announced that SAKURA is produced by TSMC on the popular 12nm FinFET technology and will be available as low-power PCI-E based development boards to participating companies of the EdgeCortix Early Access Program (EAP) in July of 2022.

“SAKURA is revolutionary from both a technical and competitive perspective, delivering well over 10X performance/watt advantage compared to current AI inference solutions based on traditional graphics processing units (GPUs), especially for real-time edge applications.”, said Sakyasingha Dasgupta, CEO and Founder of EdgeCortix, “After validating our AI processor architecture design with multiple field-programmable gate array (FPGA) customers in production, we designed SAKURA as a co-processor that can be plugged in alongside a host processor in nearly all existing systems to significantly accelerate AI inference. Using our patented runtime-reconfigurable interconnect technology, SAKURA is inherently more flexible than traditional processors and can achieve near optimal compute utilization in contrast to most AI processors that have been developed over the last 40+ years.”

SAKURA is powered by a 40 trillion operations per second (TOPS), single core Dynamic Neural Accelerator® (DNA) Intellectual Property (IP), which is EdgeCortix’s proprietary neural processing engine with built-in reconfigurable data-path connecting all compute engines. DNA enables the new SAKURA AI co-processor to run multiple deep neural network models together, with ultra-low latency, while preserving exceptional TOPS utilization. This unique attribute is key to enhancing the processing speed, energy-efficiency, and longevity of the system-on-chip, providing exceptional total cost of ownership benefits. The DNA IP is specifically optimized for inference with streaming and high-resolution data.

Subscribe Now

    Hot Topics