Kinara Ara-2

Shaping the future of Generative AI at the Edge

Meet the Kinara Ara-2 AI processor, the leader in Edge AI acceleration. This 40 TOPS powerhouse tackles the massive compute demands of Generative AI and transformer-based models with unmatched cost-effectiveness.

Built on the same flexible and efficient dataflow architecture as its predecessor, the Ara-1, the Ara-2 unleashes a beastly performance boost. Think tremendous increases in performance/Watt and performance/$ – running at less than 2W in a typical vision application, this chip delivers unmatched efficiency and power for your AI needs.

Run large language models (LLMs) at the edge to enable real-time productivity, lower usage cost, and ensure data privacy.

Fuel powerful transformer-based models, driving breakthroughs in natural language processing and machine translation.

Revolutionize edge applications from smart cities to healthcare diagnostics, all while minimizing cost and maximizing efficiency.

Ara-2 Processor
Ara-2 Processor
  • With its eight Gen 2 neural cores and ability to access up to 16GB LPDDR4(X) memory, the 40 TOPS Kinara Ara-2 powers edge servers and laptops with high performance, cost effective, and energy efficient inference. For additional operational confidence, Ara-2 supports secure boot and encrypted memory access.
KU-2 USB, KM-2 M.2
KU-2 USB, KM-2 M.2
  • The Kinara Ara-2 accelerator modules enable high performance, power-efficient AI inference for Edge AI applications including Generative AI workloads such as Stable Diffusion and LLaMA-2.
  • Powered by the 40 TOPS Kinara Ara-2, the KU-2 and KM-2 are ideal for running inference on models ranging from CNNs to complex vision transformers for a variety of applications in smart cities, retail, and gaming.
  • The KU-2 and KM-2 are available with flexible memory options – 2GB memory for conventional AI and 8GB for GenAI. Both of these modules can run multi-stream, multiple AI models simultaneously.
KP-2 PCIe Card
KP-2 PCIe Card
  • The Kinara Ara-2 PCIe AI accelerator card enables high performance, power-efficient AI inference for edge server applications including Generative AI workloads such as Stable Diffusion and LLMs.
  • The KP-2 is powered by four 40 TOPS Ara-2s, rivaling the latest GPU-based inference cards in performance but at the fraction of power and cost.
  • The KP-2 is ideal for running heavy inference workloads with multiple modern CNNs, transformers, and Generative AI models in a variety of smart city, retail, gaming, and enterprise applications.