Ambarella Brings Generative AI Capabilities to Edge Devices; Introduces N1 System-on-Chip Series for On-Premise Applications
- None.
- None.
Insights
The announcement by Ambarella, Inc. regarding its new N1 SoC series represents a significant technological advancement in the semiconductor industry. The ability to support large language models (LLMs) with low power consumption is a critical development, especially given the growing demand for edge computing solutions. The energy efficiency of Ambarella's SoC solutions, being up to three times more power-efficient per generated token compared to GPUs and other AI accelerators, indicates a potential shift in the market dynamics where power consumption is a key differentiator.
This development could lead to a broader adoption of AI capabilities in edge devices, which traditionally have been limited by power constraints. The implications for businesses in sectors like video security, robotics and industrial applications are considerable, as they stand to benefit from enhanced on-device AI without the need for extensive server infrastructure. This could result in cost savings, improved efficiency and the development of new product offerings that leverage on-device generative AI capabilities.
From a financial perspective, Ambarella's strategic move to optimize its SoC offerings for generative AI processing capabilities could lead to an increase in its market share within the AI semiconductor space. The company's focus on mid to high-end SoCs could attract a diverse range of clients looking for efficient and powerful AI solutions at a lower energy cost. The potential for immediate and cost-effective deployment aligns well with industry needs, particularly for companies aiming to reduce operational expenditures associated with data processing and analysis.
In the long term, Ambarella's positioning in the market as a provider of energy-efficient AI solutions could result in sustained revenue growth, especially if the adoption of AI in edge devices continues to accelerate. Investors should monitor the company's market penetration following this announcement and the subsequent adoption rate of the N1 SoC series by manufacturers of edge endpoint devices.
The integration of generative AI into edge endpoint devices is a burgeoning trend that is expected to grow exponentially. Ambarella's demonstration of multi-modal LLMs on its N1 SoC series at CES is not only a showcase of their technical prowess but also a strategic move to capture early market interest. With the edge AI market projected to expand, Ambarella's entry with a product that emphasizes low power consumption could resonate well with businesses looking to deploy intelligent solutions without the added cost of power and infrastructure.
Market research indicates a growing appetite for AI solutions that are both powerful and energy-efficient. Ambarella's advancements could set a new industry standard for SoC performance and efficiency, potentially influencing competitors and driving innovation in the sector. The company's focus on enabling generative AI for a wide range of applications suggests a broad market strategy that could lead to partnerships, collaborations and an expanded clientele.
Single SoC Supports One to 34 Billion-Parameter, Multi-Modal LLMs With Low Power Consumption, Enabling Generative AI for Edge Endpoint Devices
SANTA CLARA, Calif., Jan. 08, 2024 (GLOBE NEWSWIRE) -- Ambarella, Inc. (NASDAQ: AMBA), an edge AI semiconductor company, today announced during CES that it is demonstrating multi-modal large language models (LLMs) running on its new N1 SoC series at a fraction of the power-per-inference of leading GPU solutions. Ambarella aims to bring generative AI—a transformative technology that first appeared in servers due to the large processing power required—to edge endpoint devices and on-premise hardware, across a wide range of applications such as video security analysis, robotics and a multitude of industrial applications.
Ambarella will initially be offering optimized generative AI processing capabilities on its mid to high-end SoCs, from the existing CV72 for on-device performance under 5W, through to the new N1 series for server-grade performance under 50W. Compared to GPUs and other AI accelerators, Ambarella provides complete SoC solutions that are up to 3x more power-efficient per generated token, while enabling immediate and cost-effective deployment in products.
“Generative AI networks are enabling new functions across our target application markets that were just not possible before,” said Les Kohn, CTO and co-founder of Ambarella. “All edge devices are about to get a lot smarter, with our N1 series of SoCs enabling world-class multi-modal LLM processing in a very attractive power/price envelope.”
“Virtually every edge application will get enhanced by generative AI in the next 18 months,” said Alexander Harrowell, Principal Analyst, Advanced Computing at Omdia. “When moving genAI workloads to the edge, the game becomes all about performance per watt and integration with the rest of the edge ecosystem, not just raw throughput.”
All of Ambarella’s AI SoCs are supported by the company’s new Cooper™ Developer Platform. Additionally, in order to reduce customers’ time-to-market, Ambarella has pre-ported and optimized popular LLMs, such as Llama-2, as well as the Large Language and Video Assistant (LLava) model running on N1 for multi-modal vision analysis of up to 32 camera sources. These pre-trained and fine-tuned models will be available for partners to download from the Cooper Model Garden.
For many real-world applications, visual input is a key modality, in addition to language, and Ambarella’s SoC architecture is natively well-suited to process video and AI simultaneously at very low power. Providing a full-function SoC enables the highly efficient processing of multi-modal LLMs while still performing all system functions, unlike a standalone AI accelerator.
Generative AI will be a step function for computer vision processing that brings context and scene understanding to a variety of devices, from security installations and autonomous robots to industrial applications. Examples of the on-device LLM and multi-modal processing enabled by this new Ambarella offering include: smart contextual searches of security footage; robots that can be controlled with natural language commands; and different AI helpers that can perform anything from code generation to text and image generation.
Most of these systems rely heavily on both camera and natural language understanding, and will benefit from on-device generative AI processing for speed and privacy, as well as a lower total cost of ownership. The local processing enabled by Ambarella’s solutions also perfectly suits application-specific LLMs, which are typically fine-tuned on the edge for each individual scenario; versus the classical server approach of using bigger and more power-hungry LLMs to cater to every use case.
Based on Ambarella’s powerful CV3-HD architecture, initially developed for autonomous driving applications, the N1 series of SoCs repurposes all this performance for running multi-modal LLMs in an extremely low power footprint. For example, the N1 SoC runs Llama2-13B with up to 25 output tokens per second in single-streaming mode at under 50W of power. Combined with the ease-of-integration of pre-ported models, this new solution can quickly help OEMs deploy generative AI into any power-sensitive application, from an on-premise AI box to a delivery robot.
Both the N1 SoC and a demonstration of its multi-modal LLM capabilities will be on display this week at the Ambarella exhibition during CES.
About Ambarella
Ambarella’s products are used in a wide variety of human vision and edge AI applications, including video security, advanced driver assistance systems (ADAS), electronic mirror, drive recorder, driver/cabin monitoring, autonomous driving and robotics applications. Ambarella’s low-power systems-on-chip (SoCs) offer high-resolution video compression, advanced image and radar processing, and powerful deep neural network processing to enable intelligent perception, fusion and planning. For more information, please visit www.ambarella.com.
Ambarella Contacts
- Media contact: Eric Lawson, elawson@ambarella.com, +1 480-276-9572
- Investor contact: Louis Gerhardy, lgerhardy@ambarella.com, +1 408-636-2310
- Sales contact: https://www.ambarella.com/contact-us/
All brand names, product names, or trademarks belong to their respective holders. Ambarella reserves the right to alter product and service offerings, specifications, and pricing at any time without notice. © 2024 Ambarella. All rights reserved.
A photo accompanying this announcement is available at https://www.globenewswire.com/NewsRoom/AttachmentNg/10ee7319-3783-423b-b8f6-d3f6d219293b
FAQ
What did Ambarella announce at CES?
What are the target application markets for Ambarella's generative AI networks?
How does Ambarella's SoC solutions compare to GPUs and other AI accelerators?