Products

Raptor N3000

The Raptor N3000 represents Neuchips' strategic evolution in the AI accelerator market. Originally, Neuchips developed the RecAccel™ N3000 as a specialized ASIC (Application-Specific Integrated Circuit) accelerator focused on Deep Learning Recommendation Models (DLRM). Following ChatGPT's launch in late 2022, Neuchips recognized the growing importance of Large Language Models (LLMs) and swiftly adapted the existing technology to meet market demands.


Overview
Power-Efficient
LLM Accelerator
for Enterprise Local AI Deployment
The Raptor N3000 is an innovative LLM accelerator that demonstrates Neuchips' agile response to the evolving AI landscape. Originally evolved from the RecAccel™ N3000 DLRM accelerator, this solution has been strategically enhanced to address the growing demand for local LLM processing in enterprise environments. At its core, the Raptor N3000 combines efficient power consumption with practical performance capabilities. The accelerator leverages a single-chip ASIC architecture equipped with 64GB DRAM capacity and an embedded vector engine, enabling it to handle both LLM processing and vector database operations locally. This integration makes it particularly valuable for enterprises seeking to implement AI solutions while maintaining data privacy and reducing cloud dependency.

Specification
Raptor Series Specification
Embedded ARC HS48 Processors
  • CPU0 - Dual Core with MM Function
  • CPU1 - Dual Core
Embedded ARC EV72 Processors
  • Vector DSP
  • FPU
Communication InterfacePCle Gen 5x8
Memory
  • 16 LPDDRS-6400 Controllers
  • 512KB Shared Memory
Al Accelerators
  • Matrix Engines × 10
    (Dynamic MLP Engine, DME)
  • Vector Engines x 2 (Feature Cross, FX)
  • Embedding Engine

Product Features

The Enhanced SDK for LLM Support

The enhanced SDK for LLM support delivers a powerful yet accessible development toolkit designed for seamless LLM integration. The SDK features optimized model integration capabilities and intuitive API interfaces, enabling developers to quickly implement LLM functionalities. With pre-built optimization tools, developers can easily enjoy fine-tune model efficiency. The SDK maintains an optimal processing rate of 8~10 tokens per second, matching natural human reading speed for smooth interaction. This comprehensive solution streamlines development while ensuring optimal performance, making advanced LLM capabilities more accessible to developers of all skill levels.


Adaptive AI Acceleration for Enterprise

The N3000 builds upon a proven foundation, leveraging established architecture and a mature software ecosystem from previous generations. With reliable performance metrics validated through extensive real-world deployment, it delivers consistent and predictable results. The platform features integrated local vector database capabilities, enabling efficient data processing and retrieval directly on the device. This combination of proven technology and modern capabilities makes the N3000 a reliable choice for advanced processing needs.


Key Benefits
Cost Efficiency
  • Leverages existing ASIC architecture
  • Reduced development and deployment costs
  • Lower total cost of ownership
Power Optimization
  • Low power consumption
  • Reduced cooling requirements
  • Lower operating costs
  • Environmentally friendly operation
Practical Performance
  • Performance matched to human reading speed
  • Sufficient for enterprise use cases
  • Local vector search capabilities
  • Efficient memory management with 64GB DRAM
Deployment Flexibility
  • Single-chip solution
  • Compact form factor
  • Easy integration into existing systems
  • Local data processing and storage
Top