Shanghai Neardi Technology Co., Ltd. sales@neardi.com +86 17612192553
Why are more and more edge devices talking about NPUs and coprocessors? The RK3588 is already a powerful 6 TOPS (INT8) SoC, yet in complex scenes such as multi-task inference, model parallelism and video-AI analytics the compute ceiling of a single chip is still there. RK1820 was created exactly to take over that slice of load and relieve the main SoC’s “compute anxiety”. In edge-AI equipment the host processor no longer fights alone; when AI tasks outgrow the scheduling capacity of the traditional CPU/NPU, the coprocessor quietly steps in and assumes part of the intelligent workload.
![]()
RK1820 is a coprocessor purpose-built for AI inference and compute expansion; it pairs flexibly with host SoCs such as RK3588 and RK3576 and communicates with them efficiently through PCIe or USB interfaces.
| Capability Category | Key Parameters & Functions |
|---|---|
| Processor Architecture | 3× 64-bit RISC-V cores; 32 KB L1 I-cache + 32 KB L1 D-cache per core, 128 KB shared L2 cache; RISC-V H/F/D-precision FPU |
| Memory | 2.5 GB on-chip high-bandwidth DRAM + 512 KB SRAM; external support for eMMC 4.51 (HS200), SD 3.0, SPI Flash |
| Codec | JPEG encode: 16×16–65520×65520, YUV400/420/422/444; JPEG decode: 48×48–65520×65520, multiple YUV/RGB formats |
| NPU | 20 TOPS INT8; mixed-precision INT4/INT8/INT16/FP8/FP16/BF16; frameworks: TensorFlow/MXNet/PyTorch/Caffe; Qwen2.5-3B (INT4) 67 token/s, YOLOv8n (INT8) 125 FPS |
| Communication | PCIe 2.1 (2 lanes, 2.5/5 Gbps), USB 3.0 (5 Gbps, shared with PCIe) |
| Main Functions | Edge-AI inference (detection / classification / LLM), RISC-V general compute, 2-D graphics acceleration (scale / rotate), AES/SM4 security |
In the RK3588 + RK1820 system, the AI-task pipeline is decomposed into a four-tier architecture:
Application → Middleware → Co-processor Execution → Control & Presentation.
RK3588 host: handles task scheduling, data pre-processing, and result output, governing the entire workflow.
RK1820 co-processor: dedicated to high-compute AI inference, coupled to the host via PCIe, forming a “light control + heavy compute” collaboration model.
| Stage | Actor | Action |
|---|---|---|
| App Request | RK3588 | AI-task call issued from app layer (recognition/detection) |
| Dispatch | RK3588 dispatcher | Decide whether to offload to co-processor |
| Inference | RK1820 | Run deep-learning model computation |
| Return | RK1820 → RK3588 | Send back inference results; host displays or continues logic |
The application layer is where every AI task begins; it translates user requirements—image analytics, object detection, edge-side LLM Q&A, etc.—into system-executable task commands and passes them to the middleware layer through standardized APIs. This layer is handled entirely by the RK3588 host, which manages user interaction, business logic, and peripheral data.
![]()
Task reception: acquires user commands via cameras, touch panels, Ethernet, UART, etc.
Command standardization: turns unstructured input into structured task parameters
The middleware layer is the collaborative hub: it judges each task, allocates resources, preprocesses data, and governs bus traffic. It decides whether the task runs on the host or is off-loaded to the co-processor.
RK3588 only; RK1820 takes no part in PCIe configuration or interrupt management—it simply executes the inference jobs dispatched by the host.
Task classification and scheduling
Data preprocessing
![]()
Bus communication control
This layer is the inference core, driven exclusively by the RK1820 co-processor, dedicated to high-compute AI inference.
RK1820 active; RK3588 does not interfere with inference, it only waits for results. Time-out or exceptions are handled by RK3588 via PCIe reset commands.
Task reception and preparation
Receives data, model weights, and commands dispatched by RK3588; writes them into local high-bandwidth DRAM, loads the model, and configures the NPU.
NPU inference compute
Result return
This layer is the terminus of every AI task: it converts the raw inference results from RK1820 into visual or business-ready output and closes the loop.
RK3588 active; RK1820 only supplies the raw inference data.
![]()
Result post-processing
System control & feedback output
Value of synergy: not just faster, but smarter
| Stage | Actor | Action |
|---|---|---|
| App Request | RK3588 | AI-task call issued from app layer (recognition/detection) |
| Dispatch | RK3588 dispatcher | Decide whether to offload to co-processor |
| Inference | RK1820 | Run deep-learning model computation |
| Return | RK1820 → RK3588 | Send back inference results; host displays or continues logic |
Put simply: RK3588 runs the show and keeps everything on track, while RK1820 delivers raw compute bursts; together they make edge-AI devices “smarter, faster, and hassle-free.”
Follow us for more RK1820 news and SDK updates, fresh tutorials and ready-to-run demos.