NVIDIA Orin: bring your next-gen AI products with Jetson AGX Orin and NX Orin

👉 Exciting news! NVIDIA Jetson AGX Orin Developer Kit is available at Seeed now for pre-order, be the first one to get the AGX Orin Dev kit now! Check out our all-new release for NVIDIA Jetson series:

Jetson Orin Production Module Roadmap 2022

From NVIDIA’s updated Jetson roadmap, Jetson Orin production modules will be available starting in 2022. The production module will be essential to product design and following AI products roadmap.

  • Jetson AGX Orin 32GB in July, $899 over 1K units.
  • Jetson AGX Orin 64GB in October, $1599 over 1K units
  • Jetson Orin NX 16GB in September, $599 over 1K units.
  • Jetson Orin NX 8GB in December, $399 over 1K units

NVIDIA Jetson AGX Orin Developer Kit is available at Seeed now for pre-order. If you already used/considering about using AGX Xavier for your product, it will be a good time to have a test with AGX Orin Dev Kit to obtain 6 to 8 times AI performance, 64GB module will be $300 higher than AGX Xavier, however, 32GB version will stay at $899. According to NVIDIA’s roadmap, people can get AGX Orin production module for their mass product as soon as in July, 2022.

Comes to the superior AI performance, Seeed also released reServer – Jetson AGX Orin with pre-built Triton Inference Server, 256 GB NVME SSD, Dual SATA III 6.0, high-speed 10 Gigabit Ethernet port, and ensure the device Chill 24/7 with Vapor Chamber heatsink. reServer Jetson AGX Orin is prepared for the most powerful and relaible local inference server for heavy ML workload and concurrent execution of multiple deep learning applications, bringing fast and scalable AI in production.

In past three years, Xavier NX has been widely deployed into Agriculture, Healthcare, Transporation, and Retail applications. With the same form factor, now AI performance can be lifted up to 5 times if you switch to the Orin NX module. Will Orin NX be your next consideration?

But back to now, don’t miss out Seeed Jetson-20 built with Xavier NX 8GB/16GB, starting from $599, is ready to ship in July! Start your real-world development and deployment quickly with the NX module as well as rich interfaces on a carrier borad, thermal management and reComputer case.

Let’s take a close look at the specification comparison between Xavier NX and Orin NX, AGX Xavier and AGX Orin. We will keep updating our new product progress for Jetson Orin! Please also check out my previous blog Compare NVIDIA Jetson AGX Orin with AGX Xavier: 8x AI performance, in-advance Ampere GPU, CPU, Memory & Storage.

Compare Jetson Orin to Jetson AGX Xavier and Xavier NX

Jetson Benchmark: Jetson Xavier NX and Jetson AGX Orin MLPerf v2.0 Results

ModelJetson Xavier NXJetson AGX XavierJetson AGX Orin
PeopleNet124196536
Action Recognition 2D2454711577
Action Recognition 3D2132105
LPR Net70611904118
Dashcam Net4256711908
Bodypose Net105172559
ASR: Citrinet 10242734113
NLP: BERT-base5894287
TTS: Fastpitch-HifiGAN7942

Jetson AGX Orin 64GB and 32GB Modules

NVIDIA Jetson AGX Orin modules deliver up to 275 TOPS of AI performance with power configurable between 15W and 60W. This gives you up to 8X the performance of Jetson AGX Xavier in the same compact form factor for robotics and other autonomous machine use cases.

Jetson AGX Orin SeriesJetson AGX Xavier Series
Jetson AGX Orin 32GBJetson AGX Orin 64GBJetson AGX Xavier 64GBJetson AGX Xavier 32GB
Price (1K+ Units)$899$1599$1299$899
AI Performance200 TOPS275 TOPS32 TOPs32 TOPs
GPU1792-core NVIDIA Ampere GPU with
56 Tensor Cores
2048-core NVIDIA Ampere GPU with
64 Tensor Cores
512-core NVIDIA Volta™ GPU with
64 Tensor Cores
512-core NVIDIA Volta™ GPU with
64 Tensor Cores
CPU8-core NVIDIA Arm® Cortex A78AE v8.2 64-bit CPU
2MB L2 + 4MB L3
12-core NVIDIA Arm® Cortex A78AE v8.2 64-bit CPU
3MB L2 + 6MB L3
8-core NVIDIA Carmel Arm®v8.2
64-bit CPU
8MB L2 + 4MB L3
8-core NVIDIA Carmel Arm®v8.2
64-bit CPU
8MB L2 + 4MB L3
DL Accelerator2x NVDLA v2.02x NVDLA v2.02x NVDLA v12x NVDLA v1
Memory32 GB64GB64 GB32 GB
Storage64GB eMMC 5.164GB eMMC 5.132 GB eMMC 5.132 GB eMMC 5.1
Power15W – 40W15W – 60W10W | 15W | 30W10W | 15W | 30W
Mechanical100mm x 87mm
699-pin connector Integrated Thermal Transfer Plate
100mm x 87mm
699-pin connector Integrated Thermal Transfer Plate
100 mm x87 mm699-pin connector integrated Thermal Transfer Plate100 mm x87 mm699-pin connector integrated Thermal Transfer Plate

Jetson Orin NX 16GB and 8GB Modules

Jetson Orin NX modules deliver up to 100 TOPS of AI performance in the smallest Jetson form factor, with power configurable between 10W and 25W. The Orin NX Series is form factor compatible with the Jetson Xavier NX series and delivers up to 5x the performance, or up to 3X the performance at the same price. 

Jetson Xavier NX SeriesJetson Orin NX Series
Jetson Xavier NX 16GBJetson Xavier NXJetson Orin NX 8GBJetson Orin NX 16GB
$499$399$399$599
AI Performance21 TOPs21 TOPs70 TOPS100 TOPS
GPU384-core NVIDIA Volta™ GPU with 48 Tensor Cores384-core NVIDIA Volta™ GPU with 48 Tensor Cores1024-core NVIDIA Ampere GPU with
32 Tensor Cores
1024-core NVIDIA Ampere GPU with
32 Tensor Cores
CPU6-core NVIDIA Carmel Arm®v8.2 64-bit CPU
6MB L2 + 4MB L3
6-core NVIDIA Carmel Arm®v8.2 64-bit CPU
6MB L2 + 4MB L3
6-core NVIDIA Arm® Cortex A78AE v8.2 64-bit CPU
1.5MB L2 + 4MB L3
8-core NVIDIA Arm® Cortex A78AE v8.2 64-bit CPU
2MB L2 + 4MB L3
DL Accelerator2x NVDLA v12x NVDLA v11x NVDLA v2.02x NVDLA v2.0
Memory16 GB8 GB8 GB16 GB
Storage16 GB eMMC 5.1 *16 GB eMMC 5.1 *
(Supports external NVMe)

(Supports external NVMe)
Power10W | 15W | 20W10W | 15W | 20W10W – 20W10W – 25W
Mechanical69.6 mm x 45 mm260-pin SO-DIMM connector69.6 mm x 45 mm 260-pin SO-DIMM connector69.6mm x 45mm 260-pin SO-DIMM connector69.6mm x 45mm 260-pin SO-DIMM connector

reComputer Jetson: real world AI at the Edge, starts from $199

Built with Jetson Nano 4GB/ Xavier NX 8GB/16GB

  • Edge AI box fit into anywhere
  • Embedded Jetson Nano/NX Module
  • Pre-installed Jetpack for easy deployment
  • Nearly same form factor with Jetson Developer Kits, with rich set of I/Os
  • Stackable and expandable

reServer Jetson: inference center for the edge

Powered by Jetson AGX Orin / NX

reserver

  • Portable Edge AI center with up to 275 TOPS AI performance
  • Chill 24/7 with Vapor Chamber heatsink (AGX Orin only),
    40% more effective thermal performance
  • Rapid network access and hybrid connectivity
    2.5Gbes, 4G, 5G, LoRa, BLE, and WiFi (modules not included)
  • Built-in storage for multiple concurrent AI applications
  • Support Triton Inference Server and Pre-installed Jetpack
  • Rich IOs ready for AIoT
    USB 3.2 Type-A port, USB 2.0 Type-A port, HDMI port, and DP port

Software Features

In software perspective, Jetson AGX Orin Developer Kit features a powerful NVIDIA native AI software stack with support for SDKs and software platforms, including the following:

  • NVIDIA JetPack
  • NVIDIA Riva
  • NVIDIA DeepStream
  • NVIDIA Isaac
  • NVIDIA TAO

Seeed Jetson series is compatible with the entire NVIDIA Jetson software stack, industry-leading AI development, and device management platforms. Welcome developers to join communities with all the coolest AI ideas and get ready for AI production.

  • Edge Impulse: the leading development platform for machine learning on edge devices, free for developers and trusted by enterprises. Try it now for building models in minutes.
  • AlwaysAI: a complete developer platform for Computer Vision applications on IoT devices. Get started today with our tutorial, and quickly deploy CV applications at the edge.
  • YOLOv5 : YOLO is an abbreviation for the term ‘You Only Look Once’. It is an algorithm that detects and recognizes various objects in an image in real-time. YOLOv5 is the latest version of YOLO which performs much better than the previous iterations of YOLO and it is now based on the PyTorch framework.
  • Roboflow: annotate images super fast, right within your browser. Easily train a working computer vision model.
  • Deci: empowers deep learning developers to accelerate inference on edge or cloud, reach production faster, and maximize hardware potential. Check out How to Deploy and Boost Inference on NVIDIA Jetson.
  • Nimbus: log on to Nimbus, simply drag and drop to configure your robot application in the cloud.
  • Allxon: Enable Efficient Remote Hardware Management Services for NVIDIA Jetson Platform. Enjoy an exclusive 90-day free trial to unlock all functions with code H4U-NMW-CPK. Read our partnership story here.

JETSON BENCHMARKS

NVIDIA Jetson is used to deploy a wide range of popular DNN models and ML frameworks to the edge with high-performance inferencing, for tasks like real-time classification and object detection, pose estimation, semantic segmentation, and natural language processing (NLP). The tables below show inferencing benchmarks from the NVIDIA Jetson submissions to the MLPerf Inference Edge category.

Jetson Xavier NX and Jetson AGX Orin MLPerf v2.0 Results

ModelJetson Xavier NX (TensorRT)NVIDIA Orin (TensorRT)
Image Classification
ResNet-50
1243.146138.84
Object Detection
SSD-small
1782.536883.49
Object Detection
SSD-Large
36.69207.66
Speech to Text
RNN-T
260.521110.23
Natural Language
Processing
BERT-Large
61.40476.34

Jetson AGX Xavier and Jetson Xavier NX MLPerf v1.1 Results

ModelJetson Xavier NX (TensorRT)Jetson AGX Xavier 32GB (TensorRT)
Image Classification
ResNet-50
1245.102039.11
Object Detection
SSD-small
1786.912833.59
Object Detection
SSD-Large
36.9755.16
Speech to Text
RNN-T
259.67416.13
Natural Language
Processing
BERT-Large
61.3496.73

Jetson Pretrained Model Benchmarks

NVIDIA pretrained models from NGC start you off with highly accurate and optimized models and model architectures for various use cases. Pretrained models are production-ready. You can further customize these models by training with your own real or synthetic data, using the NVIDIA TAO (Train-Adapt-Optimize) workflow to quickly build an accurate and ready-to-deploy model. The table below shows inferencing benchmarks for some of pretrained models running on Jetson modules.

Jetson Xavier NX and Jetson AGX Orin MLPerf v2.0 Results

ModelJetson Xavier NXJetson AGX XavierJetson AGX Orin
PeopleNet124196536
Action Recognition 2D2454711577
Action Recognition 3D2132105
LPR Net70611904118
Dashcam Net4256711908
Bodypose Net105172559
ASR: Citrinet 10242734113
NLP: BERT-base5894287
TTS: Fastpitch-HifiGAN7942
  • These Benchmarks were run using Jetpack 5.0

Jetpack 5.0 includes L4T with Linux kernel 5.10 and a reference file system based on Ubuntu 20.04. Jetpack 5.0 enables a full compute stack update with CUDA 11.x and new versions of cuDNN and Tensor RT. It will include UEFI as a CPU bootloader and will also bring support for OP-TEE as a trusted execution environment. Finally, there will be an update to DLA support for NVDLA 2.0, as well as a VPI update to support the next generation PVA v2.

  • Each Jetson module was run with maximum performance (MAX-N mode for Jetson AGX Xavier (32 GB) and Jetson AGX Orin, and 20W mode for Jetson Xavier NX)
  • Reproduce these results by downloading these models from the NVIDIA NGC catalog

NVIDIA Jetson powered Edge AI platform at Seeed

At Seeed, you will find everything you want to work with NVIDIA Jetson Platform – official NVIDIA Jetson Dev Kits, Seeed-designed carrier boards, and edge devices, as well as accessories.

Seeed will continue working on the Jetson product line and will be ready to combine our partners’ unique technology with Seeed’s hardware expertise for an end-to-end solution. 

Join Seeed as an Edge AI Partner: Transform Your Business, Delivering Real-World AI Together

  • Integrate your unique AI technique into our current hardwares: resell or co-brand licensed devices at our channels. 
  • Build your next-gen AI product powered by the NVIDIA Jetson module and bring your product concept to the market with Seeed’s Agile Manufacturing 0-∞.
  • Co-inventing with Seeed. Combining our partners’ unique skills and Seeed’s hardware expertise, let’s unlock next AI ideas 

Calendar

April 2022
M T W T F S S
 123
45678910
11121314151617
18192021222324
252627282930