How to get, install, and use NVIDIA’s Jetson micro servers for AI at edge.
From https://developer.nvidia.com/embedded/develop/software
https://github.com/NVIDIA/GenerativeAIExamples
NVIDIA has a program for training and certifying university educators and certifying Jetson AGX Orin developers.
Launch the system compatibility check.
Http failure response for https://market-production.azurewebsites.net/api/Organization/StoreExams/0894ab5e-cc68-4f2b-a8bc-d41a576d345d/slots?startDate=2/22/2025,%2012:00:00%20AM&endDate=2/22/2025,%2011:59:59%20PM&timeZone=America/Denver&slotDurationInMinutes=60: 401 OK
NVIDIA offers a 50 (40-60) question exam in one-hour with no breaks taken online, each at $135 for each retake. It’s good for a 2-year validity period.
NVIDIA-Certified Associate: Generative AI and LLMs (NCA-GENL) validates skills in the use of generative AI and large language models:
30% Core Machine Learning and AI Knowledge
24% Software Development
22% Experimentation
14% Data Analysis and Visualization
10% Trustworthy AI
Study materials:
8-hour $90 Generative AI With Diffusion Models (to generate images from text)
NVIDIA-Certified Associate: Generative AI Multimodal (NCA-GENM)
25% Experimentation
20% Core Machine Learning and AI Knowledge
15% Multimodel Data *
15% Software Development
10% Data Analysis and Visualization
10% Performance Optimization *
5% Trustworthy AI
Notice the two topics added (marked by *).
In addition to the resources for the GENL exam:
NVIDIA-Certified Associate: AI Infrastructure and Operations (NCA-AIIO) validates fundamental skills in AI infrastructure and operations learned from Study Guide
15% Troubleshoot and Optimize
17% Systems and Network
33% Systems and Servers
35% Physical Layer Management
Manage cloud-native stack
7-hour $150 AI Infrastructure Operations Fundamentals with exam coupon. This covers compute platforms, networking, and storage solutions. The course also addresses AI operations, focusing on infrastructure management and cluster orchestration.
NVIDIA-Certified Professional: AI Infrastructure (NCP-AII), for $400 answer 50 questions in 90-minutes to validates the ability to deploy, manage, and maintain AI infrastructure by NVIDIA.
NVIDIA-Certified Professional: AI Operations (NCP-AIO) has 2-3 year preprequisite. For $400, answer 50 questions in 90-minutes to validate your ability to monitor, troubleshoot, and optimize AI infrastructure by NVIDIA.
36% Administration
Troubleshoot storage performance
7-hour $150 AI Infrastructure & Operations Fundamentals includes exam certificate. covers essential components of AI infrastructure, including compute platforms, networking, and storage solutions. The course also addresses AI operations, focusing on infrastructure management and cluster orchestration.
NVIDIA-Certified Professional: InfiniBand (NCP-IB). For $220 answer 40 questions in 90-minutes to validate skills in AI networking by NVIDIA. Correctly answer 40 questions in 90-minutes online, for $220, with a 2-year validity period for those who installs, configures, manages, troubleshoots, or monitors InfiniBand fabrics.
https://developer.nvidia.com/embedded/jetson-modules
All Jetson https://developer.nvidia.com/buy-jetson?product=all&location=US
Developer Kits:
BTW: AGX is “not an acronym persay, but it loosely means Autonomous machines accelerator technology.”
Others:
Previous :
NVIDIA has a different SDK for different hardware
https://developer.nvidia.com/embedded/learn/jetson-orin-nano-devkit-user-guide/index.html Jetson Orin Nano Developer Kit User Guide
https://www.nvidia.com/en-au/glossary/
CUDA is NVIDIA’s proprietary software for parallel computing on GPUs. Its competitor is Intel’s DPC++ (Data Parallel C++).
CUDA 12.6
CSP & 3P Service
HPC (High Performance Computing)
Jetson is NVIDIA’s proprietary GPU computing platform
Jetson Linux 36.4 provides the Linux Kernel 5.15, UEFI based bootloader, Ubuntu 22.04 based root file system, NVIDIA drivers, necessary firmwares, toolchain and more.
Jetson Platform Services (available soon.) is a collection of pre-built and cloud-native software services and reference workflows to accelerate AI applications on Jetson. These services are modular, API-driven and can be quickly configured to build Generative AI and other edge applications. There are 15+ services from AI services to system services. The services include:
Metropolis application framework to build, deploy and scale Vision AI application https://www.nvidia.com/en-us/autonomous-machines/intelligent-video-analytics-platform/
Holoscan for building high performance computing applications (HPC) with real time insights and sensor processing capabilities from edge to cloud. https://www.nvidia.com/en-us/clara/holoscan/
OpenACC
OpenVLA (Vision-Language-Action) Model https://openvla.github.io
OpenUSD
TAO Toolkit https://developer.nvidia.com/tao-toolkit
NVIDIA NIM, part of NVIDIA AI Enterprise, is a set of intuitive inference microservices designed to accelerate generative AI deployment in enterprises. NIM microservices provide interactive APIs to run inference on AI models.
Each NIM is packaged as a Docker container image on a per model or model family basis.
NIM supports a wide range of AI models—including
NIM uses NVIDIA TensorRT-LLM to optimize the models, with specialized accelerated profiles optimally selected for:
https://learn.nvidia.com/courses/course-detail?course_id=course-v1:DLI+S-OV-19+V1
Each file format can be created through Python bindings in the OpenUSD library. When creating a new stage we can pass in a string to represent a file name that ends in .usdc, .usd, .usda, or .usdz. File Formats (USD, USDC, USDA and USDZ) are used for storing and exchanging various types of 3D scene data, including meshes, cameras, lights, and shaders.
A USD (.usd) file can have either ASCII or binary format. This switch can be done at any point without breaking references for debugging.
Separate heavier data from more light weight data. When doing so, consider using .usdc and .usda explicitly to avoid obfuscation and create large .usda files unintentionally.
USDA (.usda) is a native file format used by OpenUSD to store and exchange 3D scene data. Its format is ASCII text and therefore “Human Readable” and editable. This makes USDA optimal for small files, such as a stage that is referencing external content.
USDC (.usdc) - the Crate Binary Format – is a compressed binary file format designed to minimize load time and provide a more efficient representation of the scene data compared to the human-readable ASCII format (USDA). USDC is extremely efficient for numerically-heavy data, like geometry. Various compression techniques reduce the file size and improve loading performance. It also employs memory mapping for faster file access and loading times.
USDZ (.usdz) is an atomic, uncompressed, zipped archive for delivery of all necessary assets ( a mesh with its texture) together in a single file. It’s generally intended as read-only and is optimal for XR experiences. We would not use USDZ if we are still making edits to the asset.
We may choose to use some other 3D format backed by an SdfFileFormatPlugin when we prefer to keep our source data as is and still leverage all of OpenUSD for scene manipulation and rendering.
https://www.nvidia.com/gtc/pricing/?nvid=nv-int-unbr-171401 Exhibits March 18–21 | Workshops March 16–20 | San Jose, CA & Virtual
https://forums.developer.nvidia.com/c/agx-autonomous-machines/jetson-embedded-systems/70 NVIDIA Community
TwitterX
https://www.youtube.com/@NVIDIADeveloper YouTube
https://developer.nvidia.com/embedded/learn/get-started-jetson-orin-nano-devkit The NVIDIA® Jetson Orin Nano™ Developer Kit empowers the development of AI-powered robots, smart drones, and intelligent cameras built on the Jetson Orin series.
https://learn.nvidia.com/en-us/training/self-paced-courses
https://learn.nvidia.com/courses/course-detail?course_id=course-v1:DLI+C-RX-02+V2
https://www.nvidia.com/en-us/training/ DLI (DEEP LEARNING Institute)
https://www.jetson-ai-lab.com/tutorial-intro.html
https://www.jetson-ai-lab.com/ros.html
The 22GB for nano_llm:humble container image ros2_nanollm package provides ROS2 nodes for running optimized LLM’s and VLM’s locally inside a container. These are built on NanoLLM and ROS2 Humble for deploying generative AI models onboard your robot with Jetson.
https://developer.nvidia.com/embedded/jetpack
Download Jetson Orin Nano Super Developer Kit https://developer.nvidia.com/downloads/embedded/L4T/r36_Release_v4.0/jp61-rev1-orin-nano-sd-card-image.zip
Download JETSON ORIN NANO DEVELOPER KIT SD card image from https://developer.nvidia.com/embedded/jetpack
https://docs.nvidia.com/jetson/archives/r36.4/DeveloperGuide/SD/Security/FirmwareTPM.html Firmware-based Trusted Platform Module (fTPM) on the Orin platform. Refer to the security page for all security features.
sudo apt dist-upgrade sudo apt-install nvidia-jetpack
## Keyboard Shortcuts
https://www.youtube.com/watch?v=N_OOfkEWcOk Within https://github.com/NVIDIA/GenerativeAIExamples https://github.com/NVIDIA/GenerativeAIExamples/tree/main/community/5_mins_rag_no_gpu Run using Streamlit: