Jetson Generative AI – Jetson Copilot

Jetson Generative AI – Jetson Copilot

Large Language Models become exponentially more powerful when combined with Retrieval-Augmented Generation (RAG)—Jetson Copilot demonstrates how to run open-source LLMs locally with access to your indexed knowledge base, creating a personalized AI assistant that understands your specific domain and documents in real time on your Jetson device.

In this article you’ll learn how to set up and run Jetson Copilot, a reference application that combines local LLM inference with RAG capabilities for intelligent document search and question answering.

Requirements

 

 

Hardware / Software Notes
Jetson AGX Orin (64GB)
Recommended for best performance
Jetson AGX Orin (32GB)
Good performance for most use cases        
Jetson Orin Nano (8GB)
Minimum requirement

JetPack 5 (L4T r35.x) 

JetPack 6 (L4T r36.x)
Both versions supported
NVMe SSD highly recommended
For storage speed and space                 
6GB for jetrag container
Container image storage
~4GB for default models
llama3 and mxbai-embed-large models

What is Jetson Copilot?

Jetson Copilot is a reference application for a local AI assistant that demonstrates:

  • Running open-source LLMs (large language models) on device
  • RAG (retrieval-augmented generation) to let LLM access your locally indexed knowledge
  • Document indexing and search for personalized knowledge bases
  • Web-based interface for easy interaction and management

Note: You do not need jetson-containers installed on your system. Jetson Copilot uses the jetrag container image that is managed and built separately.

Step-by-Step Setup

1.  Clone the repository

Copy to Clipboard

2.  Enter the directory

Copy to Clipboard

3.  First-time environment setup

If this is your first time running Jetson Copilot, run the setup script to ensure all necessary software is installed:

Copy to Clipboard

4.  Launch Jetson Copilot

Copy to Clipboard

5. Access the Web Interface

The console will show two access options:

Local Access (on Jetson):

Copy to Clipboard

Network Access (from other devices)

Copy to Clipboard

Tip: On Ubuntu Desktop, a frameless Chromium window will automatically open to make it look like a standalone application.

How to Use Jetson Copilot

1. Basic LLM Interaction

You can use Jetson Copilot as a standalone LLM without enabling RAG features. By default, Llama3 (8B) is downloaded on first run and used as the default model. This provides impressive capabilities but may have limitations regarding:

  •  Information after its training cutoff date
  •  Domain-specific or personal knowledge

2. Using Pre-built Knowledge Index

To enable RAG capabilities:

  1. “Toggle “Use RAG” in the side panel
  2. Select an index under the “Index” dropdown
Pre-built Sample:
  • _L4T_README index is provided as a demonstration
  • Built from README files in the L4T-README folder on Jetson desktopMounted at `/media//L4T-README/` after running:
Copy to Clipboard

Example Questions:

What IP address does Jetson get assigned when connected to a PC via USB cable in USB Device Mode?

3. Building Your Own Knowledge Index

Create custom knowledge bases from your documents:

Step 3.1 Prepare Your Documents

Copy to Clipboard

Step 3.2 Build the Index via Web Interface

  1. Open the sidebar and toggle on “Use RAG”
  2. Click “➕Build a new index” to access the Build Index page
  3. Give your index a name (e.g., “JON Carrier Board”)
  4. Select your document directory from the dropdown (e.g., `/opt/jetson_copilot/Documents/Jetson-Orin-Nano`)
  5. Add online URLs (optional) – one URL per line in the text area
  6. Ensure mxbai-embed-large is selected as the embedding model
  7. Click “Build Index” and monitor progress in the status dropdown

Step 3.3 Use Your Custom Index

Once built, return to the home screen and select your newly created index from the dropdown.

Advanced Features

Different LLM Models

You can test and switch between different language models:

  • Llama3 (8B) – Default, good balance of performance and accuracy
  • Additional models can be downloaded via the Ollama interface
  • Model selection available in the web interface settings

Different Embedding Models

For RAG functionality, you can choose embedding models:

  • mxbai-embed-large – Recommended default
  • OpenAI embedding models – Experimental support (needs testing)

Document Types Supported

Jetson Copilot can index various document formats:

  • PDF files
  • Text files
  • Markdown files
  • **Web pages** (via URL ingestion)

Development and Customization

Development Mode

For developers wanting to modify the Streamlit application:

  1. Enable auto-rerun in the web UI (top-right corner)
  2. Choose “Always rerun” to automatically update when you change source code

Manual Development Setup

For more fundamental changes:
Copy to Clipboard

Once inside the container:

Copy to Clipboard

Directory Structure

Copy to Clipboard

Troubleshooting

 

Issue Fix
Container fails to start
Ensure Docker is properly installed via `./setup_environment.sh`
Models not downloading
Check internet connection and available disk space (>10GB)
Chromium window won’t close
Manually close Chromium window; stopping container doesn’t close browser
RAG not working
Verify index is built and selected, check embedding model
Out of memory error
Close other applications, consider using smaller model

For more information about Jetson Copilot and advanced configurations, visit the Jetson Copilot GitHub repository.