Gemma

Gemma1. Model scale2. Pull Gemma3. Use Gemma3.1. Run Gemma3.2. Start a conversation3.3. End the conversationReferences

Demo Environment

Development board: Jetson Orin series motherboard

SSD: 128G

Tutorial application scope: Whether the motherboard can run is related to the available memory of the system. The user's own environment and the programs running in the background may cause the model to fail to run.

Motherboard modelRun directly with OllamaRun with Open WebUI
Jetson Orin NX 16GB
Jetson Orin NX 8GB√ (need to run the small parameter version)√ (need to run the small parameter version)
Jetson Orin Nano 8GB√ (need to run the small parameter version)√ (need to run the small parameter version)
Jetson Orin Nano 4GB√ (need to run the small parameter version)√ (need to run the small parameter version)

Gemma is a new open model developed by Google and its DeepMind team.

1. Model scale

ModelParameters
Gemma2B
Gemma7B

2. Pull Gemma

Use the pull command to automatically pull the model of the Ollama model library:

Small parameter version model: motherboards with 8G or less memory can run this

image-20250111164424774

3. Use Gemma

3.1. Run Gemma

If the system does not have a running model, the system will automatically pull the Gemma 7B model and run it:

Small parameter version model: motherboards with 8G or less memory can run this

3.2. Start a conversation

The time to reply to the question depends on the hardware configuration, please be patient!

image-20250111164600829

3.3. End the conversation

Use the Ctrl+d shortcut key or /bye to end the conversation!

image-20250111164619979

References

Ollama

Official website: https://ollama.com/

GitHub: https://github.com/ollama/ollama

Gemma

GitHub: https://github.com/google-deepmind/gemma

Ollama corresponding model: https://ollama.com/library/gemma