🌟 Gemma 3B 3N
Google Gemma 3B 3N for free?
Check out what the 3N variant is, how to use it in Colab, Python, or Inference Server with Unsloth.
🤔 What is Gemma 3N?
Gemma 3N is our quantized 3-bit version of Google’s Gemma 3B model.
- It is designed to be 3x smaller in memory while still maintaining strong accuracy.
- It can run on RTX 3060, Colab T4, and even Apple M1 / M2!
- You can also fine-tune Gemma 3N for even better results.
We call it 3N (3-bit Normal) because it’s a 3-bit quantization of the normal float16 model.
❓ Is there an official 3N from Google?
No.
- Google only released Gemma 2B, Gemma 7B, and Gemma 1.1 9B (for float16/bfloat16 and int8).
- There is no 3-bit official release from Google.
Our 3N variant is an Unsloth quantization for ultra-low-memory use cases.
📦 Where to get Gemma 3N
You can download our Unsloth Gemma 3N weights from Hugging Face:
It’s open and freely available.
💻 How to use Gemma 3N in Colab
Run the following code in Google Colab:
!pip install unsloth
from unsloth import FastLanguageModel
model, tokenizer = FastLanguageModel.from_pretrained(
"unsloth/gemma-3b-3n",
load_in_4bit = False,
)
image: images/gemma 1.png # <<< THIS IS YOUR HERO IMAGE
description: "Google Gemma 3B 3N for free?..."
linkedin_url: "https://www.linkedin.com/company/unsloth"
---