Isaac-0.1 by Perceptron

Note this is the Post-trained model Try out the model on our playground

We're introducing Isaac 0.1, our first perceptive-language model and a major step toward building AI systems that can understand and interact with the physical world. Isaac 0.1 is an open-source, 2B-parameter model built for real-world applications. It sets a new standard for efficiency, delivering capabilities that meet or exceed those of models over 50 times its size.

Founded by the team behind Meta's Chameleon multimodal models, Perceptron is tackling a fundamental challenge: bringing the power of physical AI to the dynamic, multimodal, and real-time environments we live and work in.

Isaac 0.1 is the first in our family of models built to be the intelligence layer for the physical world. It's now available open source for researchers and developers everywhere.

What’s new in Isaac 0.1

Visual QA, simply trained Strong results on standard understanding benchmarks with a straightforward, reproducible training recipe.

Grounded spatial intelligence Precise pointing and localization with robust spatial reasoning. Ask “what’s broken in this machine?” and get grounded answers with highlighted regions—handling occlusions, relationships, and object interactions.

In-context learning for perception Show a few annotated examples (defects, safety conditions, etc.) in the prompt and the model adapts—no YOLO-style fine-tuning or custom detector stacks required.

OCR & fine-grained detail Reads small text and dense scenes reliably, across resolutions, with dynamic image handling for tiny features and cluttered layouts.

Conversational Pointing A new interaction pattern where language and vision stay in lockstep: every claim is grounded and visually cited, reducing hallucinations and making reasoning auditable.

Benchmarks

visual_qa grounding

Example

pip install perceptron

Example using transformers

Learn more: Huggingface Example Repo

!git clone https://github.com/perceptron-ai-inc/perceptron.git
!cp -r perceptron/huggingface ./huggingface
from transformers import AutoTokenizer, AutoConfig, AutoModelForCausalLM
from huggingface.modular_isaac import IsaacProcessor

tokenizer = AutoTokenizer.from_pretrained("PerceptronAI/Isaac-0.1", trust_remote_code=True, use_fast=False)
config = AutoConfig.from_pretrained("PerceptronAI/Isaac-0.1", trust_remote_code=True)
processor = IsaacProcessor(tokenizer=tokenizer, config=config)
model = AutoModelForCausalLM.from_pretrained("PerceptronAI/Isaac-0.1", trust_remote_code=True)
Downloads last month
649
Safetensors
Model size
2.57B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for PerceptronAI/Isaac-0.1

Finetuned
Qwen/Qwen3-1.7B
Finetuned
(260)
this model