Skip to main content
At Perceptron, we’re building a family of vision-language models (VLMs) that can see, reason, and act. Our VLMs allow you to:
  • Ask questions about images and videos
  • Clip videos based on specific events
  • Detect and point to objects in images and videos
  • Reason over images with tool calling
Join our Discord community to get help, offer feedback, and see what others are building.

Meet the models

Demo environment

Try prompts and API calls in the browser.

Model playground (coming soon)

Create, iterate on, and save your prompts.

Tutorials