Documentation Index
Fetch the complete documentation index at: https://docs.perceptron.inc/llms.txt
Use this file to discover all available pages before exploring further.
expects="clip" and Isaac 0.3 Max will return one or more Clip objects with start/end timestamps citing the moments that justify the answer. Use it for sports highlights, robot-task success/failure labeling, compliance event detection, and any workflow that turns long video into structured temporal signal.
Basic usage
| Parameter | Type | Default | Description |
|---|---|---|---|
media_obj | VideoNode | - | Wrap your MP4 or WebM (URL or local file path) with video() |
question_text | str | - | Prompt describing what to clip |
reasoning | bool | False | Set True to let the model think through the video before localizing |
expects | str | "text" | Set "clip" to parse <clip> tags emitted by the model into Clip objects |
PerceiveResult object:
text(str): Natural-language answer with inline<clip>tags as the model emitted them.reasoning(str | None): Chain-of-thought whenreasoning=True.clips(list[Clip] | None): Parsed temporal segments. EachCliphas:timestamp.at(float): start in seconds.timestamp.until(float | None): end in seconds, orNonefor a single moment.mention(str | None): optional label the model attached.
Example: Find the shot
In this example we download a short basketball clip, ask Isaac to clip the moment the ball passes through the hoop, and inspect the returned timestamps.Best practices
- Be specific about the event: “Clip the moment the ball passes through the hoop” works better than “find interesting moments.” Tight, observable predicates produce tight clips.
- A single moment vs. a range: When
clip.timestamp.until is None, the model is pointing at a single instant rather than a span. Both are valid; treat the moment case as “approximate point in time” rather than “zero-length range.”
Run through the full Jupyter notebook here. Reach out to Perceptron support if you have questions.