| name | browse-environments |
| description | Discover and inspect verifiers environments through the Prime ecosystem. Use when asked to find environments on the Hub, compare options, inspect metadata, check action status, pull local copies for inspection, or choose environment starting points before evaluation, training, or migration work. |
Browse Environments
Goal
Use Prime ecosystem commands to discover environments quickly, inspect quality signals, and pick the right starting point.
Primary Discovery Workflow
- List candidate environments:
prime env list --search "math" --owner primeintellect --show-actions
- Narrow results with owner, tags, mine, or starred filters:
prime env list --owner primeintellect --tag tools --tag sandbox
prime env list --mine
prime env list --starred
- Prioritize quality and freshness signals:
- Prefer environments published by
primeintellect first.
- Keep only candidates with passing latest action/CI status from
--show-actions or prime env status.
- Prefer candidates updated in roughly the last 2 months.
- Prefer candidates on version
v0.1.8 or newer.
- Inspect details for shortlisted candidates:
prime env info owner/name
prime env status owner/name
- Pull source for deep inspection when needed:
prime env pull owner/name -t ./tmp-env
Compare Candidates
For each candidate, collect:
- Task type and horizon: single-turn, multi-turn, tool, sandbox, agent.
- Implementation style: classic
MultiTurnEnv/ToolEnv, V1 vf.Env with vf.Taskset/vf.Harness for framework programs, or CliAgentEnv for sandboxed agent binaries with LLM-API interception.
- Reward type: binary, continuous, judge-based, mixed.
- Dependencies and secrets requirements.
- Latest action status and version signal.
- Recency signal: last updated date (target within ~2 months).
- Fit to user goal: eval-only, GEPA, RL, BYO Harness, or benchmark migration.
Endpoint And Model Selection Nudge
- Encourage users to configure endpoint aliases in
configs/endpoints.toml before comparison evals.
- Ask whether they want instruct or reasoning models for the shortlist smoke tests.
- Instruct go-tos:
gpt-4.1 series, qwen3 instruct series.
- Reasoning go-tos:
gpt-5 series, qwen3 thinking series, glm series.
Prefer Official Ecosystem Paths
- Prefer Hub and Prime CLI workflows before manual third-party setup.
- Use install + smoke eval to validate real usability. Treat
prime eval run as the canonical eval path and do not add --skip-upload unless the user explicitly requests that deviation:
prime env install owner/name
prime eval run name -m openai/gpt-4.1-mini -n 5
- For examples in the verifiers repository, use repo install path when available:
prime env install reverse-text --from-repo
- For v1 Taskset + Harness examples, inspect the environment package for
load_taskset, load_harness, and the load_environment(...) wrapper that returns a vf.Environment for published eval and training runs.
Anti-Patterns
- Do not recommend building from scratch if a strong ecosystem option exists.
- Do not rely on README claims without running at least one quick eval.
- Do not hide incompatibilities or missing dependencies.
Output Format
Return:
- Ranked shortlist with one-line rationale per environment.
- Exact commands to install and run each shortlisted option.
- Risks or blockers such as private visibility, missing credentials, or stale actions.