A fast CLI and TUI for browsing AI models, benchmarks, and coding agents.
- Models Tab: Browse 2000+ models across 85+ providers from models.dev, categorized by type (Origin, Cloud, Inference, Gateway, Dev Tool)
- Agents Tab: Track AI coding assistants (Claude Code, Aider, Cursor, etc.) with version detection and GitHub integration
- Benchmarks Tab: Compare model performance across 15+ benchmarks from Artificial Analysis, with creator filtering by source, region, and type
models-demo.mp4
- Split cost columns — Models list now shows separate Input and Output cost columns with smart rounding
- Dynamic column widths — Model ID and Benchmark Name columns expand to fill available terminal width
- Focus-aware caret —
>indicator acts as a pseudo cursor, appearing only on the active panel - Dynamic panel titles — Models list shows the selected provider name; Benchmarks list shows the selected creator name
- Styled column headers — yellow bold headers with cyan highlight on the active sort column
g/Gkeybindings — jump to first/last item on Benchmarks and Agents tabs (already on Models)- Improved context formatting — fractional k/M values display correctly (e.g.,
131.1k,1.5M) - Creator panel truncation — long creator names truncate cleanly within the sidebar width
- Benchmarks detail panel — Region and Business Type now have their own labeled property lines
- Auto-updating benchmarks — benchmark data refreshes from CDN every 6 hours in the background, no user configuration required
- Disk cache — previously fetched data persists across sessions at
~/.config/models/benchmarks-cache.json - Offline fallback — embedded data is always available when the network is unreachable
- GitHub Action — daily automated refresh of benchmark data from the Artificial Analysis API
- Dedicated Benchmarks tab — browse ~400 model entries from Artificial Analysis with quality, speed, and pricing data
- Creator sidebar with 40+ creators, each classified by source (Open/Closed/Mixed), region (US/China/Europe/Middle East/S. Korea/Canada/Other), and type (Startup/Big Tech/Research)
- Three filter dimensions —
[4]source,[5]region,[6]type — to slice the creator and benchmark lists - Quick-sort keys —
[1]Intelligence,[2]Date,[3]Speed — press again to flip direction - Dynamic column visibility — list columns adapt based on the active sort group (knowledge, code, reasoning, math, performance)
- Detail panel — non-scrollable flat layout with indexes, benchmark scores, performance metrics, and pricing
- Null-filtering — entries missing data for the active sort column are hidden automatically
- Provider categories — filter and group providers by type (Origin, Cloud, Inference, Gateway, Dev Tool)
- OpenClaw agent added to the agents catalog
- Responsive layouts — models tab detail panel scales with terminal height
- CLI commands for scripting and quick lookups
- Interactive TUI for browsing and comparing models
- Provider categories — filter and group providers by type (Origin, Cloud, Inference, Gateway, Dev Tool)
- Cross-provider search to compare the same model across different providers
- Copy to clipboard with a single keypress
- JSON output for scripting and automation
- Curated catalog of 12+ AI coding assistants
- Version detection — automatically detects installed agents
- GitHub integration — stars, releases, changelogs, update availability
- Persistent cache — instant startup with ETag-based conditional fetching
- Customizable tracking — choose which agents to monitor
- ~400 benchmark entries from Artificial Analysis with quality, speed, and pricing scores
- Auto-updating — data refreshes from CDN every 6 hours in the background with offline fallback
- Creator sidebar with 40+ creators — filter by source (Open/Closed/Mixed), region (US/China/Europe/...), and type (Startup/Big Tech/Research)
- Quick-sort keys — instantly sort by Intelligence, Date, or Speed
- Dynamic columns — list columns adapt to show the most relevant benchmarks for the active sort
- Detail panel — full benchmark breakdown with indexes, scores, performance, and pricing
cargo install modelsdevbrew install arimxyer/tap/modelsscoop bucket add arimxyer https://github.com/arimxyer/scoop-bucket
scoop install modelsDownload the latest release for your platform from GitHub Releases.
git clone https://github.com/arimxyer/models
cd models
cargo build --release
./target/release/modelsJust run models with no arguments to launch the interactive browser:
modelsGlobal
| Key | Action |
|---|---|
] / [ |
Switch tabs (Models / Agents / Benchmarks) |
? |
Show context-aware help |
q |
Quit |
Navigation
| Key | Action |
|---|---|
j / ↓ |
Move down |
k / ↑ |
Move up |
g |
Jump to first item |
G |
Jump to last item |
Ctrl+d / PageDown |
Page down |
Ctrl+u / PageUp |
Page up |
Tab / Shift+Tab |
Switch panels |
← / → |
Switch panels |
Search
| Key | Action |
|---|---|
/ |
Enter search mode |
Enter / Esc |
Exit search mode |
Esc |
Clear search (in normal mode) |
Filters & Sort
| Key | Action |
|---|---|
s |
Cycle sort (name → date → cost → context) |
1 |
Toggle reasoning filter |
2 |
Toggle tools filter |
3 |
Toggle open weights filter |
4 |
Cycle provider category filter (All → Origin → Cloud → Inference → Gateway → Tool) |
5 |
Toggle category grouping |
Copy & Open
| Key | Action |
|---|---|
c |
Copy provider/model-id |
C |
Copy model-id only |
o |
Open provider docs in browser |
D |
Copy provider docs URL |
A |
Copy provider API URL |
Filters & Sort
| Key | Action |
|---|---|
s |
Cycle sort (name → updated → stars → status) |
1 |
Toggle installed filter |
2 |
Toggle CLI tools filter |
3 |
Toggle open source filter |
Actions
| Key | Action |
|---|---|
a |
Open tracked agents picker |
o |
Open docs in browser |
r |
Open GitHub repo |
c |
Copy agent name |
By default, models tracks 4 popular agents: Claude Code, Codex, Gemini CLI, and OpenCode.
Press a in the Agents tab to open the picker and customize which agents you track. Your preferences are saved to ~/.config/models/config.toml.
You can also add custom agents not in the catalog:
# ~/.config/models/config.toml
[[agents.custom]]
name = "My Agent"
repo = "owner/repo"
binary = "my-agent"
version_command = ["--version"]Quick Sort (press again to toggle direction)
| Key | Action |
|---|---|
1 |
Sort by Intelligence index |
2 |
Sort by Release date |
3 |
Sort by Speed (tok/s) |
Filters
| Key | Action |
|---|---|
4 |
Cycle source filter (Open / Closed / Mixed) |
5 |
Cycle region filter (US / China / Europe / Middle East / S. Korea / Canada / Other) |
6 |
Cycle type filter (Startup / Big Tech / Research) |
Sort (full cycle)
| Key | Action |
|---|---|
s |
Cycle through all 16 sort columns |
S |
Toggle sort direction (asc/desc) |
Actions
| Key | Action |
|---|---|
c |
Copy benchmark name |
o |
Open Artificial Analysis page |
models list providers# All models
models list models
# Models from a specific provider
models list models anthropicmodels show claude-opus-4-5-20251101Claude Opus 4.5
===============
ID: claude-opus-4-5-20251101
Provider: Anthropic (anthropic)
Family: claude-opus
Limits
------
Context: 200k tokens
Max Output: 64k tokens
Pricing (per million tokens)
----------------------------
Input: $5.00
Output: $25.00
Cache Read: $0.50
Cache Write: $6.25
Capabilities
------------
Reasoning: Yes
Tool Use: Yes
Attachments: Yes
Modalities: text, image, pdf -> text
Metadata
--------
Released: 2025-11-01
Updated: 2025-11-01
Knowledge: 2025-03-31
Open Weights: No
models search "gpt-4"
models search "claude opus"All commands support --json for scripting:
models list providers --json
models show claude-opus-4-5 --json
models search "llama" --jsonLots of gratitude and couldn't have made this application without these workhorses doing the legwork. Shout out to the sources!:
- Model data: Fetched from models.dev, an open-source database of AI models maintained by SST
- Benchmark data: Fetched from Artificial Analysis — quality indexes, benchmark scores, speed, and pricing for ~400 model entries
- Agent data: Curated catalog in
data/agents.json— contributions welcome! - GitHub data: Fetched from GitHub API (stars, releases, changelogs)
MIT


