Skip to content

t1094.1: Update build-agent to reference pattern data for model tier recommendations#1633

Merged
marcusquinn merged 2 commits intomainfrom
feature/t1094.1
Feb 18, 2026
Merged

t1094.1: Update build-agent to reference pattern data for model tier recommendations#1633
marcusquinn merged 2 commits intomainfrom
feature/t1094.1

Conversation

@marcusquinn
Copy link
Owner

@marcusquinn marcusquinn commented Feb 18, 2026

Update build-agent.md to use evidence-based model tier recommendations via pattern data, not just static rules.

Changes

  • New Model Tier Selection: Evidence-Based Routing section in build-agent.md
  • Quick Reference update: prompt to check pattern data before setting model: in frontmatter
  • Shows how pattern-tracker-helper.sh suggest/recommend provides data-driven tier selection
  • Concrete examples: "pattern data shows sonnet with prompt-repeat is optimal for shell-script agents (87% success, 14 samples)"
  • Decision table: when to trust pattern data vs routing rules (>75% success, 3+ samples threshold)
  • Documents how to record outcomes to build pattern history
  • Explains the philosophical shift: evidence beats rules, self-correcting as patterns accumulate

Motivation

The agent design guide previously only referenced static routing rules ("use opus for architecture"). This update adds the missing link to pattern data — empirical evidence from actual task outcomes — so agents can make data-driven model tier decisions rather than rule-based guesses.

Ref #1625

Summary by CodeRabbit

  • Documentation
    • Added comprehensive guidance on model tier selection using evidence-based routing.
    • Included decision criteria and examples for choosing between model tiers based on empirical pattern data.
    • Provided procedures for recording outcomes and validating model choices with supporting evidence.

…ommendations (t1094.1)

Add evidence-based model routing guidance to build-agent.md:
- New 'Model Tier Selection: Evidence-Based Routing' section with pattern-tracker-helper.sh examples
- Quick Reference update: prompt to check pattern data before setting model: in frontmatter
- Shows how pattern data (>75% success, 3+ samples) overrides static routing rules
- Concrete examples: 'pattern data shows sonnet with prompt-repeat is optimal for shell-script agents'
- Documents how to record outcomes to build pattern history over time
- Explains why evidence beats rules: self-correcting as patterns accumulate
@gemini-code-assist
Copy link

Summary of Changes

Hello @marcusquinn, I'm Gemini Code Assist1! I'm currently reviewing this pull request and will post my feedback shortly. In the meantime, here's a summary to help you and other reviewers quickly get up to speed!

This pull request updates the agent design guide to incorporate evidence-based model tier recommendations. It moves beyond static rules by integrating pattern data derived from actual task outcomes, enabling agents to make more informed and adaptive decisions about which language model to use. The changes aim to improve the reliability and effectiveness of agent operations by leveraging empirical evidence for model selection.

Highlights

  • Evidence-Based Model Tier Selection: Introduced a new section 'Model Tier Selection: Evidence-Based Routing' in build-agent.md to guide agents in making data-driven model tier decisions.
  • Quick Reference Update: Updated the quick reference to advise checking pattern data using pattern-tracker-helper.sh before setting the model: in frontmatter.
  • Pattern Data Utility: Documented how pattern-tracker-helper.sh suggest/recommend provides data-driven tier recommendations with concrete examples.
  • Decision Logic for Pattern Data: Provided a decision table outlining when to trust pattern data over static routing rules, specifically when pattern data shows >75% success rate with 3+ samples.
  • Outcome Recording: Explained how to record task outcomes (success/failure) to build and refine pattern history for model selection.
  • Philosophical Shift: Emphasized a shift from static, rule-based model selection to an evidence-based, self-correcting approach using empirical pattern data.
Changelog
  • .agents/tools/build-agent/build-agent.md
    • Added a new section 'Model Tier Selection: Evidence-Based Routing' to detail the use of pattern data for model selection.
    • Updated the quick reference section to include guidance on checking pattern data before specifying a model in agent frontmatter.
    • Included examples of using pattern-tracker-helper.sh for model recommendations and recording outcomes.
    • Provided a decision table for when to prioritize pattern data over static routing rules.
Using Gemini Code Assist

The full guide for Gemini Code Assist can be found on our documentation page, here are some quick tips.

Invoking Gemini

You can request assistance from Gemini at any point by creating a comment using either /gemini <command> or @gemini-code-assist <command>. Below is a summary of the supported commands on the current page.

Feature Command Description
Code Review /gemini review Performs a code review for the current pull request in its current state.
Pull Request Summary /gemini summary Provides a summary of the current pull request in its current state.
Comment @gemini-code-assist Responds in comments when explicitly tagged, both in pull request comments and review comments.
Help /gemini help Displays a list of available commands.

Customization

To customize Gemini Code Assist for GitHub experience, repository maintainers can create a configuration file and/or provide a custom code review style guide (such as PEP-8 for Python) by creating and adding files to a .gemini/ folder in the base of the repository. Detailed instructions can be found here.

Limitations & Feedback

Gemini Code Assist may make mistakes. Please leave feedback on any instances where its feedback is incorrect or counter productive. You can react with 👍 and 👎 on @gemini-code-assist comments. If you're interested in giving your feedback about your experience with Gemini Code Assist for Github and other Google products, sign up here.

You can also get AI-powered code generation, chat, as well as code reviews directly in the IDE at no cost with the Gemini Code Assist IDE Extension.

Footnotes

  1. Review the Privacy Notices, Generative AI Prohibited Use Policy, Terms of Service, and learn how to configure Gemini Code Assist in GitHub here. Gemini can make mistakes, so double check it and use code with caution.

@coderabbitai
Copy link
Contributor

coderabbitai bot commented Feb 18, 2026

Warning

Rate limit exceeded

@marcusquinn has exceeded the limit for the number of commits that can be reviewed per hour. Please wait 7 minutes and 19 seconds before requesting another review.

⌛ How to resolve this issue?

After the wait time has elapsed, a review can be triggered using the @coderabbitai review command as a PR comment. Alternatively, push new commits to this PR.

We recommend that you space out your commits to avoid hitting the rate limit.

🚦 How do rate limits work?

CodeRabbit enforces hourly rate limits for each developer per organization.

Our paid plans have higher rate limits than the trial, open-source and free plans. In all cases, we re-allow further reviews after a brief timeout.

Please see our FAQ for further information.

Walkthrough

Documentation expanded for the build-agent with a comprehensive "Model Tier Selection: Evidence-Based Routing" section. Guidance covers transitioning from static rules to pattern-data-driven model tier decisions, including criteria for trust, decision matrices, and outcome recording procedures.

Changes

Cohort / File(s) Summary
Build Agent Documentation
.agents/tools/build-agent/build-agent.md
Added extensive evidence-based model tier selection guidance (+76 lines), covering static rule mappings, pattern data validation criteria (>75% success, 3+ samples), decision matrices, outcome recording procedures, frontmatter examples, and rationale for empirical routing approaches.

Estimated code review effort

🎯 1 (Trivial) | ⏱️ ~5 minutes

Possibly related issues

Possibly related PRs

Poem

📊 Rules were static, now they learn,
Pattern whispers make insight turn,
Evidence gathered, tested, true,
The agent picks the tier for you! ✨

🚥 Pre-merge checks | ✅ 3
✅ Passed checks (3 passed)
Check name Status Explanation
Description Check ✅ Passed Check skipped - CodeRabbit’s high-level summary is enabled.
Title check ✅ Passed The title accurately describes the main change: updating build-agent documentation to reference pattern data for model tier recommendations, which aligns perfectly with the PR objectives and file changes.
Docstring Coverage ✅ Passed No functions found in the changed files to evaluate docstring coverage. Skipping docstring coverage check.

✏️ Tip: You can configure your own custom pre-merge checks in the settings.

✨ Finishing Touches
🧪 Generate unit tests (beta)
  • Create PR with unit tests
  • Post copyable unit tests in a comment
  • Commit unit tests in branch feature/t1094.1

Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out.

❤️ Share

Comment @coderabbitai help to get the list of available commands and usage tips.

@github-actions
Copy link

🔍 Code Quality Report

�[0;35m[MONITOR]�[0m Code Review Monitoring Report

�[0;34m[INFO]�[0m Latest Quality Status:
SonarCloud: 0 bugs, 0 vulnerabilities, 22 code smells

�[0;34m[INFO]�[0m Recent monitoring activity:
Wed Feb 18 02:52:24 UTC 2026: Code review monitoring started
Wed Feb 18 02:52:24 UTC 2026: SonarCloud - Bugs: 0, Vulnerabilities: 0, Code Smells: 22

📈 Current Quality Metrics

  • BUGS: 0
  • CODE SMELLS: 22
  • VULNERABILITIES: 0

Generated on: Wed Feb 18 02:52:26 UTC 2026


Generated by AI DevOps Framework Code Review Monitoring

Copy link

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request significantly improves the build-agent.md documentation by introducing an evidence-based approach for model tier selection, moving away from purely static rules. The new sections are well-written and provide helpful examples using pattern-tracker-helper.sh. My review includes a couple of suggestions to enhance clarity and consistency within the new documentation, specifically regarding the format of task-type arguments and the wording in the decision table. These changes will help ensure users can apply the new data-driven methodology correctly.

coderabbitai[bot]
coderabbitai bot previously requested changes Feb 18, 2026
Copy link
Contributor

@coderabbitai coderabbitai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actionable comments posted: 2

🤖 Prompt for all review comments with AI agents
Verify each finding against the current code and only fix it if needed.

Inline comments:
In @.agents/tools/build-agent/build-agent.md:
- Around line 425-437: The --task-type value is inconsistent between the suggest
call ("shell script agent") and the record call ("shell-script-agent"), which
can break lookups; update usages so the same canonical format is used everywhere
(pick one form, e.g., hyphenated) or ensure normalization inside
pattern-tracker-helper.sh; specifically change the call site that invokes
suggest to use the same --task-type token as record (or add normalization logic
in pattern-tracker-helper.sh to canonicalize --task-type inputs), and verify
places referencing pattern-tracker-helper.sh, suggest, and record all use the
chosen canonical token.
- Around line 441-450: The frontmatter uses a bare model alias "model: sonnet"
which violates OpenCode's required provider/model-id format; update the YAML
frontmatter key `model` to a full model identifier (for example
`anthropic/claude-sonnet-4-20250514`) or to whatever canonical provider/model-id
your project uses so the resolver can find the model (edit the `model:` line in
the block shown to replace `sonnet` with the full provider/model-id).

@marcusquinn marcusquinn dismissed coderabbitai[bot]’s stale review February 18, 2026 03:05

Auto-dismissed: bot review does not block autonomous pipeline

@github-actions
Copy link

🔍 Code Quality Report

�[0;35m[MONITOR]�[0m Code Review Monitoring Report

�[0;34m[INFO]�[0m Latest Quality Status:
SonarCloud: 0 bugs, 0 vulnerabilities, 22 code smells

�[0;34m[INFO]�[0m Recent monitoring activity:
Wed Feb 18 03:15:06 UTC 2026: Code review monitoring started
Wed Feb 18 03:15:07 UTC 2026: SonarCloud - Bugs: 0, Vulnerabilities: 0, Code Smells: 22

📈 Current Quality Metrics

  • BUGS: 0
  • CODE SMELLS: 22
  • VULNERABILITIES: 0

Generated on: Wed Feb 18 03:15:09 UTC 2026


Generated by AI DevOps Framework Code Review Monitoring

@sonarqubecloud
Copy link

@marcusquinn marcusquinn merged commit 6d9498b into main Feb 18, 2026
27 checks passed
@marcusquinn marcusquinn deleted the feature/t1094.1 branch February 18, 2026 03:28
marcusquinn added a commit that referenced this pull request Feb 18, 2026
…d t1094.1 (t1107)

All t1094 deliverables confirmed implemented:
- Prompt strategy tracking → t1095 (pr:#1629)
- Output quality gradient + failure categorization → t1096 (pr:#1632)
- Token usage → t1095 (pr:#1629)
- A/B comparison → t1098+t1099 (pr:#1637, pr:#1634)
- Prompt-repeat strategy → t1097 (pr:#1631)
- Build-agent reference → t1094.1 (pr:#1633)

t1094 parent ready to be marked complete: verified:2026-02-18
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant