Skip to content

Bug fix: Fix the Git Service when running on Ollama#219

Open
esnible wants to merge 2 commits intokagenti:mainfrom
esnible:git-service-on-ollama
Open

Bug fix: Fix the Git Service when running on Ollama#219
esnible wants to merge 2 commits intokagenti:mainfrom
esnible:git-service-on-ollama

Conversation

@esnible
Copy link
Copy Markdown
Contributor

@esnible esnible commented Apr 3, 2026

Summary

We supply a configuration for the Git Service when running on Kagenti under Ollama, but it doesn't work.

For example, the chat query What do you think about https://github.com/kagenti/agent-examples/issues/218 ? yields

Thought: The user query involves retrieving issues from a specific GitHub repository and issue number. Action: list_issues Action Input: {"owner":"kagenti","repo":"agent-examples","state":"all"}

The default port for an MCP tool in Kagenti is 8000, but this Agent example assumes the MCP server will be deployed on 9090.

(Optional) Testing Instructions

  1. Deploy the Git Tool
  2. http://kagenti-ui.localtest.me:8080/agents/import

Signed-off-by: Ed Snible <snible@us.ibm.com>
Copy link
Copy Markdown
Contributor

@rubambiza rubambiza left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Review Summary

The MCP port fix (9090 -> 8000) is correct and matches Kagenti's default tool port. The /v1 URL suffix and model ID change look intentional for OpenAI-compatible mode via litellm. One issue: the file header comment is now stale.

Areas reviewed: Config/env
Commits: 1 commit, signed-off
CI status: All passing (10/10)


# LLM configuration
TASK_MODEL_ID=ollama_chat/ibm/granite4:latest
TASK_MODEL_ID=gpt-oss:latest
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

must-fix: The header comment on line 5 still says ollama pull ibm/granite4:latest but the model ID is now gpt-oss:latest. Please update the comment to match the new model, or explain what gpt-oss is and what prerequisite pull command is needed.

# For in-cluster Ollama: http://ollama.ollama.svc:11434/v1
LLM_API_BASE=http://host.docker.internal:11434/v1
OLLAMA_API_BASE=http://host.docker.internal:11434/v1
LLM_API_KEY=ollama
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

suggestion: Adding /v1 switches from the native Ollama API to the OpenAI-compatible endpoint. This is correct for litellm with the gpt-oss model prefix, but a brief inline comment explaining why /v1 is needed would help future readers.

Copy link
Copy Markdown
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks, please re-review.

Signed-off-by: Ed Snible <esnible@acm.org>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants