Skip to content

Augment FunctionInvokingChatClient's span with token counts #6296

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 1 commit into from
Apr 15, 2025

Conversation

stephentoub
Copy link
Member

@stephentoub stephentoub commented Apr 15, 2025

@lmolkova suggested this would be valuable.

Alternatively, we could consider removing the activity use entirely from FunctionInvokingChatClient, and just tell folks that want such a span to add an OpenTelemetryChatClient before the FICC in the pipeline, even if they also have one after it:

.UseOpenTelemetry()
.UseFunctionInvocation()
.UseOpenTelemetry()

I think it would just naturally do the "right thing". Though if we were to do that, we might want to make the meter optional, so that metrics weren't counted twice.

Microsoft Reviewers: Open in CodeFlow

@stephentoub stephentoub requested a review from a team as a code owner April 15, 2025 03:31
@github-actions github-actions bot added the area-ai Microsoft.Extensions.AI libraries label Apr 15, 2025
@SteveSandersonMS
Copy link
Member

Makes sense. I think counting the usage within FICC is completely fine as an extension of the existing design decision to make FICC aware of telemetry as a concept.

I agree that making FICC unaware of telemetry would be another valid design approach and would arguably be a cleaner separation of concerns, but the drawback of needing two copies of UseOpenTelemetry in the pipeline (to provide spans representing both aggregate and per-iteration views of function invocation) would be quite bad for developer experience as very few people would understand why.

In general it makes sense that sometimes middleware should emit its own telemetry data since it might be doing something custom. Given that general rule I'm perfectly happy with FICC doing so, and it simplifies the experience so seems like a good tradeoff.

@stephentoub stephentoub merged commit 386c634 into dotnet:main Apr 15, 2025
7 checks passed
@stephentoub stephentoub deleted the funcspan branch April 15, 2025 11:45
jeffhandley added a commit that referenced this pull request Apr 30, 2025
* Merged PR 49569: Cherry pick Evaluation changes for 9.4.3 release

* Update package-lock.json for Evaluations

* Reset package-lock.json to same as main branch

* Merged PR 49585: [9.4.3] [cherry pick] A couple of minor fixes

A couple of minor fixes

1. Adds back tool tip for metrics which are useful in cases where the metric name is long and does not fit entirely in the card
2. Makes the size of the selection buttons on the left of cards consistent with the size of status icons on the right to make the alignment cleaner
3. Implement IEquatable for UrlCacheKey (per Copilot's suggestion)
4. Throw ArgumentException when the number of messages passed to ContentSafetyChatClient does not match expectations (also per Copilot's suggestion)

* Add test for optional parameters being required with RequireAllProperties (#6265)

* Add test for optional parameters being required with RequireAllProperties

* Test both requireAllProperties values

* Use AssertDeepEquals that logs the difference to other tests

* Adding reference to an unsupported built-in tool on OpenAI Chat API no longer throws (#6276)

* Update M.E.AI changelogs (#6269)

* Augment UseDistributedCache XML docs (#6256)

* Augment AIFunctionFactory.Create XML docs (#6255)

* Rename EmbeddingGeneratorExtensions.GenerateEmbedding extension methods (#6295)

To align with the the base method on IEmbeddingGenerator

* Augment FunctionInvokingChatClient's span with token counts (#6296)

* Rename ChatThreadId to ConversationId (#6300)

* In OpenAI responses client, use response ID as ChatThreadId

* Rename ChatThreadId -> ConversationId

* Related renames

* Restore deleted members as obsolete (#6304)

* Restore EmbeddingGeneratorExtensions members as obsolete

* Restore ChatThreadId as obsolete

* Restore ChatResponse.ChatThreadId and ChatResponseUpdate.ChatThreadId as obsolete.

* Remove the tests for obsolete members

* Remove the Embeddings tests for obsolete members

* Support [FromKeyedServices] in AIFunctionFactory (#6310)

* Utilize IServiceProviderIsService in AIFunctionFactory (#6317)

Add AIFunctionFactoryOptions.Services, and use it when examining function parameters to determine whether they should be resolved by default from DI.

* Remove AsChatClient/AsEmbeddingGenerator that were obsoleted in 9.4.0-preview.1.25207.5 (#6327)

* Add ChatOptions.AllowMultipleToolCalls (#6326)

* Add ChatOptions.AllowMultipleToolCalls

* Use it in OpenAI adapter

* Disable default required property schema generation and OpenAI strict mode. (#6285)

* Disable default required property schema generation and OpenAI strict mode.

* Default strictness to OpenAI client defaults.

* Fix a few failing tests.

* Undo a number of additional changes made by #6064

* Revert change to structured output defaults.

* Incorporate more test fixes.

* Address feedback.

* Tweak strict mode signature, fix failing tests.

* Roll back schemaIsStrict flag from ChatResponseFormat.

* Merged PR 49624: [9.4.3] [cherry-pick] Update readmes (#6345)

Includes name of new metric that was added to Quality package.

Also updates some doc comments.

* Bump version to 9.4.3

* Merged PR 49636: [9.4.3] [cherry-pick] Skip messages that have no text when rendering conversations as part of evaluation prompts (#6349)

* Add test for optional parameters being required with RequireAllProperties (#6265)

* Add test for optional parameters being required with RequireAllProperties

* Test both requireAllProperties values

* Use AssertDeepEquals that logs the difference to other tests

* Adding reference to an unsupported built-in tool on OpenAI Chat API no longer throws (#6276)

* Update M.E.AI changelogs (#6269)

* Augment UseDistributedCache XML docs (#6256)

* Augment AIFunctionFactory.Create XML docs (#6255)

* Rename EmbeddingGeneratorExtensions.GenerateEmbedding extension methods (#6295)

To align with the the base method on IEmbeddingGenerator

* Augment FunctionInvokingChatClient's span with token counts (#6296)

* Rename ChatThreadId to ConversationId (#6300)

* In OpenAI responses client, use response ID as ChatThreadId

* Rename ChatThreadId -> ConversationId

* Related renames

* Restore deleted members as obsolete (#6304)

* Restore EmbeddingGeneratorExtensions members as obsolete

* Restore ChatThreadId as obsolete

* Restore ChatResponse.ChatThreadId and ChatResponseUpdate.ChatThreadId as obsolete.

* Remove the tests for obsolete members

* Remove the Embeddings tests for obsolete members

* Support [FromKeyedServices] in AIFunctionFactory (#6310)

* Utilize IServiceProviderIsService in AIFunctionFactory (#6317)

Add AIFunctionFactoryOptions.Services, and use it when examining function parameters to determine whether they should be resolved by default from DI.

* Remove AsChatClient/AsEmbeddingGenerator that were obsoleted in 9.4.0-preview.1.25207.5 (#6327)

* Add ChatOptions.AllowMultipleToolCalls (#6326)

* Add ChatOptions.AllowMultipleToolCalls

* Use it in OpenAI adapter

* Disable default required property schema generation and OpenAI strict mode. (#6285)

* Disable default required property schema generation and OpenAI strict mode.

* Default strictness to OpenAI client defaults.

* Fix a few failing tests.

* Undo a number of additional changes made by #6064

* Revert change to structured output defaults.

* Incorporate more test fixes.

* Address feedback.

* Tweak strict mode signature, fix failing tests.

* Roll back schemaIsStrict flag from ChatResponseFormat.

* Bump version to 9.4.3

* Merged PR 49636: [9.4.3] [cherry-pick] Skip messages that have no text when rendering conversations as part of evaluation prompts (#6349)

* Update chat template dependencies, fix OpenAI/Aspire config, and address build warnings (#6280)

* Update chat template dependencies

* Update test snapshots

* Update Aspire version

* Revert Aspire + update CommunityToolkit.Aspire

* Open README in VS after project creation

* Add Known Issue to Aspire README for Qdrant

* Update survey template URL

* Set the Project Template package version to -preview.2

* Update template baseline

* Update template pinned versions

* Do not append template args to snapshot names

* Fix vector store index in the README. Add an AzureAISearch template test.

* Add a note to the Aspire README for trusting the development certificate

* Use AddOpenAIClient for OpenAI and AddAzureOpenAIClient for Azure OpenAI

* Remove duplicated using System.ClientModel

* Update Aspire README to specify the exception thrown for the known issue

* Augment the Aspire README for more Docker notes for Ollama and Qdrant

* Fix Microsoft.Extensions.Http.Resilience warnings with separate pinned versions

* Update test baseline for: Rename EmbeddingGeneratorExtensions.GenerateEmbedding extension methods (#6295)

* Add --managed-identity to the template developer README

* Apply the Ollama timeout recommendation in the template code

* Call http.RemoveAllResilienceHandlers before adding the handler

* Update template test baseline

---------

Co-authored-by: Jeff Handley <jeffhandley@users.noreply.github.com>
Co-authored-by: David Cantu <dacantu@microsoft.com>

* Expose AIContent constructor (#6346)

* Add PDF support to OpenAI AsIChatClient (#6344)

* Add PDF support to OpenAI AsIChatClient

* Add missing reference

* Make CreateJsonSchema tolerate JSO inputs that don't have a resolver set. (#6348)

* Update MEAI.Templates to use the just-built version of the libraries

* Enhance Function Invocation Extensibility for Microsoft.Extensions.AI (#6325)

---------

Co-authored-by: Peter Waldschmidt <Peter.Waldschmidt@microsoft.com>
Co-authored-by: Peter Waldschmidt <pewaldsc@microsoft.com>
Co-authored-by: Shyam Namboodiripad <Shyam.Namboodiripad@microsoft.com>
Co-authored-by: David Cantú <dacantu@microsoft.com>
Co-authored-by: Art Leonard <artl93@users.noreply.github.com>
Co-authored-by: Stephen Toub <stoub@microsoft.com>
Co-authored-by: Shay Rojansky <roji@roji.org>
Co-authored-by: Steve Sanderson <SteveSandersonMS@users.noreply.github.com>
Co-authored-by: Eirik Tsarpalis <eirik.tsarpalis@gmail.com>
Co-authored-by: Mackinnon Buck <mackinnon.buck@gmail.com>
Co-authored-by: Roger Barreto <19890735+RogerBarreto@users.noreply.github.com>
@github-actions github-actions bot locked and limited conversation to collaborators May 16, 2025
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
area-ai Microsoft.Extensions.AI libraries
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants