@@ -12,20 +12,18 @@ describe('ai', () => {
12
12
spans : expect . arrayContaining ( [
13
13
expect . objectContaining ( {
14
14
data : expect . objectContaining ( {
15
- 'ai.completion_tokens.used ' : 20 ,
15
+ 'gen_ai.usage.output_tokens ' : 20 ,
16
16
'ai.model.id' : 'mock-model-id' ,
17
17
'ai.model.provider' : 'mock-provider' ,
18
18
'ai.model_id' : 'mock-model-id' ,
19
19
'ai.operationId' : 'ai.generateText' ,
20
20
'ai.pipeline.name' : 'generateText' ,
21
- 'ai.prompt_tokens.used ' : 10 ,
21
+ 'gen_ai.usage.input_tokens ' : 10 ,
22
22
'ai.response.finishReason' : 'stop' ,
23
23
'ai.settings.maxRetries' : 2 ,
24
24
'ai.settings.maxSteps' : 1 ,
25
25
'ai.streaming' : false ,
26
- 'ai.total_tokens.used' : 30 ,
27
- 'ai.usage.completionTokens' : 20 ,
28
- 'ai.usage.promptTokens' : 10 ,
26
+ 'gen_ai.usage.total_tokens' : 30 ,
29
27
'operation.name' : 'ai.generateText' ,
30
28
'sentry.op' : 'ai.pipeline.generateText' ,
31
29
'sentry.origin' : 'auto.vercelai.otel' ,
@@ -51,14 +49,10 @@ describe('ai', () => {
51
49
'ai.streaming' : false ,
52
50
'ai.response.finishReason' : 'stop' ,
53
51
'ai.response.model' : 'mock-model-id' ,
54
- 'ai.usage.promptTokens' : 10 ,
55
- 'ai.usage.completionTokens' : 20 ,
56
52
'gen_ai.response.finish_reasons' : [ 'stop' ] ,
57
53
'gen_ai.usage.input_tokens' : 10 ,
58
54
'gen_ai.usage.output_tokens' : 20 ,
59
- 'ai.completion_tokens.used' : 20 ,
60
- 'ai.prompt_tokens.used' : 10 ,
61
- 'ai.total_tokens.used' : 30 ,
55
+ 'gen_ai.usage.total_tokens' : 30 ,
62
56
} ) ,
63
57
description : 'generateText.doGenerate' ,
64
58
op : 'ai.run.doGenerate' ,
@@ -67,25 +61,23 @@ describe('ai', () => {
67
61
} ) ,
68
62
expect . objectContaining ( {
69
63
data : expect . objectContaining ( {
70
- 'ai.completion_tokens.used' : 20 ,
71
64
'ai.model.id' : 'mock-model-id' ,
72
65
'ai.model.provider' : 'mock-provider' ,
73
66
'ai.model_id' : 'mock-model-id' ,
74
67
'ai.prompt' : '{"prompt":"Where is the second span?"}' ,
75
68
'ai.operationId' : 'ai.generateText' ,
76
69
'ai.pipeline.name' : 'generateText' ,
77
- 'ai.prompt_tokens.used' : 10 ,
78
70
'ai.response.finishReason' : 'stop' ,
79
71
'ai.input_messages' : '{"prompt":"Where is the second span?"}' ,
80
72
'ai.settings.maxRetries' : 2 ,
81
73
'ai.settings.maxSteps' : 1 ,
82
74
'ai.streaming' : false ,
83
- 'ai.total_tokens.used' : 30 ,
84
- 'ai.usage.completionTokens' : 20 ,
85
- 'ai.usage.promptTokens' : 10 ,
86
75
'operation.name' : 'ai.generateText' ,
87
76
'sentry.op' : 'ai.pipeline.generateText' ,
88
77
'sentry.origin' : 'auto.vercelai.otel' ,
78
+ 'gen_ai.usage.input_tokens' : 10 ,
79
+ 'gen_ai.usage.output_tokens' : 20 ,
80
+ 'gen_ai.usage.total_tokens' : 30 ,
89
81
} ) ,
90
82
description : 'generateText' ,
91
83
op : 'ai.pipeline.generateText' ,
@@ -108,14 +100,10 @@ describe('ai', () => {
108
100
'ai.streaming' : false ,
109
101
'ai.response.finishReason' : 'stop' ,
110
102
'ai.response.model' : 'mock-model-id' ,
111
- 'ai.usage.promptTokens' : 10 ,
112
- 'ai.usage.completionTokens' : 20 ,
113
103
'gen_ai.response.finish_reasons' : [ 'stop' ] ,
114
104
'gen_ai.usage.input_tokens' : 10 ,
115
105
'gen_ai.usage.output_tokens' : 20 ,
116
- 'ai.completion_tokens.used' : 20 ,
117
- 'ai.prompt_tokens.used' : 10 ,
118
- 'ai.total_tokens.used' : 30 ,
106
+ 'gen_ai.usage.total_tokens' : 30 ,
119
107
} ) ,
120
108
description : 'generateText.doGenerate' ,
121
109
op : 'ai.run.doGenerate' ,
0 commit comments