@@ -12,20 +12,18 @@ describe('ai', () => {
12
12
spans : expect . arrayContaining ( [
13
13
expect . objectContaining ( {
14
14
data : expect . objectContaining ( {
15
- 'ai.completion_tokens.used' : 20 ,
16
15
'ai.model.id' : 'mock-model-id' ,
17
16
'ai.model.provider' : 'mock-provider' ,
18
- 'ai.model_id' : 'mock-model-id' ,
19
17
'ai.operationId' : 'ai.generateText' ,
20
18
'ai.pipeline.name' : 'generateText' ,
21
- 'ai.prompt_tokens.used' : 10 ,
22
19
'ai.response.finishReason' : 'stop' ,
23
20
'ai.settings.maxRetries' : 2 ,
24
21
'ai.settings.maxSteps' : 1 ,
25
22
'ai.streaming' : false ,
26
- 'ai.total_tokens.used' : 30 ,
27
- 'ai.usage.completionTokens' : 20 ,
28
- 'ai.usage.promptTokens' : 10 ,
23
+ 'gen_ai.response.model' : 'mock-model-id' ,
24
+ 'gen_ai.usage.input_tokens' : 10 ,
25
+ 'gen_ai.usage.output_tokens' : 20 ,
26
+ 'gen_ai.usage.total_tokens' : 30 ,
29
27
'operation.name' : 'ai.generateText' ,
30
28
'sentry.op' : 'ai.pipeline.generateText' ,
31
29
'sentry.origin' : 'auto.vercelai.otel' ,
@@ -47,18 +45,17 @@ describe('ai', () => {
47
45
'gen_ai.system' : 'mock-provider' ,
48
46
'gen_ai.request.model' : 'mock-model-id' ,
49
47
'ai.pipeline.name' : 'generateText.doGenerate' ,
50
- 'ai.model_id' : 'mock-model-id' ,
51
48
'ai.streaming' : false ,
52
49
'ai.response.finishReason' : 'stop' ,
53
50
'ai.response.model' : 'mock-model-id' ,
54
- 'ai.usage.promptTokens ' : 10 ,
55
- 'ai.usage.completionTokens ' : 20 ,
51
+ 'ai.response.id ' : expect . any ( String ) ,
52
+ 'ai.response.timestamp ' : expect . any ( String ) ,
56
53
'gen_ai.response.finish_reasons' : [ 'stop' ] ,
57
54
'gen_ai.usage.input_tokens' : 10 ,
58
55
'gen_ai.usage.output_tokens' : 20 ,
59
- 'ai.completion_tokens.used ' : 20 ,
60
- 'ai.prompt_tokens.used ' : 10 ,
61
- 'ai.total_tokens.used ' : 30 ,
56
+ 'gen_ai.response.id ' : expect . any ( String ) ,
57
+ 'gen_ai.response.model ' : 'mock-model-id' ,
58
+ 'gen_ai.usage.total_tokens ' : 30 ,
62
59
} ) ,
63
60
description : 'generateText.doGenerate' ,
64
61
op : 'ai.run.doGenerate' ,
@@ -67,22 +64,21 @@ describe('ai', () => {
67
64
} ) ,
68
65
expect . objectContaining ( {
69
66
data : expect . objectContaining ( {
70
- 'ai.completion_tokens.used' : 20 ,
71
67
'ai.model.id' : 'mock-model-id' ,
72
68
'ai.model.provider' : 'mock-provider' ,
73
- 'ai.model_id' : 'mock-model-id' ,
74
- 'ai.prompt' : '{"prompt":"Where is the second span?"}' ,
75
69
'ai.operationId' : 'ai.generateText' ,
76
70
'ai.pipeline.name' : 'generateText' ,
77
- 'ai.prompt_tokens.used ' : 10 ,
71
+ 'ai.prompt ' : '{"prompt":"Where is the second span?"}' ,
78
72
'ai.response.finishReason' : 'stop' ,
79
- 'ai.input_messages ' : '{"prompt":"Where is the second span?"}' ,
73
+ 'ai.response.text ' : expect . any ( String ) ,
80
74
'ai.settings.maxRetries' : 2 ,
81
75
'ai.settings.maxSteps' : 1 ,
82
76
'ai.streaming' : false ,
83
- 'ai.total_tokens.used' : 30 ,
84
- 'ai.usage.completionTokens' : 20 ,
85
- 'ai.usage.promptTokens' : 10 ,
77
+ 'gen_ai.prompt' : '{"prompt":"Where is the second span?"}' ,
78
+ 'gen_ai.response.model' : 'mock-model-id' ,
79
+ 'gen_ai.usage.input_tokens' : 10 ,
80
+ 'gen_ai.usage.output_tokens' : 20 ,
81
+ 'gen_ai.usage.total_tokens' : 30 ,
86
82
'operation.name' : 'ai.generateText' ,
87
83
'sentry.op' : 'ai.pipeline.generateText' ,
88
84
'sentry.origin' : 'auto.vercelai.otel' ,
@@ -104,18 +100,20 @@ describe('ai', () => {
104
100
'gen_ai.system' : 'mock-provider' ,
105
101
'gen_ai.request.model' : 'mock-model-id' ,
106
102
'ai.pipeline.name' : 'generateText.doGenerate' ,
107
- 'ai.model_id' : 'mock-model-id' ,
108
103
'ai.streaming' : false ,
109
104
'ai.response.finishReason' : 'stop' ,
110
105
'ai.response.model' : 'mock-model-id' ,
111
- 'ai.usage.promptTokens' : 10 ,
112
- 'ai.usage.completionTokens' : 20 ,
106
+ 'ai.response.id' : expect . any ( String ) ,
107
+ 'ai.response.text' : expect . any ( String ) ,
108
+ 'ai.response.timestamp' : expect . any ( String ) ,
109
+ 'ai.prompt.format' : expect . any ( String ) ,
110
+ 'ai.prompt.messages' : expect . any ( String ) ,
113
111
'gen_ai.response.finish_reasons' : [ 'stop' ] ,
114
112
'gen_ai.usage.input_tokens' : 10 ,
115
113
'gen_ai.usage.output_tokens' : 20 ,
116
- 'ai.completion_tokens.used ' : 20 ,
117
- 'ai.prompt_tokens.used ' : 10 ,
118
- 'ai.total_tokens.used ' : 30 ,
114
+ 'gen_ai.response.id ' : expect . any ( String ) ,
115
+ 'gen_ai.response.model ' : 'mock-model-id' ,
116
+ 'gen_ai.usage.total_tokens ' : 30 ,
119
117
} ) ,
120
118
description : 'generateText.doGenerate' ,
121
119
op : 'ai.run.doGenerate' ,
0 commit comments