@@ -27,29 +27,9 @@ const openlayer = new Openlayer({
27
27
});
28
28
29
29
async function main () {
30
- const dataStreamResponse = await openlayer .inferencePipelines .data .stream (
31
- ' 182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e' ,
32
- {
33
- config: {
34
- inputVariableNames: [' user_query' ],
35
- outputColumnName: ' output' ,
36
- numOfTokenColumnName: ' tokens' ,
37
- costColumnName: ' cost' ,
38
- timestampColumnName: ' timestamp' ,
39
- },
40
- rows: [
41
- {
42
- user_query: " what's the meaning of life?" ,
43
- output: ' 42' ,
44
- tokens: 7 ,
45
- cost: 0.02 ,
46
- timestamp: 1620000000 ,
47
- },
48
- ],
49
- },
50
- );
51
-
52
- console .log (dataStreamResponse .success );
30
+ const projectCreateResponse = await openlayer .projects .create ({ name: ' My Project' , taskType: ' llm-base' });
31
+
32
+ console .log (projectCreateResponse .id );
53
33
}
54
34
55
35
main ();
@@ -68,26 +48,8 @@ const openlayer = new Openlayer({
68
48
});
69
49
70
50
async function main() {
71
- const params: Openlayer .InferencePipelines .DataStreamParams = {
72
- config: {
73
- inputVariableNames: [' user_query' ],
74
- outputColumnName: ' output' ,
75
- numOfTokenColumnName: ' tokens' ,
76
- costColumnName: ' cost' ,
77
- timestampColumnName: ' timestamp' ,
78
- },
79
- rows: [
80
- {
81
- user_query: " what's the meaning of life?" ,
82
- output: ' 42' ,
83
- tokens: 7 ,
84
- cost: 0.02 ,
85
- timestamp: 1620000000 ,
86
- },
87
- ],
88
- };
89
- const dataStreamResponse: Openlayer .InferencePipelines .DataStreamResponse =
90
- await openlayer .inferencePipelines .data .stream (' 182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e' , params );
51
+ const params: Openlayer .ProjectCreateParams = { name: ' My Project' , taskType: ' llm-base' };
52
+ const projectCreateResponse: Openlayer .ProjectCreateResponse = await openlayer .projects .create (params );
91
53
}
92
54
93
55
main ();
@@ -104,25 +66,8 @@ a subclass of `APIError` will be thrown:
104
66
<!-- prettier-ignore -->
105
67
``` ts
106
68
async function main() {
107
- const dataStreamResponse = await openlayer .inferencePipelines .data
108
- .stream (' 182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e' , {
109
- config: {
110
- inputVariableNames: [' user_query' ],
111
- outputColumnName: ' output' ,
112
- numOfTokenColumnName: ' tokens' ,
113
- costColumnName: ' cost' ,
114
- timestampColumnName: ' timestamp' ,
115
- },
116
- rows: [
117
- {
118
- user_query: " what's the meaning of life?" ,
119
- output: ' 42' ,
120
- tokens: 7 ,
121
- cost: 0.02 ,
122
- timestamp: 1620000000 ,
123
- },
124
- ],
125
- })
69
+ const projectCreateResponse = await openlayer .projects
70
+ .create ({ name: ' My Project' , taskType: ' llm-base' })
126
71
.catch (async (err ) => {
127
72
if (err instanceof Openlayer .APIError ) {
128
73
console .log (err .status ); // 400
@@ -166,7 +111,7 @@ const openlayer = new Openlayer({
166
111
});
167
112
168
113
// Or, configure per-request:
169
- await openlayer .inferencePipelines . data . stream ( ' 182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e ' , { config : { inputVariableNames : [ ' user_query ' ], outputColumnName : ' output ' , numOfTokenColumnName : ' tokens ' , costColumnName : ' cost ' , timestampColumnName : ' timestamp ' }, rows : [{ user_query : ' what \' s the meaning of life? ' , output : ' 42 ' , tokens : 7 , cost : 0.02 , timestamp : 1620000000 }] }, {
114
+ await openlayer .projects . create ({ name : ' My Project ' , taskType : ' llm-base ' }, {
170
115
maxRetries: 5 ,
171
116
});
172
117
```
@@ -183,7 +128,7 @@ const openlayer = new Openlayer({
183
128
});
184
129
185
130
// Override per-request:
186
- await openlayer .inferencePipelines . data . stream ( ' 182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e ' , { config: { inputVariableNames: [ ' user_query ' ], outputColumnName: ' output ' , numOfTokenColumnName : ' tokens ' , costColumnName: ' cost ' , timestampColumnName: ' timestamp ' }, rows: [{ user_query: ' what \' s the meaning of life? ' , output: ' 42 ' , tokens: 7 , cost: 0.02 , timestamp: 1620000000 }] }, {
131
+ await openlayer .projects . create ({ name: ' My Project ' , taskType : ' llm-base ' }, {
187
132
timeout: 5 * 1000 ,
188
133
});
189
134
```
@@ -204,51 +149,15 @@ You can also use the `.withResponse()` method to get the raw `Response` along wi
204
149
``` ts
205
150
const openlayer = new Openlayer ();
206
151
207
- const response = await openlayer .inferencePipelines .data
208
- .stream (' 182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e' , {
209
- config: {
210
- inputVariableNames: [' user_query' ],
211
- outputColumnName: ' output' ,
212
- numOfTokenColumnName: ' tokens' ,
213
- costColumnName: ' cost' ,
214
- timestampColumnName: ' timestamp' ,
215
- },
216
- rows: [
217
- {
218
- user_query: " what's the meaning of life?" ,
219
- output: ' 42' ,
220
- tokens: 7 ,
221
- cost: 0.02 ,
222
- timestamp: 1620000000 ,
223
- },
224
- ],
225
- })
226
- .asResponse ();
152
+ const response = await openlayer .projects .create ({ name: ' My Project' , taskType: ' llm-base' }).asResponse ();
227
153
console .log (response .headers .get (' X-My-Header' ));
228
154
console .log (response .statusText ); // access the underlying Response object
229
155
230
- const { data : dataStreamResponse, response : raw } = await openlayer .inferencePipelines .data
231
- .stream (' 182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e' , {
232
- config: {
233
- inputVariableNames: [' user_query' ],
234
- outputColumnName: ' output' ,
235
- numOfTokenColumnName: ' tokens' ,
236
- costColumnName: ' cost' ,
237
- timestampColumnName: ' timestamp' ,
238
- },
239
- rows: [
240
- {
241
- user_query: " what's the meaning of life?" ,
242
- output: ' 42' ,
243
- tokens: 7 ,
244
- cost: 0.02 ,
245
- timestamp: 1620000000 ,
246
- },
247
- ],
248
- })
156
+ const { data : projectCreateResponse, response : raw } = await openlayer .projects
157
+ .create ({ name: ' My Project' , taskType: ' llm-base' })
249
158
.withResponse ();
250
159
console .log (raw .headers .get (' X-My-Header' ));
251
- console .log (dataStreamResponse . success );
160
+ console .log (projectCreateResponse . id );
252
161
```
253
162
254
163
### Making custom/undocumented requests
@@ -352,26 +261,8 @@ const openlayer = new Openlayer({
352
261
});
353
262
354
263
// Override per-request:
355
- await openlayer .inferencePipelines .data .stream (
356
- ' 182bd5e5-6e1a-4fe4-a799-aa6d9a6ab26e' ,
357
- {
358
- config: {
359
- inputVariableNames: [' user_query' ],
360
- outputColumnName: ' output' ,
361
- numOfTokenColumnName: ' tokens' ,
362
- costColumnName: ' cost' ,
363
- timestampColumnName: ' timestamp' ,
364
- },
365
- rows: [
366
- {
367
- user_query: " what's the meaning of life?" ,
368
- output: ' 42' ,
369
- tokens: 7 ,
370
- cost: 0.02 ,
371
- timestamp: 1620000000 ,
372
- },
373
- ],
374
- },
264
+ await openlayer .projects .create (
265
+ { name: ' My Project' , taskType: ' llm-base' },
375
266
{
376
267
httpAgent: new http .Agent ({ keepAlive: false }),
377
268
},
0 commit comments