@@ -47,10 +47,9 @@ export class Prompts extends APIResource {
47
47
}
48
48
49
49
/**
50
- * Fetches the configured model parameters and messages rendered with the provided
51
- * variables mapped to the set LLM provider. This endpoint abstracts the need to
52
- * handle mapping between different providers, while still allowing direct calls to
53
- * the providers.
50
+ * Fetches the model configuration parameters for a specified prompt, including
51
+ * penalty settings, response format, and the model messages rendered with the
52
+ * given variables mapped to the set LLM provider.
54
53
*/
55
54
getParameters (
56
55
id : string ,
@@ -539,9 +538,14 @@ export namespace PromptConfiguration {
539
538
maxTokens : number | null ;
540
539
541
540
/**
542
- * The name of the model for the provider.
541
+ * Example: "gpt-3.5-turbo"
543
542
*/
544
- name : string ;
543
+ modelName : string ;
544
+
545
+ /**
546
+ * The provider of the provided model.
547
+ */
548
+ modelProvider : 'ANTHROPIC' | 'OPENAI' ;
545
549
546
550
parallelToolCalls : boolean ;
547
551
@@ -550,11 +554,6 @@ export namespace PromptConfiguration {
550
554
*/
551
555
presencePenalty : number ;
552
556
553
- /**
554
- * The LLM model provider.
555
- */
556
- provider : 'ANTHROPIC' | 'OPENAI' ;
557
-
558
557
/**
559
558
* Example: PromptResponseFormat.TEXT
560
559
*/
@@ -717,9 +716,14 @@ export namespace PromptCreateParams {
717
716
maxTokens : number | null ;
718
717
719
718
/**
720
- * The name of the model for the provider.
719
+ * Example: "gpt-3.5-turbo"
721
720
*/
722
- name : string ;
721
+ modelName : string ;
722
+
723
+ /**
724
+ * The provider of the provided model.
725
+ */
726
+ modelProvider : 'ANTHROPIC' | 'OPENAI' ;
723
727
724
728
parallelToolCalls : boolean ;
725
729
@@ -728,11 +732,6 @@ export namespace PromptCreateParams {
728
732
*/
729
733
presencePenalty : number ;
730
734
731
- /**
732
- * The LLM model provider.
733
- */
734
- provider : 'ANTHROPIC' | 'OPENAI' ;
735
-
736
735
/**
737
736
* Example: PromptResponseFormat.TEXT
738
737
*/
@@ -867,9 +866,14 @@ export namespace PromptUpdateParams {
867
866
maxTokens : number | null ;
868
867
869
868
/**
870
- * The name of the model for the provider.
869
+ * Example: "gpt-3.5-turbo"
871
870
*/
872
- name : string ;
871
+ modelName : string ;
872
+
873
+ /**
874
+ * The provider of the provided model.
875
+ */
876
+ modelProvider : 'ANTHROPIC' | 'OPENAI' ;
873
877
874
878
parallelToolCalls : boolean ;
875
879
@@ -878,11 +882,6 @@ export namespace PromptUpdateParams {
878
882
*/
879
883
presencePenalty : number ;
880
884
881
- /**
882
- * The LLM model provider.
883
- */
884
- provider : 'ANTHROPIC' | 'OPENAI' ;
885
-
886
885
/**
887
886
* Example: PromptResponseFormat.TEXT
888
887
*/
0 commit comments