Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
7 changes: 0 additions & 7 deletions src/libs/Cohere/Generated/Cohere.CohereClient.Chat.g.cs
Original file line number Diff line number Diff line change
Expand Up @@ -740,11 +740,6 @@ partial void ProcessChatResponseContent(
/// - AUTO: Cohere Platform Only<br/>
/// - AUTO_PRESERVE_ORDER: Azure, AWS Sagemaker/Bedrock, Private Deployments
/// </param>
/// <param name="rawPrompting">
/// When enabled, the user's prompt will be sent to the model without<br/>
/// any pre-processing.<br/>
/// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
/// </param>
/// <param name="responseFormat">
/// Configuration for forcing the model output to adhere to the specified format. Supported on [Command R 03-2024](https://docs.cohere.com/docs/command-r), [Command R+ 04-2024](https://docs.cohere.com/docs/command-r-plus) and newer models.<br/>
/// The model can be forced into outputting JSON objects (with up to 5 levels of nesting) by setting `{ "type": "json_object" }`.<br/>
Expand Down Expand Up @@ -839,7 +834,6 @@ partial void ProcessChatResponseContent(
string? model = default,
string? preamble = default,
global::Cohere.ChatRequestPromptTruncation? promptTruncation = default,
bool? rawPrompting = default,
global::Cohere.ResponseFormat? responseFormat = default,
global::Cohere.ChatRequestSafetyMode? safetyMode = default,
bool? searchQueriesOnly = default,
Expand Down Expand Up @@ -868,7 +862,6 @@ partial void ProcessChatResponseContent(
Preamble = preamble,
PresencePenalty = presencePenalty,
PromptTruncation = promptTruncation,
RawPrompting = rawPrompting,
ResponseFormat = responseFormat,
SafetyMode = safetyMode,
SearchQueriesOnly = searchQueriesOnly,
Expand Down
7 changes: 0 additions & 7 deletions src/libs/Cohere/Generated/Cohere.CohereClient.Chatv2.g.cs
Original file line number Diff line number Diff line change
Expand Up @@ -667,11 +667,6 @@ partial void ProcessChatv2ResponseContent(
/// Defaults to `0.0`, min value of `0.0`, max value of `1.0`.<br/>
/// Used to reduce repetitiveness of generated tokens. Similar to `frequency_penalty`, except that this penalty is applied equally to all tokens that have already appeared, regardless of their exact frequencies.
/// </param>
/// <param name="rawPrompting">
/// When enabled, the user's prompt will be sent to the model without<br/>
/// any pre-processing.<br/>
/// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
/// </param>
/// <param name="reasoningEffort">
/// The reasoning effort level of the model. This affects the model's performance and the time it takes to generate a response.
/// </param>
Expand Down Expand Up @@ -738,7 +733,6 @@ partial void ProcessChatv2ResponseContent(
int? maxTokens = default,
float? p = default,
float? presencePenalty = default,
bool? rawPrompting = default,
global::Cohere.ReasoningEffort? reasoningEffort = default,
global::Cohere.ResponseFormatV2? responseFormat = default,
global::Cohere.Chatv2RequestSafetyMode? safetyMode = default,
Expand All @@ -763,7 +757,6 @@ partial void ProcessChatv2ResponseContent(
Model = model,
P = p,
PresencePenalty = presencePenalty,
RawPrompting = rawPrompting,
ReasoningEffort = reasoningEffort,
ResponseFormat = responseFormat,
SafetyMode = safetyMode,
Expand Down
6 changes: 0 additions & 6 deletions src/libs/Cohere/Generated/Cohere.ICohereClient.Chat.g.cs
Original file line number Diff line number Diff line change
Expand Up @@ -127,11 +127,6 @@ public partial interface ICohereClient
/// - AUTO: Cohere Platform Only<br/>
/// - AUTO_PRESERVE_ORDER: Azure, AWS Sagemaker/Bedrock, Private Deployments
/// </param>
/// <param name="rawPrompting">
/// When enabled, the user's prompt will be sent to the model without<br/>
/// any pre-processing.<br/>
/// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
/// </param>
/// <param name="responseFormat">
/// Configuration for forcing the model output to adhere to the specified format. Supported on [Command R 03-2024](https://docs.cohere.com/docs/command-r), [Command R+ 04-2024](https://docs.cohere.com/docs/command-r-plus) and newer models.<br/>
/// The model can be forced into outputting JSON objects (with up to 5 levels of nesting) by setting `{ "type": "json_object" }`.<br/>
Expand Down Expand Up @@ -226,7 +221,6 @@ public partial interface ICohereClient
string? model = default,
string? preamble = default,
global::Cohere.ChatRequestPromptTruncation? promptTruncation = default,
bool? rawPrompting = default,
global::Cohere.ResponseFormat? responseFormat = default,
global::Cohere.ChatRequestSafetyMode? safetyMode = default,
bool? searchQueriesOnly = default,
Expand Down
6 changes: 0 additions & 6 deletions src/libs/Cohere/Generated/Cohere.ICohereClient.Chatv2.g.cs
Original file line number Diff line number Diff line change
Expand Up @@ -62,11 +62,6 @@ public partial interface ICohereClient
/// Defaults to `0.0`, min value of `0.0`, max value of `1.0`.<br/>
/// Used to reduce repetitiveness of generated tokens. Similar to `frequency_penalty`, except that this penalty is applied equally to all tokens that have already appeared, regardless of their exact frequencies.
/// </param>
/// <param name="rawPrompting">
/// When enabled, the user's prompt will be sent to the model without<br/>
/// any pre-processing.<br/>
/// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
/// </param>
/// <param name="reasoningEffort">
/// The reasoning effort level of the model. This affects the model's performance and the time it takes to generate a response.
/// </param>
Expand Down Expand Up @@ -133,7 +128,6 @@ public partial interface ICohereClient
int? maxTokens = default,
float? p = default,
float? presencePenalty = default,
bool? rawPrompting = default,
global::Cohere.ReasoningEffort? reasoningEffort = default,
global::Cohere.ResponseFormatV2? responseFormat = default,
global::Cohere.Chatv2RequestSafetyMode? safetyMode = default,
Expand Down
15 changes: 0 additions & 15 deletions src/libs/Cohere/Generated/Cohere.Models.ChatRequest.g.cs
Original file line number Diff line number Diff line change
Expand Up @@ -159,14 +159,6 @@ public sealed partial class ChatRequest
[global::System.Text.Json.Serialization.JsonConverter(typeof(global::Cohere.JsonConverters.ChatRequestPromptTruncationJsonConverter))]
public global::Cohere.ChatRequestPromptTruncation? PromptTruncation { get; set; }

/// <summary>
/// When enabled, the user's prompt will be sent to the model without<br/>
/// any pre-processing.<br/>
/// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
/// </summary>
[global::System.Text.Json.Serialization.JsonPropertyName("raw_prompting")]
public bool? RawPrompting { get; set; }

/// <summary>
/// Configuration for forcing the model output to adhere to the specified format. Supported on [Command R 03-2024](https://docs.cohere.com/docs/command-r), [Command R+ 04-2024](https://docs.cohere.com/docs/command-r-plus) and newer models.<br/>
/// The model can be forced into outputting JSON objects (with up to 5 levels of nesting) by setting `{ "type": "json_object" }`.<br/>
Expand Down Expand Up @@ -377,11 +369,6 @@ public sealed partial class ChatRequest
/// - AUTO: Cohere Platform Only<br/>
/// - AUTO_PRESERVE_ORDER: Azure, AWS Sagemaker/Bedrock, Private Deployments
/// </param>
/// <param name="rawPrompting">
/// When enabled, the user's prompt will be sent to the model without<br/>
/// any pre-processing.<br/>
/// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
/// </param>
/// <param name="responseFormat">
/// Configuration for forcing the model output to adhere to the specified format. Supported on [Command R 03-2024](https://docs.cohere.com/docs/command-r), [Command R+ 04-2024](https://docs.cohere.com/docs/command-r-plus) and newer models.<br/>
/// The model can be forced into outputting JSON objects (with up to 5 levels of nesting) by setting `{ "type": "json_object" }`.<br/>
Expand Down Expand Up @@ -474,7 +461,6 @@ public ChatRequest(
string? preamble,
double? presencePenalty,
global::Cohere.ChatRequestPromptTruncation? promptTruncation,
bool? rawPrompting,
global::Cohere.ResponseFormat? responseFormat,
global::Cohere.ChatRequestSafetyMode? safetyMode,
bool? searchQueriesOnly,
Expand All @@ -501,7 +487,6 @@ public ChatRequest(
this.Preamble = preamble;
this.PresencePenalty = presencePenalty;
this.PromptTruncation = promptTruncation;
this.RawPrompting = rawPrompting;
this.ResponseFormat = responseFormat;
this.SafetyMode = safetyMode;
this.SearchQueriesOnly = searchQueriesOnly;
Expand Down
15 changes: 0 additions & 15 deletions src/libs/Cohere/Generated/Cohere.Models.Chatv2Request.g.cs
Original file line number Diff line number Diff line change
Expand Up @@ -80,14 +80,6 @@ public sealed partial class Chatv2Request
[global::System.Text.Json.Serialization.JsonPropertyName("presence_penalty")]
public float? PresencePenalty { get; set; }

/// <summary>
/// When enabled, the user's prompt will be sent to the model without<br/>
/// any pre-processing.<br/>
/// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
/// </summary>
[global::System.Text.Json.Serialization.JsonPropertyName("raw_prompting")]
public bool? RawPrompting { get; set; }

/// <summary>
/// The reasoning effort level of the model. This affects the model's performance and the time it takes to generate a response.
/// </summary>
Expand Down Expand Up @@ -221,11 +213,6 @@ public sealed partial class Chatv2Request
/// Defaults to `0.0`, min value of `0.0`, max value of `1.0`.<br/>
/// Used to reduce repetitiveness of generated tokens. Similar to `frequency_penalty`, except that this penalty is applied equally to all tokens that have already appeared, regardless of their exact frequencies.
/// </param>
/// <param name="rawPrompting">
/// When enabled, the user's prompt will be sent to the model without<br/>
/// any pre-processing.<br/>
/// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
/// </param>
/// <param name="reasoningEffort">
/// The reasoning effort level of the model. This affects the model's performance and the time it takes to generate a response.
/// </param>
Expand Down Expand Up @@ -292,7 +279,6 @@ public Chatv2Request(
int? maxTokens,
float? p,
float? presencePenalty,
bool? rawPrompting,
global::Cohere.ReasoningEffort? reasoningEffort,
global::Cohere.ResponseFormatV2? responseFormat,
global::Cohere.Chatv2RequestSafetyMode? safetyMode,
Expand All @@ -314,7 +300,6 @@ public Chatv2Request(
this.MaxTokens = maxTokens;
this.P = p;
this.PresencePenalty = presencePenalty;
this.RawPrompting = rawPrompting;
this.ReasoningEffort = reasoningEffort;
this.ResponseFormat = responseFormat;
this.SafetyMode = safetyMode;
Expand Down
10 changes: 0 additions & 10 deletions src/libs/Cohere/openapi.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -148,11 +148,6 @@ paths:
description: "Defaults to `AUTO` when `connectors` are specified and `OFF` in all other cases.\n\nDictates how the prompt will be constructed.\n\nWith `prompt_truncation` set to \"AUTO\", some elements from `chat_history` and `documents` will be dropped in an attempt to construct a prompt that fits within the model's context length limit. During this process the order of the documents and chat history will be changed and ranked by relevance.\n\nWith `prompt_truncation` set to \"AUTO_PRESERVE_ORDER\", some elements from `chat_history` and `documents` will be dropped in an attempt to construct a prompt that fits within the model's context length limit. During this process the order of the documents and chat history will be preserved as they are inputted into the API.\n\nWith `prompt_truncation` set to \"OFF\", no elements will be dropped. If the sum of the inputs exceeds the model's context length limit, a `TooManyTokens` error will be returned.\n\nCompatible Deployments:\n - AUTO: Cohere Platform Only\n - AUTO_PRESERVE_ORDER: Azure, AWS Sagemaker/Bedrock, Private Deployments\n"
x-fern-audiences:
- public
raw_prompting:
type: boolean
description: "When enabled, the user's prompt will be sent to the model without\nany pre-processing.\n\nCompatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments\n"
x-fern-audiences:
- sdk-only
response_format:
$ref: '#/components/schemas/ResponseFormat'
safety_mode:
Expand Down Expand Up @@ -7445,11 +7440,6 @@ paths:
format: float
x-fern-audiences:
- public
raw_prompting:
type: boolean
description: "When enabled, the user's prompt will be sent to the model without\nany pre-processing.\n\nCompatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments\n"
x-fern-audiences:
- sdk-only
reasoning_effort:
$ref: '#/components/schemas/ReasoningEffort'
response_format:
Expand Down