From f8ed89f5b3a6ddfabb5e1db4305ef2e814b093c9 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" Date: Sun, 22 Sep 2024 03:24:59 +0000 Subject: [PATCH] feat: Updated OpenAPI spec --- .../Generated/Cohere.CohereApi.Chat.g.cs | 35 ++++++----- .../Generated/Cohere.CohereApi.Classify.g.cs | 21 ++++--- .../Cohere.CohereApi.Detokenize.g.cs | 3 +- .../Generated/Cohere.CohereApi.Embed.g.cs | 20 ++++--- .../Generated/Cohere.CohereApi.Embedv2.g.cs | 18 +++--- .../Generated/Cohere.CohereApi.Generate.g.cs | 59 +++++++++++-------- .../Generated/Cohere.CohereApi.Summarize.g.cs | 33 +++++++---- ...FinetuningClient.CreateFinetunedModel.g.cs | 35 ----------- ...FinetuningClient.UpdateFinetunedModel.g.cs | 30 ---------- .../Generated/Cohere.ICohereApi.Chat.g.cs | 35 ++++++----- .../Generated/Cohere.ICohereApi.Classify.g.cs | 21 ++++--- .../Cohere.ICohereApi.Detokenize.g.cs | 3 +- .../Generated/Cohere.ICohereApi.Embed.g.cs | 20 ++++--- .../Generated/Cohere.ICohereApi.Embedv2.g.cs | 18 +++--- .../Generated/Cohere.ICohereApi.Generate.g.cs | 59 +++++++++++-------- .../Cohere.ICohereApi.Summarize.g.cs | 33 +++++++---- ...FinetuningClient.CreateFinetunedModel.g.cs | 28 --------- ...FinetuningClient.UpdateFinetunedModel.g.cs | 24 -------- .../Generated/Cohere.Models.BaseModel.g.cs | 3 +- .../Generated/Cohere.Models.ChatRequest.g.cs | 21 ++++--- .../Cohere.Models.ClassifyExample.g.cs | 6 +- .../Cohere.Models.ClassifyRequest.g.cs | 16 +++-- ...Cohere.Models.ClassifyRequestTruncate.g.cs | 3 +- .../Cohere.Models.ClassifyResponse.g.cs | 10 ++-- ...Models.ClassifyResponseClassification.g.cs | 40 +++++++------ ...ponseClassificationClassificationType.g.cs | 3 +- ....ClassifyResponseClassificationLabels.g.cs | 3 +- .../Generated/Cohere.Models.Dataset.g.cs | 2 +- .../Cohere.Models.DetokenizeRequest.g.cs | 6 +- .../Cohere.Models.DetokenizeResponse.g.cs | 6 +- .../Cohere.Models.EmbedByTypeResponse.g.cs | 6 +- ....Models.EmbedByTypeResponseEmbeddings.g.cs | 15 +++-- .../Cohere.Models.EmbedFloatsResponse.g.cs | 6 +- .../Generated/Cohere.Models.EmbedRequest.g.cs | 12 ++-- .../Cohere.Models.EmbedRequestTruncate.g.cs | 3 +- .../Cohere.Models.EmbeddingType.g.cs | 2 +- .../Cohere.Models.Embedv2Request.g.cs | 15 +++-- .../Cohere.Models.Embedv2RequestTruncate.g.cs | 3 +- .../Cohere.Models.FinetunedModel.g.cs | 21 ++++--- .../Cohere.Models.GenerateRequest.g.cs | 40 ++++++++----- ...Cohere.Models.GenerateRequestTruncate.g.cs | 3 +- .../Cohere.Models.GetClusterJobResponse.g.cs | 30 ++++++---- ...here.Models.GetConnectorsLogsResponse.g.cs | 6 +- .../Cohere.Models.ListConnectorsResponse.g.cs | 3 +- .../Generated/Cohere.Models.ParseInfo.g.cs | 6 +- .../Cohere.Models.RerankResponseResult.g.cs | 6 +- .../Cohere.Models.Rerankv2ResponseResult.g.cs | 6 +- .../Generated/Cohere.Models.Settings.g.cs | 3 +- .../Cohere.Models.SummarizeRequest.g.cs | 24 +++++--- ...Models.SummarizeRequestExtractiveness.g.cs | 3 +- .../Cohere.Models.SummarizeRequestFormat.g.cs | 3 +- .../Cohere.Models.SummarizeRequestLength.g.cs | 3 +- .../Cohere.Models.TokenizeResponse.g.cs | 6 +- ...re.Models.UpdateFinetunedModelRequest.g.cs | 18 ++++-- 54 files changed, 449 insertions(+), 408 deletions(-) diff --git a/src/libs/Cohere/Generated/Cohere.CohereApi.Chat.g.cs b/src/libs/Cohere/Generated/Cohere.CohereApi.Chat.g.cs index 0825e0b..498bb41 100644 --- a/src/libs/Cohere/Generated/Cohere.CohereApi.Chat.g.cs +++ b/src/libs/Cohere/Generated/Cohere.CohereApi.Chat.g.cs @@ -199,24 +199,28 @@ partial void ProcessChatResponseContent( /// /// /// The maximum number of tokens the model will generate as part of the response. Note: Setting a low value may result in incomplete generations.
- /// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments + /// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
+ /// Included only in requests /// /// /// The maximum number of input tokens to send to the model. If not specified, `max_input_tokens` is the model's context length limit minus a small buffer.
/// Input will be truncated according to the `prompt_truncation` parameter.
- /// Compatible Deployments: Cohere Platform + /// Compatible Deployments: Cohere Platform
+ /// Included only in requests /// /// /// Ensures only the top `k` most likely tokens are considered for generation at each step.
/// Defaults to `0`, min value of `0`, max value of `500`.
/// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
- /// Default Value: 0 + /// Default Value: 0
+ /// Included only in requests /// /// /// Ensures that only the most likely tokens, with total probability mass of `p`, are considered for generation at each step. If both `k` and `p` are enabled, `p` acts after `k`.
/// Defaults to `0.75`. min value of `0.01`, max value of `0.99`.
/// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
- /// Default Value: 0.75 + /// Default Value: 0.75
+ /// Included only in requests /// /// /// If specified, the backend will make a best effort to sample tokens
@@ -227,17 +231,20 @@ partial void ProcessChatResponseContent( /// /// /// A list of up to 5 strings that the model will use to stop generation. If the model generates a string that matches any of the strings in the list, it will stop generating tokens and return the generated text up to that point not including the stop sequence.
- /// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments + /// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
+ /// Included only in requests /// /// /// Defaults to `0.0`, min value of `0.0`, max value of `1.0`.
/// Used to reduce repetitiveness of generated tokens. The higher the value, the stronger a penalty is applied to previously present tokens, proportional to how many times they have already appeared in the prompt or prior generation.
- /// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments + /// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
+ /// Included only in requests /// /// /// Defaults to `0.0`, min value of `0.0`, max value of `1.0`.
/// Used to reduce repetitiveness of generated tokens. Similar to `frequency_penalty`, except that this penalty is applied equally to all tokens that have already appeared, regardless of their exact frequencies.
- /// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments + /// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
+ /// Included only in requests /// /// /// A list of available tools (functions) that the model may suggest invoking before producing a text response.
@@ -288,6 +295,13 @@ partial void ProcessChatResponseContent( /// public async global::System.Threading.Tasks.Task> ChatAsync( string message, + int maxTokens, + int maxInputTokens, + int k, + double p, + global::System.Collections.Generic.IList stopSequences, + double frequencyPenalty, + double presencePenalty, string? xClientName = default, global::Cohere.ChatAccepts? accepts = default, string? model = default, @@ -301,14 +315,7 @@ partial void ProcessChatResponseContent( global::System.Collections.Generic.IList? documents = default, global::Cohere.ChatRequestCitationQuality? citationQuality = default, float? temperature = default, - int? maxTokens = default, - int? maxInputTokens = default, - int? k = 0, - double? p = 0.75, int? seed = default, - global::System.Collections.Generic.IList? stopSequences = default, - double? frequencyPenalty = default, - double? presencePenalty = default, global::System.Collections.Generic.IList? tools = default, global::System.Collections.Generic.IList? toolResults = default, bool? forceSingleStep = default, diff --git a/src/libs/Cohere/Generated/Cohere.CohereApi.Classify.g.cs b/src/libs/Cohere/Generated/Cohere.CohereApi.Classify.g.cs index 77e3896..8c3663f 100644 --- a/src/libs/Cohere/Generated/Cohere.CohereApi.Classify.g.cs +++ b/src/libs/Cohere/Generated/Cohere.CohereApi.Classify.g.cs @@ -115,34 +115,39 @@ partial void ProcessClassifyResponseContent( /// /// A list of up to 96 texts to be classified. Each one must be a non-empty string.
/// There is, however, no consistent, universal limit to the length a particular input can be. We perform classification on the first `x` tokens of each input, and `x` varies depending on which underlying model is powering classification. The maximum token length for each model is listed in the "max tokens" column [here](https://docs.cohere.com/docs/models).
- /// Note: by default the `truncate` parameter is set to `END`, so tokens exceeding the limit will be automatically dropped. This behavior can be disabled by setting `truncate` to `NONE`, which will result in validation errors for longer texts. + /// Note: by default the `truncate` parameter is set to `END`, so tokens exceeding the limit will be automatically dropped. This behavior can be disabled by setting `truncate` to `NONE`, which will result in validation errors for longer texts.
+ /// Included only in requests /// /// /// An array of examples to provide context to the model. Each example is a text string and its associated label/class. Each unique label requires at least 2 examples associated with it; the maximum number of examples is 2500, and each example has a maximum length of 512 tokens. The values should be structured as `{text: "...",label: "..."}`.
- /// Note: [Fine-tuned Models](https://docs.cohere.com/docs/classify-fine-tuning) trained on classification examples don't require the `examples` parameter to be passed in explicitly. + /// Note: [Fine-tuned Models](https://docs.cohere.com/docs/classify-fine-tuning) trained on classification examples don't require the `examples` parameter to be passed in explicitly.
+ /// Included only in requests /// /// - /// The identifier of the model. Currently available models are `embed-multilingual-v2.0`, `embed-english-light-v2.0`, and `embed-english-v2.0` (default). Smaller "light" models are faster, while larger models will perform better. [Fine-tuned models](https://docs.cohere.com/docs/fine-tuning) can also be supplied with their full ID. + /// The identifier of the model. Currently available models are `embed-multilingual-v2.0`, `embed-english-light-v2.0`, and `embed-english-v2.0` (default). Smaller "light" models are faster, while larger models will perform better. [Fine-tuned models](https://docs.cohere.com/docs/fine-tuning) can also be supplied with their full ID.
+ /// Included only in requests /// /// /// The ID of a custom playground preset. You can create presets in the [playground](https://dashboard.cohere.com/playground/classify?model=large). If you use a preset, all other parameters become optional, and any included parameters will override the preset's parameters.
+ /// Included only in requests
/// Example: my-preset-a58sbd /// /// /// One of `NONE|START|END` to specify how the API will handle inputs longer than the maximum token length.
/// Passing `START` will discard the start of the input. `END` will discard the end of the input. In both cases, input is discarded until the remaining input is exactly the maximum input token length for the model.
/// If `NONE` is selected, when the input exceeds the maximum input token length an error will be returned.
- /// Default Value: END + /// Default Value: END
+ /// Included only in requests /// /// The token to cancel the operation with /// public async global::System.Threading.Tasks.Task ClassifyAsync( global::System.Collections.Generic.IList inputs, + global::System.Collections.Generic.IList examples, + string model, + string preset, + global::Cohere.ClassifyRequestTruncate truncate, string? xClientName = default, - global::System.Collections.Generic.IList? examples = default, - string? model = default, - string? preset = default, - global::Cohere.ClassifyRequestTruncate? truncate = global::Cohere.ClassifyRequestTruncate.END, global::System.Threading.CancellationToken cancellationToken = default) { var request = new global::Cohere.ClassifyRequest diff --git a/src/libs/Cohere/Generated/Cohere.CohereApi.Detokenize.g.cs b/src/libs/Cohere/Generated/Cohere.CohereApi.Detokenize.g.cs index aaa3588..e7eb27b 100644 --- a/src/libs/Cohere/Generated/Cohere.CohereApi.Detokenize.g.cs +++ b/src/libs/Cohere/Generated/Cohere.CohereApi.Detokenize.g.cs @@ -111,7 +111,8 @@ partial void ProcessDetokenizeResponseContent( /// /// /// - /// The list of tokens to be detokenized. + /// The list of tokens to be detokenized.
+ /// Included only in requests /// /// /// An optional parameter to provide the model name. This will ensure that the detokenization is done by the tokenizer used by that model. diff --git a/src/libs/Cohere/Generated/Cohere.CohereApi.Embed.g.cs b/src/libs/Cohere/Generated/Cohere.CohereApi.Embed.g.cs index 9f681ee..5c76788 100644 --- a/src/libs/Cohere/Generated/Cohere.CohereApi.Embed.g.cs +++ b/src/libs/Cohere/Generated/Cohere.CohereApi.Embed.g.cs @@ -115,7 +115,8 @@ partial void ProcessEmbedResponseContent( /// /// /// - /// An array of strings for the model to embed. Maximum number of texts per call is `96`. We recommend reducing the length of each text to be under `512` tokens for optimal quality. + /// An array of strings for the model to embed. Maximum number of texts per call is `96`. We recommend reducing the length of each text to be under `512` tokens for optimal quality.
+ /// Included only in requests /// /// /// Defaults to embed-english-v2.0
@@ -127,7 +128,8 @@ partial void ProcessEmbedResponseContent( /// * `embed-multilingual-light-v3.0` 384
/// * `embed-english-v2.0` 4096
/// * `embed-english-light-v2.0` 1024
- /// * `embed-multilingual-v2.0` 768 + /// * `embed-multilingual-v2.0` 768
+ /// Included only in requests /// /// /// Specifies the type of input passed to the model. Required for embedding models v3 and higher.
@@ -143,23 +145,25 @@ partial void ProcessEmbedResponseContent( /// * `"int8"`: Use this when you want to get back signed int8 embeddings. Valid for only v3 models.
/// * `"uint8"`: Use this when you want to get back unsigned int8 embeddings. Valid for only v3 models.
/// * `"binary"`: Use this when you want to get back signed binary embeddings. Valid for only v3 models.
- /// * `"ubinary"`: Use this when you want to get back unsigned binary embeddings. Valid for only v3 models. + /// * `"ubinary"`: Use this when you want to get back unsigned binary embeddings. Valid for only v3 models.
+ /// Included only in requests /// /// /// One of `NONE|START|END` to specify how the API will handle inputs longer than the maximum token length.
/// Passing `START` will discard the start of the input. `END` will discard the end of the input. In both cases, input is discarded until the remaining input is exactly the maximum input token length for the model.
/// If `NONE` is selected, when the input exceeds the maximum input token length an error will be returned.
- /// Default Value: END + /// Default Value: END
+ /// Included only in requests /// /// The token to cancel the operation with /// public async global::System.Threading.Tasks.Task> EmbedAsync( + global::System.Collections.Generic.IList texts, + string model, + global::System.Collections.Generic.IList embeddingTypes, + global::Cohere.EmbedRequestTruncate truncate, string? xClientName = default, - global::System.Collections.Generic.IList? texts = default, - string? model = default, global::Cohere.EmbedInputType? inputType = default, - global::System.Collections.Generic.IList? embeddingTypes = default, - global::Cohere.EmbedRequestTruncate? truncate = global::Cohere.EmbedRequestTruncate.END, global::System.Threading.CancellationToken cancellationToken = default) { var request = new global::Cohere.EmbedRequest diff --git a/src/libs/Cohere/Generated/Cohere.CohereApi.Embedv2.g.cs b/src/libs/Cohere/Generated/Cohere.CohereApi.Embedv2.g.cs index c7c5000..87123ed 100644 --- a/src/libs/Cohere/Generated/Cohere.CohereApi.Embedv2.g.cs +++ b/src/libs/Cohere/Generated/Cohere.CohereApi.Embedv2.g.cs @@ -115,7 +115,8 @@ partial void ProcessEmbedv2ResponseContent( /// /// /// - /// An array of strings for the model to embed. Maximum number of texts per call is `96`. We recommend reducing the length of each text to be under `512` tokens for optimal quality. + /// An array of strings for the model to embed. Maximum number of texts per call is `96`. We recommend reducing the length of each text to be under `512` tokens for optimal quality.
+ /// Included only in requests /// /// /// Defaults to embed-english-v2.0
@@ -127,7 +128,8 @@ partial void ProcessEmbedv2ResponseContent( /// * `embed-multilingual-light-v3.0` 384
/// * `embed-english-v2.0` 4096
/// * `embed-english-light-v2.0` 1024
- /// * `embed-multilingual-v2.0` 768 + /// * `embed-multilingual-v2.0` 768
+ /// Included only in requests /// /// /// Specifies the type of input passed to the model. Required for embedding models v3 and higher.
@@ -143,23 +145,25 @@ partial void ProcessEmbedv2ResponseContent( /// * `"int8"`: Use this when you want to get back signed int8 embeddings. Valid for only v3 models.
/// * `"uint8"`: Use this when you want to get back unsigned int8 embeddings. Valid for only v3 models.
/// * `"binary"`: Use this when you want to get back signed binary embeddings. Valid for only v3 models.
- /// * `"ubinary"`: Use this when you want to get back unsigned binary embeddings. Valid for only v3 models. + /// * `"ubinary"`: Use this when you want to get back unsigned binary embeddings. Valid for only v3 models.
+ /// Included only in requests /// /// /// One of `NONE|START|END` to specify how the API will handle inputs longer than the maximum token length.
/// Passing `START` will discard the start of the input. `END` will discard the end of the input. In both cases, input is discarded until the remaining input is exactly the maximum input token length for the model.
/// If `NONE` is selected, when the input exceeds the maximum input token length an error will be returned.
- /// Default Value: END + /// Default Value: END
+ /// Included only in requests /// /// The token to cancel the operation with /// public async global::System.Threading.Tasks.Task Embedv2Async( + global::System.Collections.Generic.IList texts, string model, + global::System.Collections.Generic.IList embeddingTypes, + global::Cohere.Embedv2RequestTruncate truncate, string? xClientName = default, - global::System.Collections.Generic.IList? texts = default, global::Cohere.EmbedInputType? inputType = default, - global::System.Collections.Generic.IList? embeddingTypes = default, - global::Cohere.Embedv2RequestTruncate? truncate = global::Cohere.Embedv2RequestTruncate.END, global::System.Threading.CancellationToken cancellationToken = default) { var request = new global::Cohere.Embedv2Request diff --git a/src/libs/Cohere/Generated/Cohere.CohereApi.Generate.g.cs b/src/libs/Cohere/Generated/Cohere.CohereApi.Generate.g.cs index 418288a..94ac229 100644 --- a/src/libs/Cohere/Generated/Cohere.CohereApi.Generate.g.cs +++ b/src/libs/Cohere/Generated/Cohere.CohereApi.Generate.g.cs @@ -119,14 +119,17 @@ partial void ProcessGenerateResponseContent( /// /// The input text that serves as the starting point for generating the response.
/// Note: The prompt will be pre-processed and modified before reaching the model.
+ /// Included only in requests
/// Example: Please explain to me how LLMs work /// /// /// The identifier of the model to generate with. Currently available models are `command` (default), `command-nightly` (experimental), `command-light`, and `command-light-nightly` (experimental).
- /// Smaller, "light" models are faster, while larger models will perform better. [Custom models](/docs/training-custom-models) can also be supplied with their full ID. + /// Smaller, "light" models are faster, while larger models will perform better. [Custom models](/docs/training-custom-models) can also be supplied with their full ID.
+ /// Included only in requests /// /// - /// The maximum number of generations that will be returned. Defaults to `1`, min value of `1`, max value of `5`. + /// The maximum number of generations that will be returned. Defaults to `1`, min value of `1`, max value of `5`.
+ /// Included only in requests /// /// /// When `true`, the response will be a JSON stream of events. Streaming is beneficial for user interfaces that render the contents of the response piece by piece, as it gets generated.
@@ -139,17 +142,20 @@ partial void ProcessGenerateResponseContent( /// /// The maximum number of tokens the model will generate as part of the response. Note: Setting a low value may result in incomplete generations.
/// This parameter is off by default, and if it's not specified, the model will continue generating until it emits an EOS completion token. See [BPE Tokens](/bpe-tokens-wiki) for more details.
- /// Can only be set to `0` if `return_likelihoods` is set to `ALL` to get the likelihood of the prompt. + /// Can only be set to `0` if `return_likelihoods` is set to `ALL` to get the likelihood of the prompt.
+ /// Included only in requests /// /// /// One of `NONE|START|END` to specify how the API will handle inputs longer than the maximum token length.
/// Passing `START` will discard the start of the input. `END` will discard the end of the input. In both cases, input is discarded until the remaining input is exactly the maximum input token length for the model.
/// If `NONE` is selected, when the input exceeds the maximum input token length an error will be returned.
- /// Default Value: END + /// Default Value: END
+ /// Included only in requests /// /// /// A non-negative float that tunes the degree of randomness in generation. Lower temperatures mean less random generations. See [Temperature](/temperature-wiki) for more details.
- /// Defaults to `0.75`, min value of `0.0`, max value of `5.0`. + /// Defaults to `0.75`, min value of `0.0`, max value of `5.0`.
+ /// Included only in requests /// /// /// If specified, the backend will make a best effort to sample tokens
@@ -161,30 +167,37 @@ partial void ProcessGenerateResponseContent( /// /// Identifier of a custom preset. A preset is a combination of parameters, such as prompt, temperature etc. You can create presets in the [playground](https://dashboard.cohere.com/playground/generate).
/// When a preset is specified, the `prompt` parameter becomes optional, and any included parameters will override the preset's parameters.
+ /// Included only in requests
/// Example: my-preset-a58sbd /// /// - /// The generated text will be cut at the beginning of the earliest occurrence of an end sequence. The sequence will be excluded from the text. + /// The generated text will be cut at the beginning of the earliest occurrence of an end sequence. The sequence will be excluded from the text.
+ /// Included only in requests /// /// - /// The generated text will be cut at the end of the earliest occurrence of a stop sequence. The sequence will be included the text. + /// The generated text will be cut at the end of the earliest occurrence of a stop sequence. The sequence will be included the text.
+ /// Included only in requests /// /// /// Ensures only the top `k` most likely tokens are considered for generation at each step.
- /// Defaults to `0`, min value of `0`, max value of `500`. + /// Defaults to `0`, min value of `0`, max value of `500`.
+ /// Included only in requests /// /// /// Ensures that only the most likely tokens, with total probability mass of `p`, are considered for generation at each step. If both `k` and `p` are enabled, `p` acts after `k`.
- /// Defaults to `0.75`. min value of `0.01`, max value of `0.99`. + /// Defaults to `0.75`. min value of `0.01`, max value of `0.99`.
+ /// Included only in requests /// /// /// Used to reduce repetitiveness of generated tokens. The higher the value, the stronger a penalty is applied to previously present tokens, proportional to how many times they have already appeared in the prompt or prior generation.
- /// Using `frequency_penalty` in combination with `presence_penalty` is not supported on newer models. + /// Using `frequency_penalty` in combination with `presence_penalty` is not supported on newer models.
+ /// Included only in requests /// /// /// Defaults to `0.0`, min value of `0.0`, max value of `1.0`.
/// Can be used to reduce repetitiveness of generated tokens. Similar to `frequency_penalty`, except that this penalty is applied equally to all tokens that have already appeared, regardless of their exact frequencies.
- /// Using `frequency_penalty` in combination with `presence_penalty` is not supported on newer models. + /// Using `frequency_penalty` in combination with `presence_penalty` is not supported on newer models.
+ /// Included only in requests /// /// /// One of `GENERATION|ALL|NONE` to specify how and if the token likelihoods are returned with the response. Defaults to `NONE`.
@@ -199,21 +212,21 @@ partial void ProcessGenerateResponseContent( /// public async global::System.Threading.Tasks.Task GenerateAsync( string prompt, + string model, + int numGenerations, + int maxTokens, + global::Cohere.GenerateRequestTruncate truncate, + double temperature, + string preset, + global::System.Collections.Generic.IList endSequences, + global::System.Collections.Generic.IList stopSequences, + int k, + double p, + double frequencyPenalty, + double presencePenalty, string? xClientName = default, - string? model = default, - int? numGenerations = default, bool? stream = default, - int? maxTokens = default, - global::Cohere.GenerateRequestTruncate? truncate = global::Cohere.GenerateRequestTruncate.END, - double? temperature = default, int? seed = default, - string? preset = default, - global::System.Collections.Generic.IList? endSequences = default, - global::System.Collections.Generic.IList? stopSequences = default, - int? k = default, - double? p = default, - double? frequencyPenalty = default, - double? presencePenalty = default, global::Cohere.GenerateRequestReturnLikelihoods? returnLikelihoods = global::Cohere.GenerateRequestReturnLikelihoods.NONE, bool? rawPrompting = default, global::System.Threading.CancellationToken cancellationToken = default) diff --git a/src/libs/Cohere/Generated/Cohere.CohereApi.Summarize.g.cs b/src/libs/Cohere/Generated/Cohere.CohereApi.Summarize.g.cs index aaedf8e..f990be1 100644 --- a/src/libs/Cohere/Generated/Cohere.CohereApi.Summarize.g.cs +++ b/src/libs/Cohere/Generated/Cohere.CohereApi.Summarize.g.cs @@ -117,41 +117,48 @@ partial void ProcessSummarizeResponseContent( /// /// /// - /// The text to generate a summary for. Can be up to 100,000 characters long. Currently the only supported language is English. + /// The text to generate a summary for. Can be up to 100,000 characters long. Currently the only supported language is English.
+ /// Included only in requests /// /// /// One of `short`, `medium`, `long`, or `auto` defaults to `auto`. Indicates the approximate length of the summary. If `auto` is selected, the best option will be picked based on the input text.
- /// Default Value: medium + /// Default Value: medium
+ /// Included only in requests /// /// /// One of `paragraph`, `bullets`, or `auto`, defaults to `auto`. Indicates the style in which the summary will be delivered - in a free form paragraph or in bullet points. If `auto` is selected, the best option will be picked based on the input text.
- /// Default Value: paragraph + /// Default Value: paragraph
+ /// Included only in requests /// /// - /// The identifier of the model to generate the summary with. Currently available models are `command` (default), `command-nightly` (experimental), `command-light`, and `command-light-nightly` (experimental). Smaller, "light" models are faster, while larger models will perform better. + /// The identifier of the model to generate the summary with. Currently available models are `command` (default), `command-nightly` (experimental), `command-light`, and `command-light-nightly` (experimental). Smaller, "light" models are faster, while larger models will perform better.
+ /// Included only in requests /// /// /// One of `low`, `medium`, `high`, or `auto`, defaults to `auto`. Controls how close to the original text the summary is. `high` extractiveness summaries will lean towards reusing sentences verbatim, while `low` extractiveness summaries will tend to paraphrase more. If `auto` is selected, the best option will be picked based on the input text.
- /// Default Value: low + /// Default Value: low
+ /// Included only in requests /// /// /// Ranges from 0 to 5. Controls the randomness of the output. Lower values tend to generate more “predictable” output, while higher values tend to generate more “creative” output. The sweet spot is typically between 0 and 1.
- /// Default Value: 0.3 + /// Default Value: 0.3
+ /// Included only in requests /// /// - /// A free-form instruction for modifying how the summaries get generated. Should complete the sentence "Generate a summary _". Eg. "focusing on the next steps" or "written by Yoda" + /// A free-form instruction for modifying how the summaries get generated. Should complete the sentence "Generate a summary _". Eg. "focusing on the next steps" or "written by Yoda"
+ /// Included only in requests /// /// The token to cancel the operation with /// public async global::System.Threading.Tasks.Task SummarizeAsync( string text, + global::Cohere.SummarizeRequestLength length, + global::Cohere.SummarizeRequestFormat format, + string model, + global::Cohere.SummarizeRequestExtractiveness extractiveness, + double temperature, + string additionalCommand, string? xClientName = default, - global::Cohere.SummarizeRequestLength? length = global::Cohere.SummarizeRequestLength.Medium, - global::Cohere.SummarizeRequestFormat? format = global::Cohere.SummarizeRequestFormat.Paragraph, - string? model = default, - global::Cohere.SummarizeRequestExtractiveness? extractiveness = global::Cohere.SummarizeRequestExtractiveness.Low, - double? temperature = 0.3, - string? additionalCommand = default, global::System.Threading.CancellationToken cancellationToken = default) { var request = new global::Cohere.SummarizeRequest diff --git a/src/libs/Cohere/Generated/Cohere.FinetuningClient.CreateFinetunedModel.g.cs b/src/libs/Cohere/Generated/Cohere.FinetuningClient.CreateFinetunedModel.g.cs index 1e583fc..27d5dd5 100644 --- a/src/libs/Cohere/Generated/Cohere.FinetuningClient.CreateFinetunedModel.g.cs +++ b/src/libs/Cohere/Generated/Cohere.FinetuningClient.CreateFinetunedModel.g.cs @@ -108,18 +108,9 @@ partial void ProcessCreateFinetunedModelResponseContent( /// Trains and deploys a fine-tuned model. /// /// - /// - /// read-only. FinetunedModel ID. - /// /// /// FinetunedModel name (e.g. `foobar`). /// - /// - /// read-only. User ID of the creator. - /// - /// - /// read-only. Organization ID. - /// /// /// The configuration used for fine-tuning. /// @@ -136,46 +127,20 @@ partial void ProcessCreateFinetunedModelResponseContent( /// - STATUS_QUEUED: The fine-tuned model is queued for training.
/// Default Value: STATUS_UNSPECIFIED /// - /// - /// read-only. Creation timestamp. - /// - /// - /// read-only. Latest update timestamp. - /// - /// - /// read-only. Timestamp for the completed fine-tuning. - /// - /// - /// read-only. Timestamp for the latest request to this fine-tuned model. - /// /// The token to cancel the operation with /// public async global::System.Threading.Tasks.Task CreateFinetunedModelAsync( string name, global::Cohere.Settings settings, string? xClientName = default, - string? id = default, - string? creatorId = default, - string? organizationId = default, global::Cohere.Status? status = global::Cohere.Status.STATUSUNSPECIFIED, - global::System.DateTime? createdAt = default, - global::System.DateTime? updatedAt = default, - global::System.DateTime? completedAt = default, - global::System.DateTime? lastUsed = default, global::System.Threading.CancellationToken cancellationToken = default) { var request = new global::Cohere.FinetunedModel { - Id = id, Name = name, - CreatorId = creatorId, - OrganizationId = organizationId, Settings = settings, Status = status, - CreatedAt = createdAt, - UpdatedAt = updatedAt, - CompletedAt = completedAt, - LastUsed = lastUsed, }; return await CreateFinetunedModelAsync( diff --git a/src/libs/Cohere/Generated/Cohere.FinetuningClient.UpdateFinetunedModel.g.cs b/src/libs/Cohere/Generated/Cohere.FinetuningClient.UpdateFinetunedModel.g.cs index 969a567..8db8198 100644 --- a/src/libs/Cohere/Generated/Cohere.FinetuningClient.UpdateFinetunedModel.g.cs +++ b/src/libs/Cohere/Generated/Cohere.FinetuningClient.UpdateFinetunedModel.g.cs @@ -118,12 +118,6 @@ partial void ProcessUpdateFinetunedModelResponseContent( /// /// FinetunedModel name (e.g. `foobar`). /// - /// - /// User ID of the creator. - /// - /// - /// Organization ID. - /// /// /// The configuration used for fine-tuning. /// @@ -140,18 +134,6 @@ partial void ProcessUpdateFinetunedModelResponseContent( /// - STATUS_QUEUED: The fine-tuned model is queued for training.
/// Default Value: STATUS_UNSPECIFIED /// - /// - /// Creation timestamp. - /// - /// - /// Latest update timestamp. - /// - /// - /// Timestamp for the completed fine-tuning. - /// - /// - /// Timestamp for the latest request to this fine-tuned model. - /// /// The token to cancel the operation with /// public async global::System.Threading.Tasks.Task UpdateFinetunedModelAsync( @@ -159,26 +141,14 @@ partial void ProcessUpdateFinetunedModelResponseContent( string name, global::Cohere.Settings settings, string? xClientName = default, - string? creatorId = default, - string? organizationId = default, global::Cohere.Status? status = global::Cohere.Status.STATUSUNSPECIFIED, - global::System.DateTime? createdAt = default, - global::System.DateTime? updatedAt = default, - global::System.DateTime? completedAt = default, - global::System.DateTime? lastUsed = default, global::System.Threading.CancellationToken cancellationToken = default) { var request = new global::Cohere.UpdateFinetunedModelRequest { Name = name, - CreatorId = creatorId, - OrganizationId = organizationId, Settings = settings, Status = status, - CreatedAt = createdAt, - UpdatedAt = updatedAt, - CompletedAt = completedAt, - LastUsed = lastUsed, }; return await UpdateFinetunedModelAsync( diff --git a/src/libs/Cohere/Generated/Cohere.ICohereApi.Chat.g.cs b/src/libs/Cohere/Generated/Cohere.ICohereApi.Chat.g.cs index c83fc6b..29b68ce 100644 --- a/src/libs/Cohere/Generated/Cohere.ICohereApi.Chat.g.cs +++ b/src/libs/Cohere/Generated/Cohere.ICohereApi.Chat.g.cs @@ -107,24 +107,28 @@ public partial interface ICohereApi /// /// /// The maximum number of tokens the model will generate as part of the response. Note: Setting a low value may result in incomplete generations.
- /// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments + /// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
+ /// Included only in requests /// /// /// The maximum number of input tokens to send to the model. If not specified, `max_input_tokens` is the model's context length limit minus a small buffer.
/// Input will be truncated according to the `prompt_truncation` parameter.
- /// Compatible Deployments: Cohere Platform + /// Compatible Deployments: Cohere Platform
+ /// Included only in requests /// /// /// Ensures only the top `k` most likely tokens are considered for generation at each step.
/// Defaults to `0`, min value of `0`, max value of `500`.
/// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
- /// Default Value: 0 + /// Default Value: 0
+ /// Included only in requests /// /// /// Ensures that only the most likely tokens, with total probability mass of `p`, are considered for generation at each step. If both `k` and `p` are enabled, `p` acts after `k`.
/// Defaults to `0.75`. min value of `0.01`, max value of `0.99`.
/// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
- /// Default Value: 0.75 + /// Default Value: 0.75
+ /// Included only in requests /// /// /// If specified, the backend will make a best effort to sample tokens
@@ -135,17 +139,20 @@ public partial interface ICohereApi /// /// /// A list of up to 5 strings that the model will use to stop generation. If the model generates a string that matches any of the strings in the list, it will stop generating tokens and return the generated text up to that point not including the stop sequence.
- /// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments + /// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
+ /// Included only in requests /// /// /// Defaults to `0.0`, min value of `0.0`, max value of `1.0`.
/// Used to reduce repetitiveness of generated tokens. The higher the value, the stronger a penalty is applied to previously present tokens, proportional to how many times they have already appeared in the prompt or prior generation.
- /// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments + /// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
+ /// Included only in requests /// /// /// Defaults to `0.0`, min value of `0.0`, max value of `1.0`.
/// Used to reduce repetitiveness of generated tokens. Similar to `frequency_penalty`, except that this penalty is applied equally to all tokens that have already appeared, regardless of their exact frequencies.
- /// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments + /// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
+ /// Included only in requests /// /// /// A list of available tools (functions) that the model may suggest invoking before producing a text response.
@@ -196,6 +203,13 @@ public partial interface ICohereApi /// global::System.Threading.Tasks.Task> ChatAsync( string message, + int maxTokens, + int maxInputTokens, + int k, + double p, + global::System.Collections.Generic.IList stopSequences, + double frequencyPenalty, + double presencePenalty, string? xClientName = default, global::Cohere.ChatAccepts? accepts = default, string? model = default, @@ -209,14 +223,7 @@ public partial interface ICohereApi global::System.Collections.Generic.IList? documents = default, global::Cohere.ChatRequestCitationQuality? citationQuality = default, float? temperature = default, - int? maxTokens = default, - int? maxInputTokens = default, - int? k = 0, - double? p = 0.75, int? seed = default, - global::System.Collections.Generic.IList? stopSequences = default, - double? frequencyPenalty = default, - double? presencePenalty = default, global::System.Collections.Generic.IList? tools = default, global::System.Collections.Generic.IList? toolResults = default, bool? forceSingleStep = default, diff --git a/src/libs/Cohere/Generated/Cohere.ICohereApi.Classify.g.cs b/src/libs/Cohere/Generated/Cohere.ICohereApi.Classify.g.cs index 8f8373f..5699187 100644 --- a/src/libs/Cohere/Generated/Cohere.ICohereApi.Classify.g.cs +++ b/src/libs/Cohere/Generated/Cohere.ICohereApi.Classify.g.cs @@ -27,34 +27,39 @@ public partial interface ICohereApi /// /// A list of up to 96 texts to be classified. Each one must be a non-empty string.
/// There is, however, no consistent, universal limit to the length a particular input can be. We perform classification on the first `x` tokens of each input, and `x` varies depending on which underlying model is powering classification. The maximum token length for each model is listed in the "max tokens" column [here](https://docs.cohere.com/docs/models).
- /// Note: by default the `truncate` parameter is set to `END`, so tokens exceeding the limit will be automatically dropped. This behavior can be disabled by setting `truncate` to `NONE`, which will result in validation errors for longer texts. + /// Note: by default the `truncate` parameter is set to `END`, so tokens exceeding the limit will be automatically dropped. This behavior can be disabled by setting `truncate` to `NONE`, which will result in validation errors for longer texts.
+ /// Included only in requests /// /// /// An array of examples to provide context to the model. Each example is a text string and its associated label/class. Each unique label requires at least 2 examples associated with it; the maximum number of examples is 2500, and each example has a maximum length of 512 tokens. The values should be structured as `{text: "...",label: "..."}`.
- /// Note: [Fine-tuned Models](https://docs.cohere.com/docs/classify-fine-tuning) trained on classification examples don't require the `examples` parameter to be passed in explicitly. + /// Note: [Fine-tuned Models](https://docs.cohere.com/docs/classify-fine-tuning) trained on classification examples don't require the `examples` parameter to be passed in explicitly.
+ /// Included only in requests /// /// - /// The identifier of the model. Currently available models are `embed-multilingual-v2.0`, `embed-english-light-v2.0`, and `embed-english-v2.0` (default). Smaller "light" models are faster, while larger models will perform better. [Fine-tuned models](https://docs.cohere.com/docs/fine-tuning) can also be supplied with their full ID. + /// The identifier of the model. Currently available models are `embed-multilingual-v2.0`, `embed-english-light-v2.0`, and `embed-english-v2.0` (default). Smaller "light" models are faster, while larger models will perform better. [Fine-tuned models](https://docs.cohere.com/docs/fine-tuning) can also be supplied with their full ID.
+ /// Included only in requests /// /// /// The ID of a custom playground preset. You can create presets in the [playground](https://dashboard.cohere.com/playground/classify?model=large). If you use a preset, all other parameters become optional, and any included parameters will override the preset's parameters.
+ /// Included only in requests
/// Example: my-preset-a58sbd /// /// /// One of `NONE|START|END` to specify how the API will handle inputs longer than the maximum token length.
/// Passing `START` will discard the start of the input. `END` will discard the end of the input. In both cases, input is discarded until the remaining input is exactly the maximum input token length for the model.
/// If `NONE` is selected, when the input exceeds the maximum input token length an error will be returned.
- /// Default Value: END + /// Default Value: END
+ /// Included only in requests /// /// The token to cancel the operation with /// global::System.Threading.Tasks.Task ClassifyAsync( global::System.Collections.Generic.IList inputs, + global::System.Collections.Generic.IList examples, + string model, + string preset, + global::Cohere.ClassifyRequestTruncate truncate, string? xClientName = default, - global::System.Collections.Generic.IList? examples = default, - string? model = default, - string? preset = default, - global::Cohere.ClassifyRequestTruncate? truncate = global::Cohere.ClassifyRequestTruncate.END, global::System.Threading.CancellationToken cancellationToken = default); } } \ No newline at end of file diff --git a/src/libs/Cohere/Generated/Cohere.ICohereApi.Detokenize.g.cs b/src/libs/Cohere/Generated/Cohere.ICohereApi.Detokenize.g.cs index bb2cd07..cccaf05 100644 --- a/src/libs/Cohere/Generated/Cohere.ICohereApi.Detokenize.g.cs +++ b/src/libs/Cohere/Generated/Cohere.ICohereApi.Detokenize.g.cs @@ -23,7 +23,8 @@ public partial interface ICohereApi /// /// /// - /// The list of tokens to be detokenized. + /// The list of tokens to be detokenized.
+ /// Included only in requests /// /// /// An optional parameter to provide the model name. This will ensure that the detokenization is done by the tokenizer used by that model. diff --git a/src/libs/Cohere/Generated/Cohere.ICohereApi.Embed.g.cs b/src/libs/Cohere/Generated/Cohere.ICohereApi.Embed.g.cs index 56227ca..30d2306 100644 --- a/src/libs/Cohere/Generated/Cohere.ICohereApi.Embed.g.cs +++ b/src/libs/Cohere/Generated/Cohere.ICohereApi.Embed.g.cs @@ -27,7 +27,8 @@ public partial interface ICohereApi /// /// /// - /// An array of strings for the model to embed. Maximum number of texts per call is `96`. We recommend reducing the length of each text to be under `512` tokens for optimal quality. + /// An array of strings for the model to embed. Maximum number of texts per call is `96`. We recommend reducing the length of each text to be under `512` tokens for optimal quality.
+ /// Included only in requests /// /// /// Defaults to embed-english-v2.0
@@ -39,7 +40,8 @@ public partial interface ICohereApi /// * `embed-multilingual-light-v3.0` 384
/// * `embed-english-v2.0` 4096
/// * `embed-english-light-v2.0` 1024
- /// * `embed-multilingual-v2.0` 768 + /// * `embed-multilingual-v2.0` 768
+ /// Included only in requests /// /// /// Specifies the type of input passed to the model. Required for embedding models v3 and higher.
@@ -55,23 +57,25 @@ public partial interface ICohereApi /// * `"int8"`: Use this when you want to get back signed int8 embeddings. Valid for only v3 models.
/// * `"uint8"`: Use this when you want to get back unsigned int8 embeddings. Valid for only v3 models.
/// * `"binary"`: Use this when you want to get back signed binary embeddings. Valid for only v3 models.
- /// * `"ubinary"`: Use this when you want to get back unsigned binary embeddings. Valid for only v3 models. + /// * `"ubinary"`: Use this when you want to get back unsigned binary embeddings. Valid for only v3 models.
+ /// Included only in requests /// /// /// One of `NONE|START|END` to specify how the API will handle inputs longer than the maximum token length.
/// Passing `START` will discard the start of the input. `END` will discard the end of the input. In both cases, input is discarded until the remaining input is exactly the maximum input token length for the model.
/// If `NONE` is selected, when the input exceeds the maximum input token length an error will be returned.
- /// Default Value: END + /// Default Value: END
+ /// Included only in requests /// /// The token to cancel the operation with /// global::System.Threading.Tasks.Task> EmbedAsync( + global::System.Collections.Generic.IList texts, + string model, + global::System.Collections.Generic.IList embeddingTypes, + global::Cohere.EmbedRequestTruncate truncate, string? xClientName = default, - global::System.Collections.Generic.IList? texts = default, - string? model = default, global::Cohere.EmbedInputType? inputType = default, - global::System.Collections.Generic.IList? embeddingTypes = default, - global::Cohere.EmbedRequestTruncate? truncate = global::Cohere.EmbedRequestTruncate.END, global::System.Threading.CancellationToken cancellationToken = default); } } \ No newline at end of file diff --git a/src/libs/Cohere/Generated/Cohere.ICohereApi.Embedv2.g.cs b/src/libs/Cohere/Generated/Cohere.ICohereApi.Embedv2.g.cs index 994b7b6..e13c77c 100644 --- a/src/libs/Cohere/Generated/Cohere.ICohereApi.Embedv2.g.cs +++ b/src/libs/Cohere/Generated/Cohere.ICohereApi.Embedv2.g.cs @@ -27,7 +27,8 @@ public partial interface ICohereApi /// /// /// - /// An array of strings for the model to embed. Maximum number of texts per call is `96`. We recommend reducing the length of each text to be under `512` tokens for optimal quality. + /// An array of strings for the model to embed. Maximum number of texts per call is `96`. We recommend reducing the length of each text to be under `512` tokens for optimal quality.
+ /// Included only in requests /// /// /// Defaults to embed-english-v2.0
@@ -39,7 +40,8 @@ public partial interface ICohereApi /// * `embed-multilingual-light-v3.0` 384
/// * `embed-english-v2.0` 4096
/// * `embed-english-light-v2.0` 1024
- /// * `embed-multilingual-v2.0` 768 + /// * `embed-multilingual-v2.0` 768
+ /// Included only in requests /// /// /// Specifies the type of input passed to the model. Required for embedding models v3 and higher.
@@ -55,23 +57,25 @@ public partial interface ICohereApi /// * `"int8"`: Use this when you want to get back signed int8 embeddings. Valid for only v3 models.
/// * `"uint8"`: Use this when you want to get back unsigned int8 embeddings. Valid for only v3 models.
/// * `"binary"`: Use this when you want to get back signed binary embeddings. Valid for only v3 models.
- /// * `"ubinary"`: Use this when you want to get back unsigned binary embeddings. Valid for only v3 models. + /// * `"ubinary"`: Use this when you want to get back unsigned binary embeddings. Valid for only v3 models.
+ /// Included only in requests /// /// /// One of `NONE|START|END` to specify how the API will handle inputs longer than the maximum token length.
/// Passing `START` will discard the start of the input. `END` will discard the end of the input. In both cases, input is discarded until the remaining input is exactly the maximum input token length for the model.
/// If `NONE` is selected, when the input exceeds the maximum input token length an error will be returned.
- /// Default Value: END + /// Default Value: END
+ /// Included only in requests /// /// The token to cancel the operation with /// global::System.Threading.Tasks.Task Embedv2Async( + global::System.Collections.Generic.IList texts, string model, + global::System.Collections.Generic.IList embeddingTypes, + global::Cohere.Embedv2RequestTruncate truncate, string? xClientName = default, - global::System.Collections.Generic.IList? texts = default, global::Cohere.EmbedInputType? inputType = default, - global::System.Collections.Generic.IList? embeddingTypes = default, - global::Cohere.Embedv2RequestTruncate? truncate = global::Cohere.Embedv2RequestTruncate.END, global::System.Threading.CancellationToken cancellationToken = default); } } \ No newline at end of file diff --git a/src/libs/Cohere/Generated/Cohere.ICohereApi.Generate.g.cs b/src/libs/Cohere/Generated/Cohere.ICohereApi.Generate.g.cs index 51a11cc..b2eb682 100644 --- a/src/libs/Cohere/Generated/Cohere.ICohereApi.Generate.g.cs +++ b/src/libs/Cohere/Generated/Cohere.ICohereApi.Generate.g.cs @@ -31,14 +31,17 @@ public partial interface ICohereApi /// /// The input text that serves as the starting point for generating the response.
/// Note: The prompt will be pre-processed and modified before reaching the model.
+ /// Included only in requests
/// Example: Please explain to me how LLMs work /// /// /// The identifier of the model to generate with. Currently available models are `command` (default), `command-nightly` (experimental), `command-light`, and `command-light-nightly` (experimental).
- /// Smaller, "light" models are faster, while larger models will perform better. [Custom models](/docs/training-custom-models) can also be supplied with their full ID. + /// Smaller, "light" models are faster, while larger models will perform better. [Custom models](/docs/training-custom-models) can also be supplied with their full ID.
+ /// Included only in requests /// /// - /// The maximum number of generations that will be returned. Defaults to `1`, min value of `1`, max value of `5`. + /// The maximum number of generations that will be returned. Defaults to `1`, min value of `1`, max value of `5`.
+ /// Included only in requests /// /// /// When `true`, the response will be a JSON stream of events. Streaming is beneficial for user interfaces that render the contents of the response piece by piece, as it gets generated.
@@ -51,17 +54,20 @@ public partial interface ICohereApi /// /// The maximum number of tokens the model will generate as part of the response. Note: Setting a low value may result in incomplete generations.
/// This parameter is off by default, and if it's not specified, the model will continue generating until it emits an EOS completion token. See [BPE Tokens](/bpe-tokens-wiki) for more details.
- /// Can only be set to `0` if `return_likelihoods` is set to `ALL` to get the likelihood of the prompt. + /// Can only be set to `0` if `return_likelihoods` is set to `ALL` to get the likelihood of the prompt.
+ /// Included only in requests /// /// /// One of `NONE|START|END` to specify how the API will handle inputs longer than the maximum token length.
/// Passing `START` will discard the start of the input. `END` will discard the end of the input. In both cases, input is discarded until the remaining input is exactly the maximum input token length for the model.
/// If `NONE` is selected, when the input exceeds the maximum input token length an error will be returned.
- /// Default Value: END + /// Default Value: END
+ /// Included only in requests /// /// /// A non-negative float that tunes the degree of randomness in generation. Lower temperatures mean less random generations. See [Temperature](/temperature-wiki) for more details.
- /// Defaults to `0.75`, min value of `0.0`, max value of `5.0`. + /// Defaults to `0.75`, min value of `0.0`, max value of `5.0`.
+ /// Included only in requests /// /// /// If specified, the backend will make a best effort to sample tokens
@@ -73,30 +79,37 @@ public partial interface ICohereApi /// /// Identifier of a custom preset. A preset is a combination of parameters, such as prompt, temperature etc. You can create presets in the [playground](https://dashboard.cohere.com/playground/generate).
/// When a preset is specified, the `prompt` parameter becomes optional, and any included parameters will override the preset's parameters.
+ /// Included only in requests
/// Example: my-preset-a58sbd /// /// - /// The generated text will be cut at the beginning of the earliest occurrence of an end sequence. The sequence will be excluded from the text. + /// The generated text will be cut at the beginning of the earliest occurrence of an end sequence. The sequence will be excluded from the text.
+ /// Included only in requests /// /// - /// The generated text will be cut at the end of the earliest occurrence of a stop sequence. The sequence will be included the text. + /// The generated text will be cut at the end of the earliest occurrence of a stop sequence. The sequence will be included the text.
+ /// Included only in requests /// /// /// Ensures only the top `k` most likely tokens are considered for generation at each step.
- /// Defaults to `0`, min value of `0`, max value of `500`. + /// Defaults to `0`, min value of `0`, max value of `500`.
+ /// Included only in requests /// /// /// Ensures that only the most likely tokens, with total probability mass of `p`, are considered for generation at each step. If both `k` and `p` are enabled, `p` acts after `k`.
- /// Defaults to `0.75`. min value of `0.01`, max value of `0.99`. + /// Defaults to `0.75`. min value of `0.01`, max value of `0.99`.
+ /// Included only in requests /// /// /// Used to reduce repetitiveness of generated tokens. The higher the value, the stronger a penalty is applied to previously present tokens, proportional to how many times they have already appeared in the prompt or prior generation.
- /// Using `frequency_penalty` in combination with `presence_penalty` is not supported on newer models. + /// Using `frequency_penalty` in combination with `presence_penalty` is not supported on newer models.
+ /// Included only in requests /// /// /// Defaults to `0.0`, min value of `0.0`, max value of `1.0`.
/// Can be used to reduce repetitiveness of generated tokens. Similar to `frequency_penalty`, except that this penalty is applied equally to all tokens that have already appeared, regardless of their exact frequencies.
- /// Using `frequency_penalty` in combination with `presence_penalty` is not supported on newer models. + /// Using `frequency_penalty` in combination with `presence_penalty` is not supported on newer models.
+ /// Included only in requests /// /// /// One of `GENERATION|ALL|NONE` to specify how and if the token likelihoods are returned with the response. Defaults to `NONE`.
@@ -111,21 +124,21 @@ public partial interface ICohereApi /// global::System.Threading.Tasks.Task GenerateAsync( string prompt, + string model, + int numGenerations, + int maxTokens, + global::Cohere.GenerateRequestTruncate truncate, + double temperature, + string preset, + global::System.Collections.Generic.IList endSequences, + global::System.Collections.Generic.IList stopSequences, + int k, + double p, + double frequencyPenalty, + double presencePenalty, string? xClientName = default, - string? model = default, - int? numGenerations = default, bool? stream = default, - int? maxTokens = default, - global::Cohere.GenerateRequestTruncate? truncate = global::Cohere.GenerateRequestTruncate.END, - double? temperature = default, int? seed = default, - string? preset = default, - global::System.Collections.Generic.IList? endSequences = default, - global::System.Collections.Generic.IList? stopSequences = default, - int? k = default, - double? p = default, - double? frequencyPenalty = default, - double? presencePenalty = default, global::Cohere.GenerateRequestReturnLikelihoods? returnLikelihoods = global::Cohere.GenerateRequestReturnLikelihoods.NONE, bool? rawPrompting = default, global::System.Threading.CancellationToken cancellationToken = default); diff --git a/src/libs/Cohere/Generated/Cohere.ICohereApi.Summarize.g.cs b/src/libs/Cohere/Generated/Cohere.ICohereApi.Summarize.g.cs index cc72a07..dd63f46 100644 --- a/src/libs/Cohere/Generated/Cohere.ICohereApi.Summarize.g.cs +++ b/src/libs/Cohere/Generated/Cohere.ICohereApi.Summarize.g.cs @@ -29,41 +29,48 @@ public partial interface ICohereApi /// /// /// - /// The text to generate a summary for. Can be up to 100,000 characters long. Currently the only supported language is English. + /// The text to generate a summary for. Can be up to 100,000 characters long. Currently the only supported language is English.
+ /// Included only in requests /// /// /// One of `short`, `medium`, `long`, or `auto` defaults to `auto`. Indicates the approximate length of the summary. If `auto` is selected, the best option will be picked based on the input text.
- /// Default Value: medium + /// Default Value: medium
+ /// Included only in requests /// /// /// One of `paragraph`, `bullets`, or `auto`, defaults to `auto`. Indicates the style in which the summary will be delivered - in a free form paragraph or in bullet points. If `auto` is selected, the best option will be picked based on the input text.
- /// Default Value: paragraph + /// Default Value: paragraph
+ /// Included only in requests /// /// - /// The identifier of the model to generate the summary with. Currently available models are `command` (default), `command-nightly` (experimental), `command-light`, and `command-light-nightly` (experimental). Smaller, "light" models are faster, while larger models will perform better. + /// The identifier of the model to generate the summary with. Currently available models are `command` (default), `command-nightly` (experimental), `command-light`, and `command-light-nightly` (experimental). Smaller, "light" models are faster, while larger models will perform better.
+ /// Included only in requests /// /// /// One of `low`, `medium`, `high`, or `auto`, defaults to `auto`. Controls how close to the original text the summary is. `high` extractiveness summaries will lean towards reusing sentences verbatim, while `low` extractiveness summaries will tend to paraphrase more. If `auto` is selected, the best option will be picked based on the input text.
- /// Default Value: low + /// Default Value: low
+ /// Included only in requests /// /// /// Ranges from 0 to 5. Controls the randomness of the output. Lower values tend to generate more “predictable” output, while higher values tend to generate more “creative” output. The sweet spot is typically between 0 and 1.
- /// Default Value: 0.3 + /// Default Value: 0.3
+ /// Included only in requests /// /// - /// A free-form instruction for modifying how the summaries get generated. Should complete the sentence "Generate a summary _". Eg. "focusing on the next steps" or "written by Yoda" + /// A free-form instruction for modifying how the summaries get generated. Should complete the sentence "Generate a summary _". Eg. "focusing on the next steps" or "written by Yoda"
+ /// Included only in requests /// /// The token to cancel the operation with /// global::System.Threading.Tasks.Task SummarizeAsync( string text, + global::Cohere.SummarizeRequestLength length, + global::Cohere.SummarizeRequestFormat format, + string model, + global::Cohere.SummarizeRequestExtractiveness extractiveness, + double temperature, + string additionalCommand, string? xClientName = default, - global::Cohere.SummarizeRequestLength? length = global::Cohere.SummarizeRequestLength.Medium, - global::Cohere.SummarizeRequestFormat? format = global::Cohere.SummarizeRequestFormat.Paragraph, - string? model = default, - global::Cohere.SummarizeRequestExtractiveness? extractiveness = global::Cohere.SummarizeRequestExtractiveness.Low, - double? temperature = 0.3, - string? additionalCommand = default, global::System.Threading.CancellationToken cancellationToken = default); } } \ No newline at end of file diff --git a/src/libs/Cohere/Generated/Cohere.IFinetuningClient.CreateFinetunedModel.g.cs b/src/libs/Cohere/Generated/Cohere.IFinetuningClient.CreateFinetunedModel.g.cs index d487c43..191e968 100644 --- a/src/libs/Cohere/Generated/Cohere.IFinetuningClient.CreateFinetunedModel.g.cs +++ b/src/libs/Cohere/Generated/Cohere.IFinetuningClient.CreateFinetunedModel.g.cs @@ -20,18 +20,9 @@ public partial interface IFinetuningClient /// Trains and deploys a fine-tuned model. /// /// - /// - /// read-only. FinetunedModel ID. - /// /// /// FinetunedModel name (e.g. `foobar`). /// - /// - /// read-only. User ID of the creator. - /// - /// - /// read-only. Organization ID. - /// /// /// The configuration used for fine-tuning. /// @@ -48,32 +39,13 @@ public partial interface IFinetuningClient /// - STATUS_QUEUED: The fine-tuned model is queued for training.
/// Default Value: STATUS_UNSPECIFIED /// - /// - /// read-only. Creation timestamp. - /// - /// - /// read-only. Latest update timestamp. - /// - /// - /// read-only. Timestamp for the completed fine-tuning. - /// - /// - /// read-only. Timestamp for the latest request to this fine-tuned model. - /// /// The token to cancel the operation with /// global::System.Threading.Tasks.Task CreateFinetunedModelAsync( string name, global::Cohere.Settings settings, string? xClientName = default, - string? id = default, - string? creatorId = default, - string? organizationId = default, global::Cohere.Status? status = global::Cohere.Status.STATUSUNSPECIFIED, - global::System.DateTime? createdAt = default, - global::System.DateTime? updatedAt = default, - global::System.DateTime? completedAt = default, - global::System.DateTime? lastUsed = default, global::System.Threading.CancellationToken cancellationToken = default); } } \ No newline at end of file diff --git a/src/libs/Cohere/Generated/Cohere.IFinetuningClient.UpdateFinetunedModel.g.cs b/src/libs/Cohere/Generated/Cohere.IFinetuningClient.UpdateFinetunedModel.g.cs index 79df9a5..854ebe4 100644 --- a/src/libs/Cohere/Generated/Cohere.IFinetuningClient.UpdateFinetunedModel.g.cs +++ b/src/libs/Cohere/Generated/Cohere.IFinetuningClient.UpdateFinetunedModel.g.cs @@ -26,12 +26,6 @@ public partial interface IFinetuningClient /// /// FinetunedModel name (e.g. `foobar`). /// - /// - /// User ID of the creator. - /// - /// - /// Organization ID. - /// /// /// The configuration used for fine-tuning. /// @@ -48,18 +42,6 @@ public partial interface IFinetuningClient /// - STATUS_QUEUED: The fine-tuned model is queued for training.
/// Default Value: STATUS_UNSPECIFIED /// - /// - /// Creation timestamp. - /// - /// - /// Latest update timestamp. - /// - /// - /// Timestamp for the completed fine-tuning. - /// - /// - /// Timestamp for the latest request to this fine-tuned model. - /// /// The token to cancel the operation with /// global::System.Threading.Tasks.Task UpdateFinetunedModelAsync( @@ -67,13 +49,7 @@ public partial interface IFinetuningClient string name, global::Cohere.Settings settings, string? xClientName = default, - string? creatorId = default, - string? organizationId = default, global::Cohere.Status? status = global::Cohere.Status.STATUSUNSPECIFIED, - global::System.DateTime? createdAt = default, - global::System.DateTime? updatedAt = default, - global::System.DateTime? completedAt = default, - global::System.DateTime? lastUsed = default, global::System.Threading.CancellationToken cancellationToken = default); } } \ No newline at end of file diff --git a/src/libs/Cohere/Generated/Cohere.Models.BaseModel.g.cs b/src/libs/Cohere/Generated/Cohere.Models.BaseModel.g.cs index 92952bc..f99019f 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.BaseModel.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.BaseModel.g.cs @@ -15,7 +15,8 @@ public sealed partial class BaseModel public string? Name { get; set; } /// - /// read-only. The version of the base model. + /// read-only. The version of the base model.
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("version")] public string? Version { get; set; } diff --git a/src/libs/Cohere/Generated/Cohere.Models.ChatRequest.g.cs b/src/libs/Cohere/Generated/Cohere.Models.ChatRequest.g.cs index 9151e51..0004e13 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.ChatRequest.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.ChatRequest.g.cs @@ -127,7 +127,8 @@ public sealed partial class ChatRequest /// /// The maximum number of tokens the model will generate as part of the response. Note: Setting a low value may result in incomplete generations.
- /// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments + /// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("max_tokens")] public int? MaxTokens { get; set; } @@ -135,7 +136,8 @@ public sealed partial class ChatRequest /// /// The maximum number of input tokens to send to the model. If not specified, `max_input_tokens` is the model's context length limit minus a small buffer.
/// Input will be truncated according to the `prompt_truncation` parameter.
- /// Compatible Deployments: Cohere Platform + /// Compatible Deployments: Cohere Platform
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("max_input_tokens")] public int? MaxInputTokens { get; set; } @@ -144,7 +146,8 @@ public sealed partial class ChatRequest /// Ensures only the top `k` most likely tokens are considered for generation at each step.
/// Defaults to `0`, min value of `0`, max value of `500`.
/// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
- /// Default Value: 0 + /// Default Value: 0
+ /// Included only in requests /// [global::System.Text.Json.Serialization.JsonPropertyName("k")] public int? K { get; set; } = 0; @@ -153,7 +156,8 @@ public sealed partial class ChatRequest /// Ensures that only the most likely tokens, with total probability mass of `p`, are considered for generation at each step. If both `k` and `p` are enabled, `p` acts after `k`.
/// Defaults to `0.75`. min value of `0.01`, max value of `0.99`.
/// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
- /// Default Value: 0.75 + /// Default Value: 0.75
+ /// Included only in requests /// [global::System.Text.Json.Serialization.JsonPropertyName("p")] public double? P { get; set; } = 0.75; @@ -170,7 +174,8 @@ public sealed partial class ChatRequest /// /// A list of up to 5 strings that the model will use to stop generation. If the model generates a string that matches any of the strings in the list, it will stop generating tokens and return the generated text up to that point not including the stop sequence.
- /// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments + /// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("stop_sequences")] public global::System.Collections.Generic.IList? StopSequences { get; set; } @@ -178,7 +183,8 @@ public sealed partial class ChatRequest /// /// Defaults to `0.0`, min value of `0.0`, max value of `1.0`.
/// Used to reduce repetitiveness of generated tokens. The higher the value, the stronger a penalty is applied to previously present tokens, proportional to how many times they have already appeared in the prompt or prior generation.
- /// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments + /// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("frequency_penalty")] public double? FrequencyPenalty { get; set; } @@ -186,7 +192,8 @@ public sealed partial class ChatRequest /// /// Defaults to `0.0`, min value of `0.0`, max value of `1.0`.
/// Used to reduce repetitiveness of generated tokens. Similar to `frequency_penalty`, except that this penalty is applied equally to all tokens that have already appeared, regardless of their exact frequencies.
- /// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments + /// Compatible Deployments: Cohere Platform, Azure, AWS Sagemaker/Bedrock, Private Deployments
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("presence_penalty")] public double? PresencePenalty { get; set; } diff --git a/src/libs/Cohere/Generated/Cohere.Models.ClassifyExample.g.cs b/src/libs/Cohere/Generated/Cohere.Models.ClassifyExample.g.cs index b727a94..bf7455e 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.ClassifyExample.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.ClassifyExample.g.cs @@ -4,18 +4,18 @@ namespace Cohere { /// - /// + /// Included only in requests /// public sealed partial class ClassifyExample { /// - /// + /// Included only in requests /// [global::System.Text.Json.Serialization.JsonPropertyName("text")] public string? Text { get; set; } /// - /// + /// Included only in requests /// [global::System.Text.Json.Serialization.JsonPropertyName("label")] public string? Label { get; set; } diff --git a/src/libs/Cohere/Generated/Cohere.Models.ClassifyRequest.g.cs b/src/libs/Cohere/Generated/Cohere.Models.ClassifyRequest.g.cs index 460f286..242d0d2 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.ClassifyRequest.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.ClassifyRequest.g.cs @@ -11,27 +11,30 @@ public sealed partial class ClassifyRequest /// /// A list of up to 96 texts to be classified. Each one must be a non-empty string.
/// There is, however, no consistent, universal limit to the length a particular input can be. We perform classification on the first `x` tokens of each input, and `x` varies depending on which underlying model is powering classification. The maximum token length for each model is listed in the "max tokens" column [here](https://docs.cohere.com/docs/models).
- /// Note: by default the `truncate` parameter is set to `END`, so tokens exceeding the limit will be automatically dropped. This behavior can be disabled by setting `truncate` to `NONE`, which will result in validation errors for longer texts. + /// Note: by default the `truncate` parameter is set to `END`, so tokens exceeding the limit will be automatically dropped. This behavior can be disabled by setting `truncate` to `NONE`, which will result in validation errors for longer texts.
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("inputs")] - [global::System.Text.Json.Serialization.JsonRequired] - public required global::System.Collections.Generic.IList Inputs { get; set; } + public global::System.Collections.Generic.IList? Inputs { get; set; } /// /// An array of examples to provide context to the model. Each example is a text string and its associated label/class. Each unique label requires at least 2 examples associated with it; the maximum number of examples is 2500, and each example has a maximum length of 512 tokens. The values should be structured as `{text: "...",label: "..."}`.
- /// Note: [Fine-tuned Models](https://docs.cohere.com/docs/classify-fine-tuning) trained on classification examples don't require the `examples` parameter to be passed in explicitly. + /// Note: [Fine-tuned Models](https://docs.cohere.com/docs/classify-fine-tuning) trained on classification examples don't require the `examples` parameter to be passed in explicitly.
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("examples")] public global::System.Collections.Generic.IList? Examples { get; set; } /// - /// The identifier of the model. Currently available models are `embed-multilingual-v2.0`, `embed-english-light-v2.0`, and `embed-english-v2.0` (default). Smaller "light" models are faster, while larger models will perform better. [Fine-tuned models](https://docs.cohere.com/docs/fine-tuning) can also be supplied with their full ID. + /// The identifier of the model. Currently available models are `embed-multilingual-v2.0`, `embed-english-light-v2.0`, and `embed-english-v2.0` (default). Smaller "light" models are faster, while larger models will perform better. [Fine-tuned models](https://docs.cohere.com/docs/fine-tuning) can also be supplied with their full ID.
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("model")] public string? Model { get; set; } /// /// The ID of a custom playground preset. You can create presets in the [playground](https://dashboard.cohere.com/playground/classify?model=large). If you use a preset, all other parameters become optional, and any included parameters will override the preset's parameters.
+ /// Included only in requests
/// Example: my-preset-a58sbd ///
[global::System.Text.Json.Serialization.JsonPropertyName("preset")] @@ -41,7 +44,8 @@ public sealed partial class ClassifyRequest /// One of `NONE|START|END` to specify how the API will handle inputs longer than the maximum token length.
/// Passing `START` will discard the start of the input. `END` will discard the end of the input. In both cases, input is discarded until the remaining input is exactly the maximum input token length for the model.
/// If `NONE` is selected, when the input exceeds the maximum input token length an error will be returned.
- /// Default Value: END + /// Default Value: END
+ /// Included only in requests /// [global::System.Text.Json.Serialization.JsonPropertyName("truncate")] [global::System.Text.Json.Serialization.JsonConverter(typeof(global::Cohere.JsonConverters.ClassifyRequestTruncateJsonConverter))] diff --git a/src/libs/Cohere/Generated/Cohere.Models.ClassifyRequestTruncate.g.cs b/src/libs/Cohere/Generated/Cohere.Models.ClassifyRequestTruncate.g.cs index 9105f82..24d6b0f 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.ClassifyRequestTruncate.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.ClassifyRequestTruncate.g.cs @@ -7,7 +7,8 @@ namespace Cohere /// One of `NONE|START|END` to specify how the API will handle inputs longer than the maximum token length.
/// Passing `START` will discard the start of the input. `END` will discard the end of the input. In both cases, input is discarded until the remaining input is exactly the maximum input token length for the model.
/// If `NONE` is selected, when the input exceeds the maximum input token length an error will be returned.
- /// Default Value: END + /// Default Value: END
+ /// Included only in requests /// public enum ClassifyRequestTruncate { diff --git a/src/libs/Cohere/Generated/Cohere.Models.ClassifyResponse.g.cs b/src/libs/Cohere/Generated/Cohere.Models.ClassifyResponse.g.cs index 3984198..b444212 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.ClassifyResponse.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.ClassifyResponse.g.cs @@ -9,18 +9,16 @@ namespace Cohere public sealed partial class ClassifyResponse { /// - /// + /// Included only in responses /// [global::System.Text.Json.Serialization.JsonPropertyName("id")] - [global::System.Text.Json.Serialization.JsonRequired] - public required string Id { get; set; } + public string Id { get; set; } = default!; /// - /// + /// Included only in responses /// [global::System.Text.Json.Serialization.JsonPropertyName("classifications")] - [global::System.Text.Json.Serialization.JsonRequired] - public required global::System.Collections.Generic.IList Classifications { get; set; } + public global::System.Collections.Generic.IList Classifications { get; set; } = default!; /// /// diff --git a/src/libs/Cohere/Generated/Cohere.Models.ClassifyResponseClassification.g.cs b/src/libs/Cohere/Generated/Cohere.Models.ClassifyResponseClassification.g.cs index 1b3077a..515cfd3 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.ClassifyResponseClassification.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.ClassifyResponseClassification.g.cs @@ -4,65 +4,67 @@ namespace Cohere { /// - /// + /// Included only in responses /// public sealed partial class ClassifyResponseClassification { /// - /// + /// Included only in responses /// [global::System.Text.Json.Serialization.JsonPropertyName("id")] - [global::System.Text.Json.Serialization.JsonRequired] - public required string Id { get; set; } + public string Id { get; set; } = default!; /// - /// The input text that was classified + /// The input text that was classified
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("input")] public string? Input { get; set; } /// - /// The predicted label for the associated query (only filled for single-label models) + /// The predicted label for the associated query (only filled for single-label models)
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("prediction")] [global::System.Obsolete("This property marked as deprecated.")] public string? Prediction { get; set; } /// - /// An array containing the predicted labels for the associated query (only filled for single-label classification) + /// An array containing the predicted labels for the associated query (only filled for single-label classification)
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("predictions")] - [global::System.Text.Json.Serialization.JsonRequired] - public required global::System.Collections.Generic.IList Predictions { get; set; } + public global::System.Collections.Generic.IList Predictions { get; set; } = default!; /// - /// The confidence score for the top predicted class (only filled for single-label classification) + /// The confidence score for the top predicted class (only filled for single-label classification)
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("confidence")] [global::System.Obsolete("This property marked as deprecated.")] public float? Confidence { get; set; } /// - /// An array containing the confidence scores of all the predictions in the same order + /// An array containing the confidence scores of all the predictions in the same order
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("confidences")] - [global::System.Text.Json.Serialization.JsonRequired] - public required global::System.Collections.Generic.IList Confidences { get; set; } + public global::System.Collections.Generic.IList Confidences { get; set; } = default!; /// - /// A map containing each label and its confidence score according to the classifier. All the confidence scores add up to 1 for single-label classification. For multi-label classification the label confidences are independent of each other, so they don't have to sum up to 1. + /// A map containing each label and its confidence score according to the classifier. All the confidence scores add up to 1 for single-label classification. For multi-label classification the label confidences are independent of each other, so they don't have to sum up to 1.
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("labels")] - [global::System.Text.Json.Serialization.JsonRequired] - public required global::Cohere.ClassifyResponseClassificationLabels Labels { get; set; } + public global::Cohere.ClassifyResponseClassificationLabels Labels { get; set; } = default!; /// - /// The type of classification performed + /// The type of classification performed
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("classification_type")] [global::System.Text.Json.Serialization.JsonConverter(typeof(global::Cohere.JsonConverters.ClassifyResponseClassificationClassificationTypeJsonConverter))] - [global::System.Text.Json.Serialization.JsonRequired] - public required global::Cohere.ClassifyResponseClassificationClassificationType ClassificationType { get; set; } + public global::Cohere.ClassifyResponseClassificationClassificationType ClassificationType { get; set; } = default!; /// /// Additional properties that are not explicitly defined in the schema diff --git a/src/libs/Cohere/Generated/Cohere.Models.ClassifyResponseClassificationClassificationType.g.cs b/src/libs/Cohere/Generated/Cohere.Models.ClassifyResponseClassificationClassificationType.g.cs index f6963fd..1cbef47 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.ClassifyResponseClassificationClassificationType.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.ClassifyResponseClassificationClassificationType.g.cs @@ -4,7 +4,8 @@ namespace Cohere { /// - /// The type of classification performed + /// The type of classification performed
+ /// Included only in responses ///
public enum ClassifyResponseClassificationClassificationType { diff --git a/src/libs/Cohere/Generated/Cohere.Models.ClassifyResponseClassificationLabels.g.cs b/src/libs/Cohere/Generated/Cohere.Models.ClassifyResponseClassificationLabels.g.cs index a1550a2..266e30e 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.ClassifyResponseClassificationLabels.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.ClassifyResponseClassificationLabels.g.cs @@ -4,7 +4,8 @@ namespace Cohere { /// - /// A map containing each label and its confidence score according to the classifier. All the confidence scores add up to 1 for single-label classification. For multi-label classification the label confidences are independent of each other, so they don't have to sum up to 1. + /// A map containing each label and its confidence score according to the classifier. All the confidence scores add up to 1 for single-label classification. For multi-label classification the label confidences are independent of each other, so they don't have to sum up to 1.
+ /// Included only in responses ///
public sealed partial class ClassifyResponseClassificationLabels { diff --git a/src/libs/Cohere/Generated/Cohere.Models.Dataset.g.cs b/src/libs/Cohere/Generated/Cohere.Models.Dataset.g.cs index 61da078..fbec19d 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.Dataset.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.Dataset.g.cs @@ -89,7 +89,7 @@ public sealed partial class Dataset public global::System.Collections.Generic.IList? ValidationWarnings { get; set; } /// - /// + /// Included only in requests /// [global::System.Text.Json.Serialization.JsonPropertyName("parse_info")] public global::Cohere.ParseInfo? ParseInfo { get; set; } diff --git a/src/libs/Cohere/Generated/Cohere.Models.DetokenizeRequest.g.cs b/src/libs/Cohere/Generated/Cohere.Models.DetokenizeRequest.g.cs index 7b895f2..3d7e21d 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.DetokenizeRequest.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.DetokenizeRequest.g.cs @@ -9,11 +9,11 @@ namespace Cohere public sealed partial class DetokenizeRequest { /// - /// The list of tokens to be detokenized. + /// The list of tokens to be detokenized.
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("tokens")] - [global::System.Text.Json.Serialization.JsonRequired] - public required global::System.Collections.Generic.IList Tokens { get; set; } + public global::System.Collections.Generic.IList? Tokens { get; set; } /// /// An optional parameter to provide the model name. This will ensure that the detokenization is done by the tokenizer used by that model. diff --git a/src/libs/Cohere/Generated/Cohere.Models.DetokenizeResponse.g.cs b/src/libs/Cohere/Generated/Cohere.Models.DetokenizeResponse.g.cs index 79d4112..73340ba 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.DetokenizeResponse.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.DetokenizeResponse.g.cs @@ -9,11 +9,11 @@ namespace Cohere public sealed partial class DetokenizeResponse { /// - /// A string representing the list of tokens. + /// A string representing the list of tokens.
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("text")] - [global::System.Text.Json.Serialization.JsonRequired] - public required string Text { get; set; } + public string Text { get; set; } = default!; /// /// diff --git a/src/libs/Cohere/Generated/Cohere.Models.EmbedByTypeResponse.g.cs b/src/libs/Cohere/Generated/Cohere.Models.EmbedByTypeResponse.g.cs index ff4794b..9416888 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.EmbedByTypeResponse.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.EmbedByTypeResponse.g.cs @@ -23,11 +23,11 @@ public sealed partial class EmbedByTypeResponse public required string Id { get; set; } /// - /// An object with different embedding types. The length of each embedding type array will be the same as the length of the original `texts` array. + /// An object with different embedding types. The length of each embedding type array will be the same as the length of the original `texts` array.
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("embeddings")] - [global::System.Text.Json.Serialization.JsonRequired] - public required global::Cohere.EmbedByTypeResponseEmbeddings Embeddings { get; set; } + public global::Cohere.EmbedByTypeResponseEmbeddings Embeddings { get; set; } = default!; /// /// The text entries for which embeddings were returned. diff --git a/src/libs/Cohere/Generated/Cohere.Models.EmbedByTypeResponseEmbeddings.g.cs b/src/libs/Cohere/Generated/Cohere.Models.EmbedByTypeResponseEmbeddings.g.cs index e922197..55cbfd4 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.EmbedByTypeResponseEmbeddings.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.EmbedByTypeResponseEmbeddings.g.cs @@ -4,30 +4,35 @@ namespace Cohere { /// - /// An object with different embedding types. The length of each embedding type array will be the same as the length of the original `texts` array. + /// An object with different embedding types. The length of each embedding type array will be the same as the length of the original `texts` array.
+ /// Included only in responses ///
public sealed partial class EmbedByTypeResponseEmbeddings { /// - /// An array of float embeddings. + /// An array of float embeddings.
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("float")] public global::System.Collections.Generic.IList>? Float { get; set; } /// - /// An array of signed int8 embeddings. Each value is between -128 and 127. + /// An array of signed int8 embeddings. Each value is between -128 and 127.
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("int8")] public global::System.Collections.Generic.IList>? Int8 { get; set; } /// - /// An array of unsigned int8 embeddings. Each value is between 0 and 255. + /// An array of unsigned int8 embeddings. Each value is between 0 and 255.
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("uint8")] public global::System.Collections.Generic.IList>? Uint8 { get; set; } /// - /// An array of packed signed binary embeddings. The length of each binary embedding is 1/8 the length of the float embeddings of the provided model. Each value is between -128 and 127. + /// An array of packed signed binary embeddings. The length of each binary embedding is 1/8 the length of the float embeddings of the provided model. Each value is between -128 and 127.
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("binary")] public global::System.Collections.Generic.IList>? Binary { get; set; } diff --git a/src/libs/Cohere/Generated/Cohere.Models.EmbedFloatsResponse.g.cs b/src/libs/Cohere/Generated/Cohere.Models.EmbedFloatsResponse.g.cs index 5c91c95..9129708 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.EmbedFloatsResponse.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.EmbedFloatsResponse.g.cs @@ -23,11 +23,11 @@ public sealed partial class EmbedFloatsResponse public required string Id { get; set; } /// - /// An array of embeddings, where each embedding is an array of floats. The length of the `embeddings` array will be the same as the length of the original `texts` array. + /// An array of embeddings, where each embedding is an array of floats. The length of the `embeddings` array will be the same as the length of the original `texts` array.
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("embeddings")] - [global::System.Text.Json.Serialization.JsonRequired] - public required global::System.Collections.Generic.IList> Embeddings { get; set; } + public global::System.Collections.Generic.IList> Embeddings { get; set; } = default!; /// /// The text entries for which embeddings were returned. diff --git a/src/libs/Cohere/Generated/Cohere.Models.EmbedRequest.g.cs b/src/libs/Cohere/Generated/Cohere.Models.EmbedRequest.g.cs index 0e89293..657a18a 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.EmbedRequest.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.EmbedRequest.g.cs @@ -9,7 +9,8 @@ namespace Cohere public sealed partial class EmbedRequest { /// - /// An array of strings for the model to embed. Maximum number of texts per call is `96`. We recommend reducing the length of each text to be under `512` tokens for optimal quality. + /// An array of strings for the model to embed. Maximum number of texts per call is `96`. We recommend reducing the length of each text to be under `512` tokens for optimal quality.
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("texts")] public global::System.Collections.Generic.IList? Texts { get; set; } @@ -24,7 +25,8 @@ public sealed partial class EmbedRequest /// * `embed-multilingual-light-v3.0` 384
/// * `embed-english-v2.0` 4096
/// * `embed-english-light-v2.0` 1024
- /// * `embed-multilingual-v2.0` 768 + /// * `embed-multilingual-v2.0` 768
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("model")] public string? Model { get; set; } @@ -47,7 +49,8 @@ public sealed partial class EmbedRequest /// * `"int8"`: Use this when you want to get back signed int8 embeddings. Valid for only v3 models.
/// * `"uint8"`: Use this when you want to get back unsigned int8 embeddings. Valid for only v3 models.
/// * `"binary"`: Use this when you want to get back signed binary embeddings. Valid for only v3 models.
- /// * `"ubinary"`: Use this when you want to get back unsigned binary embeddings. Valid for only v3 models. + /// * `"ubinary"`: Use this when you want to get back unsigned binary embeddings. Valid for only v3 models.
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("embedding_types")] public global::System.Collections.Generic.IList? EmbeddingTypes { get; set; } @@ -56,7 +59,8 @@ public sealed partial class EmbedRequest /// One of `NONE|START|END` to specify how the API will handle inputs longer than the maximum token length.
/// Passing `START` will discard the start of the input. `END` will discard the end of the input. In both cases, input is discarded until the remaining input is exactly the maximum input token length for the model.
/// If `NONE` is selected, when the input exceeds the maximum input token length an error will be returned.
- /// Default Value: END + /// Default Value: END
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("truncate")] [global::System.Text.Json.Serialization.JsonConverter(typeof(global::Cohere.JsonConverters.EmbedRequestTruncateJsonConverter))] diff --git a/src/libs/Cohere/Generated/Cohere.Models.EmbedRequestTruncate.g.cs b/src/libs/Cohere/Generated/Cohere.Models.EmbedRequestTruncate.g.cs index 599b144..079c2a8 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.EmbedRequestTruncate.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.EmbedRequestTruncate.g.cs @@ -7,7 +7,8 @@ namespace Cohere /// One of `NONE|START|END` to specify how the API will handle inputs longer than the maximum token length.
/// Passing `START` will discard the start of the input. `END` will discard the end of the input. In both cases, input is discarded until the remaining input is exactly the maximum input token length for the model.
/// If `NONE` is selected, when the input exceeds the maximum input token length an error will be returned.
- /// Default Value: END + /// Default Value: END
+ /// Included only in requests ///
public enum EmbedRequestTruncate { diff --git a/src/libs/Cohere/Generated/Cohere.Models.EmbeddingType.g.cs b/src/libs/Cohere/Generated/Cohere.Models.EmbeddingType.g.cs index 462320a..6afee1a 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.EmbeddingType.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.EmbeddingType.g.cs @@ -4,7 +4,7 @@ namespace Cohere { /// - /// + /// Included only in requests /// public enum EmbeddingType { diff --git a/src/libs/Cohere/Generated/Cohere.Models.Embedv2Request.g.cs b/src/libs/Cohere/Generated/Cohere.Models.Embedv2Request.g.cs index 6f07455..b4ab575 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.Embedv2Request.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.Embedv2Request.g.cs @@ -9,7 +9,8 @@ namespace Cohere public sealed partial class Embedv2Request { /// - /// An array of strings for the model to embed. Maximum number of texts per call is `96`. We recommend reducing the length of each text to be under `512` tokens for optimal quality. + /// An array of strings for the model to embed. Maximum number of texts per call is `96`. We recommend reducing the length of each text to be under `512` tokens for optimal quality.
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("texts")] public global::System.Collections.Generic.IList? Texts { get; set; } @@ -24,11 +25,11 @@ public sealed partial class Embedv2Request /// * `embed-multilingual-light-v3.0` 384
/// * `embed-english-v2.0` 4096
/// * `embed-english-light-v2.0` 1024
- /// * `embed-multilingual-v2.0` 768 + /// * `embed-multilingual-v2.0` 768
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("model")] - [global::System.Text.Json.Serialization.JsonRequired] - public required string Model { get; set; } + public string? Model { get; set; } /// /// Specifies the type of input passed to the model. Required for embedding models v3 and higher.
@@ -48,7 +49,8 @@ public sealed partial class Embedv2Request /// * `"int8"`: Use this when you want to get back signed int8 embeddings. Valid for only v3 models.
/// * `"uint8"`: Use this when you want to get back unsigned int8 embeddings. Valid for only v3 models.
/// * `"binary"`: Use this when you want to get back signed binary embeddings. Valid for only v3 models.
- /// * `"ubinary"`: Use this when you want to get back unsigned binary embeddings. Valid for only v3 models. + /// * `"ubinary"`: Use this when you want to get back unsigned binary embeddings. Valid for only v3 models.
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("embedding_types")] public global::System.Collections.Generic.IList? EmbeddingTypes { get; set; } @@ -57,7 +59,8 @@ public sealed partial class Embedv2Request /// One of `NONE|START|END` to specify how the API will handle inputs longer than the maximum token length.
/// Passing `START` will discard the start of the input. `END` will discard the end of the input. In both cases, input is discarded until the remaining input is exactly the maximum input token length for the model.
/// If `NONE` is selected, when the input exceeds the maximum input token length an error will be returned.
- /// Default Value: END + /// Default Value: END
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("truncate")] [global::System.Text.Json.Serialization.JsonConverter(typeof(global::Cohere.JsonConverters.Embedv2RequestTruncateJsonConverter))] diff --git a/src/libs/Cohere/Generated/Cohere.Models.Embedv2RequestTruncate.g.cs b/src/libs/Cohere/Generated/Cohere.Models.Embedv2RequestTruncate.g.cs index 3f21d4d..9972678 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.Embedv2RequestTruncate.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.Embedv2RequestTruncate.g.cs @@ -7,7 +7,8 @@ namespace Cohere /// One of `NONE|START|END` to specify how the API will handle inputs longer than the maximum token length.
/// Passing `START` will discard the start of the input. `END` will discard the end of the input. In both cases, input is discarded until the remaining input is exactly the maximum input token length for the model.
/// If `NONE` is selected, when the input exceeds the maximum input token length an error will be returned.
- /// Default Value: END + /// Default Value: END
+ /// Included only in requests /// public enum Embedv2RequestTruncate { diff --git a/src/libs/Cohere/Generated/Cohere.Models.FinetunedModel.g.cs b/src/libs/Cohere/Generated/Cohere.Models.FinetunedModel.g.cs index 2d2c6ac..f46bcde 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.FinetunedModel.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.FinetunedModel.g.cs @@ -9,7 +9,8 @@ namespace Cohere public sealed partial class FinetunedModel { /// - /// read-only. FinetunedModel ID. + /// read-only. FinetunedModel ID.
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("id")] public string? Id { get; set; } @@ -22,13 +23,15 @@ public sealed partial class FinetunedModel public required string Name { get; set; } /// - /// read-only. User ID of the creator. + /// read-only. User ID of the creator.
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("creator_id")] public string? CreatorId { get; set; } /// - /// read-only. Organization ID. + /// read-only. Organization ID.
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("organization_id")] public string? OrganizationId { get; set; } @@ -58,25 +61,29 @@ public sealed partial class FinetunedModel public global::Cohere.Status? Status { get; set; } = global::Cohere.Status.STATUSUNSPECIFIED; /// - /// read-only. Creation timestamp. + /// read-only. Creation timestamp.
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("created_at")] public global::System.DateTime? CreatedAt { get; set; } /// - /// read-only. Latest update timestamp. + /// read-only. Latest update timestamp.
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("updated_at")] public global::System.DateTime? UpdatedAt { get; set; } /// - /// read-only. Timestamp for the completed fine-tuning. + /// read-only. Timestamp for the completed fine-tuning.
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("completed_at")] public global::System.DateTime? CompletedAt { get; set; } /// - /// read-only. Timestamp for the latest request to this fine-tuned model. + /// read-only. Timestamp for the latest request to this fine-tuned model.
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("last_used")] public global::System.DateTime? LastUsed { get; set; } diff --git a/src/libs/Cohere/Generated/Cohere.Models.GenerateRequest.g.cs b/src/libs/Cohere/Generated/Cohere.Models.GenerateRequest.g.cs index 622f3fc..48221b6 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.GenerateRequest.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.GenerateRequest.g.cs @@ -4,28 +4,30 @@ namespace Cohere { /// - /// + /// Included only in requests /// public sealed partial class GenerateRequest { /// /// The input text that serves as the starting point for generating the response.
/// Note: The prompt will be pre-processed and modified before reaching the model.
+ /// Included only in requests
/// Example: Please explain to me how LLMs work ///
[global::System.Text.Json.Serialization.JsonPropertyName("prompt")] - [global::System.Text.Json.Serialization.JsonRequired] - public required string Prompt { get; set; } + public string? Prompt { get; set; } /// /// The identifier of the model to generate with. Currently available models are `command` (default), `command-nightly` (experimental), `command-light`, and `command-light-nightly` (experimental).
- /// Smaller, "light" models are faster, while larger models will perform better. [Custom models](/docs/training-custom-models) can also be supplied with their full ID. + /// Smaller, "light" models are faster, while larger models will perform better. [Custom models](/docs/training-custom-models) can also be supplied with their full ID.
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("model")] public string? Model { get; set; } /// - /// The maximum number of generations that will be returned. Defaults to `1`, min value of `1`, max value of `5`. + /// The maximum number of generations that will be returned. Defaults to `1`, min value of `1`, max value of `5`.
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("num_generations")] public int? NumGenerations { get; set; } @@ -44,7 +46,8 @@ public sealed partial class GenerateRequest /// /// The maximum number of tokens the model will generate as part of the response. Note: Setting a low value may result in incomplete generations.
/// This parameter is off by default, and if it's not specified, the model will continue generating until it emits an EOS completion token. See [BPE Tokens](/bpe-tokens-wiki) for more details.
- /// Can only be set to `0` if `return_likelihoods` is set to `ALL` to get the likelihood of the prompt. + /// Can only be set to `0` if `return_likelihoods` is set to `ALL` to get the likelihood of the prompt.
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("max_tokens")] public int? MaxTokens { get; set; } @@ -53,7 +56,8 @@ public sealed partial class GenerateRequest /// One of `NONE|START|END` to specify how the API will handle inputs longer than the maximum token length.
/// Passing `START` will discard the start of the input. `END` will discard the end of the input. In both cases, input is discarded until the remaining input is exactly the maximum input token length for the model.
/// If `NONE` is selected, when the input exceeds the maximum input token length an error will be returned.
- /// Default Value: END + /// Default Value: END
+ /// Included only in requests /// [global::System.Text.Json.Serialization.JsonPropertyName("truncate")] [global::System.Text.Json.Serialization.JsonConverter(typeof(global::Cohere.JsonConverters.GenerateRequestTruncateJsonConverter))] @@ -61,7 +65,8 @@ public sealed partial class GenerateRequest /// /// A non-negative float that tunes the degree of randomness in generation. Lower temperatures mean less random generations. See [Temperature](/temperature-wiki) for more details.
- /// Defaults to `0.75`, min value of `0.0`, max value of `5.0`. + /// Defaults to `0.75`, min value of `0.0`, max value of `5.0`.
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("temperature")] public double? Temperature { get; set; } @@ -79,40 +84,46 @@ public sealed partial class GenerateRequest /// /// Identifier of a custom preset. A preset is a combination of parameters, such as prompt, temperature etc. You can create presets in the [playground](https://dashboard.cohere.com/playground/generate).
/// When a preset is specified, the `prompt` parameter becomes optional, and any included parameters will override the preset's parameters.
+ /// Included only in requests
/// Example: my-preset-a58sbd ///
[global::System.Text.Json.Serialization.JsonPropertyName("preset")] public string? Preset { get; set; } /// - /// The generated text will be cut at the beginning of the earliest occurrence of an end sequence. The sequence will be excluded from the text. + /// The generated text will be cut at the beginning of the earliest occurrence of an end sequence. The sequence will be excluded from the text.
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("end_sequences")] public global::System.Collections.Generic.IList? EndSequences { get; set; } /// - /// The generated text will be cut at the end of the earliest occurrence of a stop sequence. The sequence will be included the text. + /// The generated text will be cut at the end of the earliest occurrence of a stop sequence. The sequence will be included the text.
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("stop_sequences")] public global::System.Collections.Generic.IList? StopSequences { get; set; } /// /// Ensures only the top `k` most likely tokens are considered for generation at each step.
- /// Defaults to `0`, min value of `0`, max value of `500`. + /// Defaults to `0`, min value of `0`, max value of `500`.
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("k")] public int? K { get; set; } /// /// Ensures that only the most likely tokens, with total probability mass of `p`, are considered for generation at each step. If both `k` and `p` are enabled, `p` acts after `k`.
- /// Defaults to `0.75`. min value of `0.01`, max value of `0.99`. + /// Defaults to `0.75`. min value of `0.01`, max value of `0.99`.
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("p")] public double? P { get; set; } /// /// Used to reduce repetitiveness of generated tokens. The higher the value, the stronger a penalty is applied to previously present tokens, proportional to how many times they have already appeared in the prompt or prior generation.
- /// Using `frequency_penalty` in combination with `presence_penalty` is not supported on newer models. + /// Using `frequency_penalty` in combination with `presence_penalty` is not supported on newer models.
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("frequency_penalty")] public double? FrequencyPenalty { get; set; } @@ -120,7 +131,8 @@ public sealed partial class GenerateRequest /// /// Defaults to `0.0`, min value of `0.0`, max value of `1.0`.
/// Can be used to reduce repetitiveness of generated tokens. Similar to `frequency_penalty`, except that this penalty is applied equally to all tokens that have already appeared, regardless of their exact frequencies.
- /// Using `frequency_penalty` in combination with `presence_penalty` is not supported on newer models. + /// Using `frequency_penalty` in combination with `presence_penalty` is not supported on newer models.
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("presence_penalty")] public double? PresencePenalty { get; set; } diff --git a/src/libs/Cohere/Generated/Cohere.Models.GenerateRequestTruncate.g.cs b/src/libs/Cohere/Generated/Cohere.Models.GenerateRequestTruncate.g.cs index 56d222f..9e87827 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.GenerateRequestTruncate.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.GenerateRequestTruncate.g.cs @@ -7,7 +7,8 @@ namespace Cohere /// One of `NONE|START|END` to specify how the API will handle inputs longer than the maximum token length.
/// Passing `START` will discard the start of the input. `END` will discard the end of the input. In both cases, input is discarded until the remaining input is exactly the maximum input token length for the model.
/// If `NONE` is selected, when the input exceeds the maximum input token length an error will be returned.
- /// Default Value: END + /// Default Value: END
+ /// Included only in requests /// public enum GenerateRequestTruncate { diff --git a/src/libs/Cohere/Generated/Cohere.Models.GetClusterJobResponse.g.cs b/src/libs/Cohere/Generated/Cohere.Models.GetClusterJobResponse.g.cs index 777b488..571c982 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.GetClusterJobResponse.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.GetClusterJobResponse.g.cs @@ -16,37 +16,43 @@ public sealed partial class GetClusterJobResponse public required string JobId { get; set; } /// - /// Time of job creation in RFC3339 format + /// Time of job creation in RFC3339 format
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("created_at")] public global::System.DateTime? CreatedAt { get; set; } /// - /// The input file URL used for the job + /// The input file URL used for the job
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("embeddings_url")] public string? EmbeddingsUrl { get; set; } /// - /// The input dataset ID used for the job + /// The input dataset ID used for the job
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("input_dataset_id")] public string? InputDatasetId { get; set; } /// - /// The parameter used in the job creation. Please refer to the job creation endpoint for more details + /// The parameter used in the job creation. Please refer to the job creation endpoint for more details
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("min_cluster_size")] public int? MinClusterSize { get; set; } /// - /// The parameter used in the job creation. Please refer to the job creation endpoint for more details + /// The parameter used in the job creation. Please refer to the job creation endpoint for more details
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("n_neighbors")] public int? NNeighbors { get; set; } /// - /// The parameter used in the job creation. Please refer to the job creation endpoint for more details + /// The parameter used in the job creation. Please refer to the job creation endpoint for more details
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("is_deterministic")] public bool? IsDeterministic { get; set; } @@ -59,25 +65,29 @@ public sealed partial class GetClusterJobResponse public global::Cohere.GetClusterJobResponseStatus? Status { get; set; } /// - /// A boolean indicating whether the job is in a final state, whether completed or failed + /// A boolean indicating whether the job is in a final state, whether completed or failed
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("is_final_state")] public bool? IsFinalState { get; set; } /// - /// The output file URL for the clusters (signed url that expires) + /// The output file URL for the clusters (signed url that expires)
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("output_clusters_url")] public string? OutputClustersUrl { get; set; } /// - /// The output file URL for the outliers (signed url that expires) + /// The output file URL for the outliers (signed url that expires)
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("output_outliers_url")] public string? OutputOutliersUrl { get; set; } /// - /// The list of cluster summaries for the job + /// The list of cluster summaries for the job
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("clusters")] public global::System.Collections.Generic.IList? Clusters { get; set; } diff --git a/src/libs/Cohere/Generated/Cohere.Models.GetConnectorsLogsResponse.g.cs b/src/libs/Cohere/Generated/Cohere.Models.GetConnectorsLogsResponse.g.cs index 9d45bb4..2253ae9 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.GetConnectorsLogsResponse.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.GetConnectorsLogsResponse.g.cs @@ -16,11 +16,11 @@ public sealed partial class GetConnectorsLogsResponse public required global::System.Collections.Generic.IList ConnectorLogs { get; set; } /// - /// Total number of connector logs + /// Total number of connector logs
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("total")] - [global::System.Text.Json.Serialization.JsonRequired] - public required double Total { get; set; } + public double Total { get; set; } = default!; /// /// Additional properties that are not explicitly defined in the schema diff --git a/src/libs/Cohere/Generated/Cohere.Models.ListConnectorsResponse.g.cs b/src/libs/Cohere/Generated/Cohere.Models.ListConnectorsResponse.g.cs index f5ecd0e..c6d6390 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.ListConnectorsResponse.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.ListConnectorsResponse.g.cs @@ -16,7 +16,8 @@ public sealed partial class ListConnectorsResponse public required global::System.Collections.Generic.IList Connectors { get; set; } /// - /// Total number of connectors. + /// Total number of connectors.
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("total_count")] public double? TotalCount { get; set; } diff --git a/src/libs/Cohere/Generated/Cohere.Models.ParseInfo.g.cs b/src/libs/Cohere/Generated/Cohere.Models.ParseInfo.g.cs index 8eb0d96..a09014c 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.ParseInfo.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.ParseInfo.g.cs @@ -4,18 +4,18 @@ namespace Cohere { /// - /// + /// Included only in requests /// public sealed partial class ParseInfo { /// - /// + /// Included only in requests /// [global::System.Text.Json.Serialization.JsonPropertyName("separator")] public string? Separator { get; set; } /// - /// + /// Included only in requests /// [global::System.Text.Json.Serialization.JsonPropertyName("delimiter")] public string? Delimiter { get; set; } diff --git a/src/libs/Cohere/Generated/Cohere.Models.RerankResponseResult.g.cs b/src/libs/Cohere/Generated/Cohere.Models.RerankResponseResult.g.cs index c0496a9..b9ca225 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.RerankResponseResult.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.RerankResponseResult.g.cs @@ -22,11 +22,11 @@ public sealed partial class RerankResponseResult public required int Index { get; set; } /// - /// Relevance scores are normalized to be in the range `[0, 1]`. Scores close to `1` indicate a high relevance to the query, and scores closer to `0` indicate low relevance. It is not accurate to assume a score of 0.9 means the document is 2x more relevant than a document with a score of 0.45 + /// Relevance scores are normalized to be in the range `[0, 1]`. Scores close to `1` indicate a high relevance to the query, and scores closer to `0` indicate low relevance. It is not accurate to assume a score of 0.9 means the document is 2x more relevant than a document with a score of 0.45
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("relevance_score")] - [global::System.Text.Json.Serialization.JsonRequired] - public required float RelevanceScore { get; set; } + public float RelevanceScore { get; set; } = default!; /// /// Additional properties that are not explicitly defined in the schema diff --git a/src/libs/Cohere/Generated/Cohere.Models.Rerankv2ResponseResult.g.cs b/src/libs/Cohere/Generated/Cohere.Models.Rerankv2ResponseResult.g.cs index 3e82214..a33fb7e 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.Rerankv2ResponseResult.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.Rerankv2ResponseResult.g.cs @@ -22,11 +22,11 @@ public sealed partial class Rerankv2ResponseResult public required int Index { get; set; } /// - /// Relevance scores are normalized to be in the range `[0, 1]`. Scores close to `1` indicate a high relevance to the query, and scores closer to `0` indicate low relevance. It is not accurate to assume a score of 0.9 means the document is 2x more relevant than a document with a score of 0.45 + /// Relevance scores are normalized to be in the range `[0, 1]`. Scores close to `1` indicate a high relevance to the query, and scores closer to `0` indicate low relevance. It is not accurate to assume a score of 0.9 means the document is 2x more relevant than a document with a score of 0.45
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("relevance_score")] - [global::System.Text.Json.Serialization.JsonRequired] - public required float RelevanceScore { get; set; } + public float RelevanceScore { get; set; } = default!; /// /// Additional properties that are not explicitly defined in the schema diff --git a/src/libs/Cohere/Generated/Cohere.Models.Settings.g.cs b/src/libs/Cohere/Generated/Cohere.Models.Settings.g.cs index 9032312..0165c1b 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.Settings.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.Settings.g.cs @@ -29,7 +29,8 @@ public sealed partial class Settings public global::Cohere.Hyperparameters? Hyperparameters { get; set; } /// - /// read-only. Whether the model is single-label or multi-label (only for classification). + /// read-only. Whether the model is single-label or multi-label (only for classification).
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("multi_label")] public bool? MultiLabel { get; set; } diff --git a/src/libs/Cohere/Generated/Cohere.Models.SummarizeRequest.g.cs b/src/libs/Cohere/Generated/Cohere.Models.SummarizeRequest.g.cs index 4463774..e22dda1 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.SummarizeRequest.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.SummarizeRequest.g.cs @@ -9,15 +9,16 @@ namespace Cohere public sealed partial class SummarizeRequest { /// - /// The text to generate a summary for. Can be up to 100,000 characters long. Currently the only supported language is English. + /// The text to generate a summary for. Can be up to 100,000 characters long. Currently the only supported language is English.
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("text")] - [global::System.Text.Json.Serialization.JsonRequired] - public required string Text { get; set; } + public string? Text { get; set; } /// /// One of `short`, `medium`, `long`, or `auto` defaults to `auto`. Indicates the approximate length of the summary. If `auto` is selected, the best option will be picked based on the input text.
- /// Default Value: medium + /// Default Value: medium
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("length")] [global::System.Text.Json.Serialization.JsonConverter(typeof(global::Cohere.JsonConverters.SummarizeRequestLengthJsonConverter))] @@ -25,21 +26,24 @@ public sealed partial class SummarizeRequest /// /// One of `paragraph`, `bullets`, or `auto`, defaults to `auto`. Indicates the style in which the summary will be delivered - in a free form paragraph or in bullet points. If `auto` is selected, the best option will be picked based on the input text.
- /// Default Value: paragraph + /// Default Value: paragraph
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("format")] [global::System.Text.Json.Serialization.JsonConverter(typeof(global::Cohere.JsonConverters.SummarizeRequestFormatJsonConverter))] public global::Cohere.SummarizeRequestFormat? Format { get; set; } = global::Cohere.SummarizeRequestFormat.Paragraph; /// - /// The identifier of the model to generate the summary with. Currently available models are `command` (default), `command-nightly` (experimental), `command-light`, and `command-light-nightly` (experimental). Smaller, "light" models are faster, while larger models will perform better. + /// The identifier of the model to generate the summary with. Currently available models are `command` (default), `command-nightly` (experimental), `command-light`, and `command-light-nightly` (experimental). Smaller, "light" models are faster, while larger models will perform better.
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("model")] public string? Model { get; set; } /// /// One of `low`, `medium`, `high`, or `auto`, defaults to `auto`. Controls how close to the original text the summary is. `high` extractiveness summaries will lean towards reusing sentences verbatim, while `low` extractiveness summaries will tend to paraphrase more. If `auto` is selected, the best option will be picked based on the input text.
- /// Default Value: low + /// Default Value: low
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("extractiveness")] [global::System.Text.Json.Serialization.JsonConverter(typeof(global::Cohere.JsonConverters.SummarizeRequestExtractivenessJsonConverter))] @@ -47,13 +51,15 @@ public sealed partial class SummarizeRequest /// /// Ranges from 0 to 5. Controls the randomness of the output. Lower values tend to generate more “predictable” output, while higher values tend to generate more “creative” output. The sweet spot is typically between 0 and 1.
- /// Default Value: 0.3 + /// Default Value: 0.3
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("temperature")] public double? Temperature { get; set; } = 0.3; /// - /// A free-form instruction for modifying how the summaries get generated. Should complete the sentence "Generate a summary _". Eg. "focusing on the next steps" or "written by Yoda" + /// A free-form instruction for modifying how the summaries get generated. Should complete the sentence "Generate a summary _". Eg. "focusing on the next steps" or "written by Yoda"
+ /// Included only in requests ///
[global::System.Text.Json.Serialization.JsonPropertyName("additional_command")] public string? AdditionalCommand { get; set; } diff --git a/src/libs/Cohere/Generated/Cohere.Models.SummarizeRequestExtractiveness.g.cs b/src/libs/Cohere/Generated/Cohere.Models.SummarizeRequestExtractiveness.g.cs index 0ee1fd8..05ab31a 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.SummarizeRequestExtractiveness.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.SummarizeRequestExtractiveness.g.cs @@ -5,7 +5,8 @@ namespace Cohere { /// /// One of `low`, `medium`, `high`, or `auto`, defaults to `auto`. Controls how close to the original text the summary is. `high` extractiveness summaries will lean towards reusing sentences verbatim, while `low` extractiveness summaries will tend to paraphrase more. If `auto` is selected, the best option will be picked based on the input text.
- /// Default Value: low + /// Default Value: low
+ /// Included only in requests ///
public enum SummarizeRequestExtractiveness { diff --git a/src/libs/Cohere/Generated/Cohere.Models.SummarizeRequestFormat.g.cs b/src/libs/Cohere/Generated/Cohere.Models.SummarizeRequestFormat.g.cs index 4b497bc..3e7960f 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.SummarizeRequestFormat.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.SummarizeRequestFormat.g.cs @@ -5,7 +5,8 @@ namespace Cohere { /// /// One of `paragraph`, `bullets`, or `auto`, defaults to `auto`. Indicates the style in which the summary will be delivered - in a free form paragraph or in bullet points. If `auto` is selected, the best option will be picked based on the input text.
- /// Default Value: paragraph + /// Default Value: paragraph
+ /// Included only in requests ///
public enum SummarizeRequestFormat { diff --git a/src/libs/Cohere/Generated/Cohere.Models.SummarizeRequestLength.g.cs b/src/libs/Cohere/Generated/Cohere.Models.SummarizeRequestLength.g.cs index 7dae069..ca0e6b2 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.SummarizeRequestLength.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.SummarizeRequestLength.g.cs @@ -5,7 +5,8 @@ namespace Cohere { /// /// One of `short`, `medium`, `long`, or `auto` defaults to `auto`. Indicates the approximate length of the summary. If `auto` is selected, the best option will be picked based on the input text.
- /// Default Value: medium + /// Default Value: medium
+ /// Included only in requests ///
public enum SummarizeRequestLength { diff --git a/src/libs/Cohere/Generated/Cohere.Models.TokenizeResponse.g.cs b/src/libs/Cohere/Generated/Cohere.Models.TokenizeResponse.g.cs index 5352022..2000b1b 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.TokenizeResponse.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.TokenizeResponse.g.cs @@ -9,11 +9,11 @@ namespace Cohere public sealed partial class TokenizeResponse { /// - /// An array of tokens, where each token is an integer. + /// An array of tokens, where each token is an integer.
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("tokens")] - [global::System.Text.Json.Serialization.JsonRequired] - public required global::System.Collections.Generic.IList Tokens { get; set; } + public global::System.Collections.Generic.IList Tokens { get; set; } = default!; /// /// diff --git a/src/libs/Cohere/Generated/Cohere.Models.UpdateFinetunedModelRequest.g.cs b/src/libs/Cohere/Generated/Cohere.Models.UpdateFinetunedModelRequest.g.cs index ec367c6..ba351bc 100644 --- a/src/libs/Cohere/Generated/Cohere.Models.UpdateFinetunedModelRequest.g.cs +++ b/src/libs/Cohere/Generated/Cohere.Models.UpdateFinetunedModelRequest.g.cs @@ -16,13 +16,15 @@ public sealed partial class UpdateFinetunedModelRequest public required string Name { get; set; } /// - /// User ID of the creator. + /// User ID of the creator.
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("creator_id")] public string? CreatorId { get; set; } /// - /// Organization ID. + /// Organization ID.
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("organization_id")] public string? OrganizationId { get; set; } @@ -52,25 +54,29 @@ public sealed partial class UpdateFinetunedModelRequest public global::Cohere.Status? Status { get; set; } = global::Cohere.Status.STATUSUNSPECIFIED; /// - /// Creation timestamp. + /// Creation timestamp.
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("created_at")] public global::System.DateTime? CreatedAt { get; set; } /// - /// Latest update timestamp. + /// Latest update timestamp.
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("updated_at")] public global::System.DateTime? UpdatedAt { get; set; } /// - /// Timestamp for the completed fine-tuning. + /// Timestamp for the completed fine-tuning.
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("completed_at")] public global::System.DateTime? CompletedAt { get; set; } /// - /// Timestamp for the latest request to this fine-tuned model. + /// Timestamp for the latest request to this fine-tuned model.
+ /// Included only in responses ///
[global::System.Text.Json.Serialization.JsonPropertyName("last_used")] public global::System.DateTime? LastUsed { get; set; }