Skip to content

Commit

Permalink
fix: Fixed issues with new parameters with default values.
Browse files Browse the repository at this point in the history
  • Loading branch information
HavenDV committed Oct 24, 2024
1 parent 837b253 commit 512d13b
Show file tree
Hide file tree
Showing 88 changed files with 213 additions and 213 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -166,8 +166,8 @@ partial void ProcessCreateAssistantResponseContent(
global::System.Collections.Generic.IList<global::OpenAI.ToolsItem2>? tools = default,
global::OpenAI.CreateAssistantRequestToolResources? toolResources = default,
object? metadata = default,
double? temperature = 1,
double? topP = 1,
double? temperature = default,
double? topP = default,
global::OpenAI.AssistantsApiResponseFormatOption? responseFormat = default,
global::System.Threading.CancellationToken cancellationToken = default)
{
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -204,8 +204,8 @@ partial void ProcessCreateRunResponseContent(
global::System.Collections.Generic.IList<global::OpenAI.CreateMessageRequest>? additionalMessages = default,
global::System.Collections.Generic.IList<global::OpenAI.ToolsItem4>? tools = default,
object? metadata = default,
double? temperature = 1,
double? topP = 1,
double? temperature = default,
double? topP = default,
bool? stream = default,
int? maxPromptTokens = default,
int? maxCompletionTokens = default,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -186,8 +186,8 @@ partial void ProcessCreateThreadAndRunResponseContent(
global::System.Collections.Generic.IList<global::OpenAI.ToolsItem5>? tools = default,
global::OpenAI.CreateThreadAndRunRequestToolResources? toolResources = default,
object? metadata = default,
double? temperature = 1,
double? topP = 1,
double? temperature = default,
double? topP = default,
bool? stream = default,
int? maxPromptTokens = default,
int? maxCompletionTokens = default,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -41,8 +41,8 @@ partial void ProcessListAssistantsResponseContent(
/// <param name="cancellationToken">The token to cancel the operation with</param>
/// <exception cref="global::System.InvalidOperationException"></exception>
public async global::System.Threading.Tasks.Task<global::OpenAI.ListAssistantsResponse> ListAssistantsAsync(
int? limit = 20,
global::OpenAI.ListAssistantsOrder? order = global::OpenAI.ListAssistantsOrder.Desc,
int? limit = default,
global::OpenAI.ListAssistantsOrder? order = default,
string? after = default,
string? before = default,
global::System.Threading.CancellationToken cancellationToken = default)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -48,8 +48,8 @@ partial void ProcessListMessagesResponseContent(
/// <exception cref="global::System.InvalidOperationException"></exception>
public async global::System.Threading.Tasks.Task<global::OpenAI.ListMessagesResponse> ListMessagesAsync(
string threadId,
int? limit = 20,
global::OpenAI.ListMessagesOrder? order = global::OpenAI.ListMessagesOrder.Desc,
int? limit = default,
global::OpenAI.ListMessagesOrder? order = default,
string? after = default,
string? before = default,
string? runId = default,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -52,8 +52,8 @@ partial void ProcessListRunStepsResponseContent(
public async global::System.Threading.Tasks.Task<global::OpenAI.ListRunStepsResponse> ListRunStepsAsync(
string threadId,
string runId,
int? limit = 20,
global::OpenAI.ListRunStepsOrder? order = global::OpenAI.ListRunStepsOrder.Desc,
int? limit = default,
global::OpenAI.ListRunStepsOrder? order = default,
string? after = default,
string? before = default,
global::System.Collections.Generic.IList<global::OpenAI.ListRunStepsIncludeItem>? include = default,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -45,8 +45,8 @@ partial void ProcessListRunsResponseContent(
/// <exception cref="global::System.InvalidOperationException"></exception>
public async global::System.Threading.Tasks.Task<global::OpenAI.ListRunsResponse> ListRunsAsync(
string threadId,
int? limit = 20,
global::OpenAI.ListRunsOrder? order = global::OpenAI.ListRunsOrder.Desc,
int? limit = default,
global::OpenAI.ListRunsOrder? order = default,
string? after = default,
string? before = default,
global::System.Threading.CancellationToken cancellationToken = default)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -173,8 +173,8 @@ partial void ProcessModifyAssistantResponseContent(
global::System.Collections.Generic.IList<global::OpenAI.ToolsItem7>? tools = default,
global::OpenAI.ModifyAssistantRequestToolResources? toolResources = default,
object? metadata = default,
double? temperature = 1,
double? topP = 1,
double? temperature = default,
double? topP = default,
global::OpenAI.AssistantsApiResponseFormatOption? responseFormat = default,
global::System.Threading.CancellationToken cancellationToken = default)
{
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -128,8 +128,8 @@ partial void ProcessCreateSpeechResponseContent(
global::OpenAI.AnyOf<string, global::OpenAI.CreateSpeechRequestModel?> model,
string input,
global::OpenAI.CreateSpeechRequestVoice voice,
global::OpenAI.CreateSpeechRequestResponseFormat? responseFormat = global::OpenAI.CreateSpeechRequestResponseFormat.Mp3,
double? speed = 1,
global::OpenAI.CreateSpeechRequestResponseFormat? responseFormat = default,
double? speed = default,
global::System.Threading.CancellationToken cancellationToken = default)
{
var __request = new global::OpenAI.CreateSpeechRequest
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -88,13 +88,13 @@ partial void ProcessCreateTranscriptionResponseContent(
content: new global::System.Net.Http.StringContent($"{request.Prompt}"),
name: "prompt");
}
if (request.ResponseFormat != global::OpenAI.AudioResponseFormat.Json)
if (request.ResponseFormat != default)
{
__httpRequestContent.Add(
content: new global::System.Net.Http.StringContent($"{request.ResponseFormat?.ToValueString()}"),
name: "response_format");
}
if (request.Temperature != 0)
if (request.Temperature != default)
{
__httpRequestContent.Add(
content: new global::System.Net.Http.StringContent($"{request.Temperature}"),
Expand Down Expand Up @@ -192,8 +192,8 @@ partial void ProcessCreateTranscriptionResponseContent(
global::OpenAI.AnyOf<string, global::OpenAI.CreateTranscriptionRequestModel?> model,
string? language = default,
string? prompt = default,
global::OpenAI.AudioResponseFormat? responseFormat = global::OpenAI.AudioResponseFormat.Json,
double? temperature = 0,
global::OpenAI.AudioResponseFormat? responseFormat = default,
double? temperature = default,
global::System.Collections.Generic.IList<global::OpenAI.CreateTranscriptionRequestTimestampGranularitie>? timestampGranularities = default,
global::System.Threading.CancellationToken cancellationToken = default)
{
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -82,13 +82,13 @@ partial void ProcessCreateTranslationResponseContent(
content: new global::System.Net.Http.StringContent($"{request.Prompt}"),
name: "prompt");
}
if (request.ResponseFormat != global::OpenAI.AudioResponseFormat.Json)
if (request.ResponseFormat != default)
{
__httpRequestContent.Add(
content: new global::System.Net.Http.StringContent($"{request.ResponseFormat?.ToValueString()}"),
name: "response_format");
}
if (request.Temperature != 0)
if (request.Temperature != default)
{
__httpRequestContent.Add(
content: new global::System.Net.Http.StringContent($"{request.Temperature}"),
Expand Down Expand Up @@ -172,8 +172,8 @@ partial void ProcessCreateTranslationResponseContent(
string filename,
global::OpenAI.AnyOf<string, global::OpenAI.CreateTranslationRequestModel?> model,
string? prompt = default,
global::OpenAI.AudioResponseFormat? responseFormat = global::OpenAI.AudioResponseFormat.Json,
double? temperature = 0,
global::OpenAI.AudioResponseFormat? responseFormat = default,
double? temperature = default,
global::System.Threading.CancellationToken cancellationToken = default)
{
var __request = new global::OpenAI.CreateTranslationRequest
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -60,7 +60,7 @@ partial void ProcessListAuditLogsResponseContent(
global::System.Collections.Generic.IList<string>? actorIds = default,
global::System.Collections.Generic.IList<string>? actorEmails = default,
global::System.Collections.Generic.IList<string>? resourceIds = default,
int? limit = 20,
int? limit = default,
string? after = default,
string? before = default,
global::System.Threading.CancellationToken cancellationToken = default)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -34,7 +34,7 @@ partial void ProcessListBatchesResponseContent(
/// <exception cref="global::System.InvalidOperationException"></exception>
public async global::System.Threading.Tasks.Task<global::OpenAI.ListBatchesResponse> ListBatchesAsync(
string? after = default,
int? limit = 20,
int? limit = default,
global::System.Threading.CancellationToken cancellationToken = default)
{
PrepareArguments(
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -247,25 +247,25 @@ partial void ProcessCreateChatCompletionResponseContent(
public async global::System.Threading.Tasks.Task<global::OpenAI.CreateChatCompletionResponse> CreateChatCompletionAsync(
global::System.Collections.Generic.IList<global::OpenAI.ChatCompletionRequestMessage> messages,
global::OpenAI.AnyOf<string, global::OpenAI.CreateChatCompletionRequestModel?> model,
bool? store = false,
bool? store = default,
global::System.Collections.Generic.Dictionary<string, string>? metadata = default,
double? frequencyPenalty = 0,
double? frequencyPenalty = default,
global::System.Collections.Generic.Dictionary<string, int>? logitBias = default,
bool? logprobs = false,
bool? logprobs = default,
int? topLogprobs = default,
int? maxCompletionTokens = default,
int? n = 1,
int? n = default,
global::System.Collections.Generic.IList<global::OpenAI.ChatCompletionModalitie>? modalities = default,
global::OpenAI.CreateChatCompletionRequestAudio? audio = default,
double? presencePenalty = 0,
double? presencePenalty = default,
global::OpenAI.ResponseFormat? responseFormat = default,
int? seed = default,
global::OpenAI.CreateChatCompletionRequestServiceTier? serviceTier = global::OpenAI.CreateChatCompletionRequestServiceTier.Auto,
global::OpenAI.CreateChatCompletionRequestServiceTier? serviceTier = default,
global::OpenAI.OneOf<string, global::System.Collections.Generic.IList<string>>? stop = default,
bool? stream = false,
bool? stream = default,
global::OpenAI.ChatCompletionStreamOptions? streamOptions = default,
double? temperature = 1,
double? topP = 1,
double? temperature = default,
double? topP = default,
global::System.Collections.Generic.IList<global::OpenAI.ChatCompletionTool>? tools = default,
global::OpenAI.ChatCompletionToolChoiceOption? toolChoice = default,
bool? parallelToolCalls = default,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -206,21 +206,21 @@ partial void ProcessCreateCompletionResponseContent(
public async global::System.Threading.Tasks.Task<global::OpenAI.CreateCompletionResponse> CreateCompletionAsync(
global::OpenAI.AnyOf<string, global::OpenAI.CreateCompletionRequestModel?> model,
global::OpenAI.OneOf<string, global::System.Collections.Generic.IList<string>, global::System.Collections.Generic.IList<int>, global::System.Collections.Generic.IList<global::System.Collections.Generic.IList<int>>>? prompt,
int? bestOf = 1,
bool? echo = false,
double? frequencyPenalty = 0,
int? bestOf = default,
bool? echo = default,
double? frequencyPenalty = default,
global::System.Collections.Generic.Dictionary<string, int>? logitBias = default,
int? logprobs = default,
int? maxTokens = 16,
int? n = 1,
double? presencePenalty = 0,
int? maxTokens = default,
int? n = default,
double? presencePenalty = default,
int? seed = default,
global::OpenAI.OneOf<string, global::System.Collections.Generic.IList<string>>? stop = default,
bool? stream = false,
bool? stream = default,
global::OpenAI.ChatCompletionStreamOptions? streamOptions = default,
string? suffix = default,
double? temperature = 1,
double? topP = 1,
double? temperature = default,
double? topP = default,
string? user = default,
global::System.Threading.CancellationToken cancellationToken = default)
{
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -142,7 +142,7 @@ partial void ProcessCreateEmbeddingResponseContent(
public async global::System.Threading.Tasks.Task<global::OpenAI.CreateEmbeddingResponse> CreateEmbeddingAsync(
global::OpenAI.OneOf<string, global::System.Collections.Generic.IList<string>, global::System.Collections.Generic.IList<int>, global::System.Collections.Generic.IList<global::System.Collections.Generic.IList<int>>> input,
global::OpenAI.AnyOf<string, global::OpenAI.CreateEmbeddingRequestModel?> model,
global::OpenAI.CreateEmbeddingRequestEncodingFormat? encodingFormat = global::OpenAI.CreateEmbeddingRequestEncodingFormat.Float,
global::OpenAI.CreateEmbeddingRequestEncodingFormat? encodingFormat = default,
int? dimensions = default,
string? user = default,
global::System.Threading.CancellationToken cancellationToken = default)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -40,7 +40,7 @@ partial void ProcessListFineTuningEventsResponseContent(
public async global::System.Threading.Tasks.Task<global::OpenAI.ListFineTuningJobEventsResponse> ListFineTuningEventsAsync(
string fineTuningJobId,
string? after = default,
int? limit = 20,
int? limit = default,
global::System.Threading.CancellationToken cancellationToken = default)
{
PrepareArguments(
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -40,7 +40,7 @@ partial void ProcessListFineTuningJobCheckpointsResponseContent(
public async global::System.Threading.Tasks.Task<global::OpenAI.ListFineTuningJobCheckpointsResponse> ListFineTuningJobCheckpointsAsync(
string fineTuningJobId,
string? after = default,
int? limit = 10,
int? limit = default,
global::System.Threading.CancellationToken cancellationToken = default)
{
PrepareArguments(
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -34,7 +34,7 @@ partial void ProcessListPaginatedFineTuningJobsResponseContent(
/// <exception cref="global::System.InvalidOperationException"></exception>
public async global::System.Threading.Tasks.Task<global::OpenAI.ListPaginatedFineTuningJobsResponse> ListPaginatedFineTuningJobsAsync(
string? after = default,
int? limit = 20,
int? limit = default,
global::System.Threading.CancellationToken cancellationToken = default)
{
PrepareArguments(
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -66,8 +66,8 @@ public partial interface IAssistantsClient
global::System.Collections.Generic.IList<global::OpenAI.ToolsItem2>? tools = default,
global::OpenAI.CreateAssistantRequestToolResources? toolResources = default,
object? metadata = default,
double? temperature = 1,
double? topP = 1,
double? temperature = default,
double? topP = default,
global::OpenAI.AssistantsApiResponseFormatOption? responseFormat = default,
global::System.Threading.CancellationToken cancellationToken = default);
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -96,8 +96,8 @@ public partial interface IAssistantsClient
global::System.Collections.Generic.IList<global::OpenAI.CreateMessageRequest>? additionalMessages = default,
global::System.Collections.Generic.IList<global::OpenAI.ToolsItem4>? tools = default,
object? metadata = default,
double? temperature = 1,
double? topP = 1,
double? temperature = default,
double? topP = default,
bool? stream = default,
int? maxPromptTokens = default,
int? maxCompletionTokens = default,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -86,8 +86,8 @@ public partial interface IAssistantsClient
global::System.Collections.Generic.IList<global::OpenAI.ToolsItem5>? tools = default,
global::OpenAI.CreateThreadAndRunRequestToolResources? toolResources = default,
object? metadata = default,
double? temperature = 1,
double? topP = 1,
double? temperature = default,
double? topP = default,
bool? stream = default,
int? maxPromptTokens = default,
int? maxCompletionTokens = default,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -18,8 +18,8 @@ public partial interface IAssistantsClient
/// <param name="cancellationToken">The token to cancel the operation with</param>
/// <exception cref="global::System.InvalidOperationException"></exception>
global::System.Threading.Tasks.Task<global::OpenAI.ListAssistantsResponse> ListAssistantsAsync(
int? limit = 20,
global::OpenAI.ListAssistantsOrder? order = global::OpenAI.ListAssistantsOrder.Desc,
int? limit = default,
global::OpenAI.ListAssistantsOrder? order = default,
string? after = default,
string? before = default,
global::System.Threading.CancellationToken cancellationToken = default);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -21,8 +21,8 @@ public partial interface IAssistantsClient
/// <exception cref="global::System.InvalidOperationException"></exception>
global::System.Threading.Tasks.Task<global::OpenAI.ListMessagesResponse> ListMessagesAsync(
string threadId,
int? limit = 20,
global::OpenAI.ListMessagesOrder? order = global::OpenAI.ListMessagesOrder.Desc,
int? limit = default,
global::OpenAI.ListMessagesOrder? order = default,
string? after = default,
string? before = default,
string? runId = default,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -23,8 +23,8 @@ public partial interface IAssistantsClient
global::System.Threading.Tasks.Task<global::OpenAI.ListRunStepsResponse> ListRunStepsAsync(
string threadId,
string runId,
int? limit = 20,
global::OpenAI.ListRunStepsOrder? order = global::OpenAI.ListRunStepsOrder.Desc,
int? limit = default,
global::OpenAI.ListRunStepsOrder? order = default,
string? after = default,
string? before = default,
global::System.Collections.Generic.IList<global::OpenAI.ListRunStepsIncludeItem>? include = default,
Expand Down
Loading

0 comments on commit 512d13b

Please sign in to comment.