feat: Introduce an OpenAI-compatible client to replace the Groq-specific client and enable multiple LLM providers.
This commit is contained in:
@@ -8,16 +8,16 @@ namespace Toak.Serialization;
|
||||
[JsonSourceGenerationOptions(WriteIndented = true, DefaultIgnoreCondition = JsonIgnoreCondition.WhenWritingNull)]
|
||||
[JsonSerializable(typeof(ToakConfig))]
|
||||
[JsonSerializable(typeof(WhisperResponse))]
|
||||
[JsonSerializable(typeof(LlamaRequest))]
|
||||
[JsonSerializable(typeof(LlamaRequestMessage))]
|
||||
[JsonSerializable(typeof(LlamaResponse))]
|
||||
[JsonSerializable(typeof(LlamaChoice))]
|
||||
[JsonSerializable(typeof(LlamaRequestMessage[]))]
|
||||
[JsonSerializable(typeof(LlamaChoice[]))]
|
||||
[JsonSerializable(typeof(LlamaStreamResponse))]
|
||||
[JsonSerializable(typeof(LlamaStreamChoice))]
|
||||
[JsonSerializable(typeof(LlamaStreamDelta))]
|
||||
[JsonSerializable(typeof(LlamaStreamChoice[]))]
|
||||
[JsonSerializable(typeof(OpenAiRequest))]
|
||||
[JsonSerializable(typeof(OpenAiRequestMessage))]
|
||||
[JsonSerializable(typeof(OpenAiResponse))]
|
||||
[JsonSerializable(typeof(OpenAiChoice))]
|
||||
[JsonSerializable(typeof(OpenAiRequestMessage[]))]
|
||||
[JsonSerializable(typeof(OpenAiChoice[]))]
|
||||
[JsonSerializable(typeof(OpenAiStreamResponse))]
|
||||
[JsonSerializable(typeof(OpenAiStreamChoice))]
|
||||
[JsonSerializable(typeof(OpenAiStreamDelta))]
|
||||
[JsonSerializable(typeof(OpenAiStreamChoice[]))]
|
||||
[JsonSerializable(typeof(Toak.Core.Skills.SkillDefinition))]
|
||||
[JsonSerializable(typeof(Toak.Core.HistoryEntry))]
|
||||
internal partial class AppJsonSerializerContext : JsonSerializerContext
|
||||
|
||||
Reference in New Issue
Block a user