feat: Implement compact JSON serialization for history entries, improve history loading robustness, and add a new LLM model option.
This commit is contained in:
@@ -26,8 +26,8 @@ public static class OnboardCommand
|
||||
config.LlmModel = AnsiConsole.Prompt(
|
||||
new SelectionPrompt<string>()
|
||||
.Title("Select [green]LLM Model[/]:")
|
||||
.AddChoices(new[] { "openai/gpt-oss-20b", "llama-3.1-8b-instant" })
|
||||
.UseConverter(c => c == "openai/gpt-oss-20b" ? "openai/gpt-oss-20b (Fastest)" : "llama-3.1-8b-instant (Cheapest)"));
|
||||
.AddChoices(new[] { "openai/gpt-oss-20b", "llama-3.1-8b-instant", "llama-3.3-70b-versatile" })
|
||||
.UseConverter(c => c == "openai/gpt-oss-20b" ? "openai/gpt-oss-20b (Fastest)" : c == "llama-3.1-8b-instant" ? "llama-3.1-8b-instant (Cheapest)" : "llama-3.3-70b-versatile (More Accurate)"));
|
||||
|
||||
if (config.LlmModel.Contains(" ")) config.LlmModel = config.LlmModel.Split(' ')[0];
|
||||
|
||||
|
||||
@@ -38,7 +38,7 @@ public static class StatsCommand
|
||||
.Select(g => g.Key)
|
||||
.ToList();
|
||||
|
||||
AnsiConsole.MarkupLine("[bold blue]Toak Usage Statistics[/]");
|
||||
AnsiConsole.MarkupLine("\n[bold blue]Toak Usage Statistics[/]");
|
||||
AnsiConsole.MarkupLine($"[dim]Total recordings:[/] {totalCount}");
|
||||
AnsiConsole.MarkupLine($"[dim]Total duration:[/] {totalDuration.TotalMinutes:F1}m");
|
||||
AnsiConsole.MarkupLine($"[dim]Average processing latency:[/] {avgDuration.TotalSeconds:F2}s");
|
||||
|
||||
Reference in New Issue
Block a user