diff --git a/screenpipe-app-tauri/components/settings/ai-section.tsx b/screenpipe-app-tauri/components/settings/ai-section.tsx index 94d2002f3..cfc7898ef 100644 --- a/screenpipe-app-tauri/components/settings/ai-section.tsx +++ b/screenpipe-app-tauri/components/settings/ai-section.tsx @@ -543,10 +543,10 @@ const AISection = () => {

- maximum number of characters (think 3 characters per token) + maximum number of characters (think 4 characters per token) to send to the ai model.
usually, openai models support up to 128k tokens, which is - roughly 30k-40k characters.
+ roughly 512k characters.
we'll use this for UI purposes to show you how much you can send.

diff --git a/screenpipe-app-tauri/lib/hooks/use-settings.tsx b/screenpipe-app-tauri/lib/hooks/use-settings.tsx index 4791be437..7aeb8fc38 100644 --- a/screenpipe-app-tauri/lib/hooks/use-settings.tsx +++ b/screenpipe-app-tauri/lib/hooks/use-settings.tsx @@ -117,7 +117,7 @@ const DEFAULT_SETTINGS: Settings = { includedWindows: [], aiProviderType: "openai", aiUrl: "https://api.openai.com/v1", - aiMaxContextChars: 30000, + aiMaxContextChars: 512000, fps: 0.5, vadSensitivity: "high", analyticsEnabled: true,