diff --git a/README.md b/README.md index 8a4d974..c8385e9 100644 --- a/README.md +++ b/README.md @@ -8,12 +8,12 @@ Try it here: [https://future.attejuvonen.fi](https://future.attejuvonen.fi) ### Dev -1. Set up environment variables +1. Set up environment variables (e.g. in /.vscode/launch.json) - OPENAI_API_KEY - OPENAI_ORGANIZATION - LOG_ENDPOINT -2. Run server.js with node v17 +2. Run server.js with node v17 (e.g. with VSCode run -> run without debugging) 3. Open http://localhost:3000 diff --git a/fly.toml b/fly.toml index 6dd8ee2..f751b9b 100644 --- a/fly.toml +++ b/fly.toml @@ -6,7 +6,7 @@ kill_timeout = 5 processes = [] [build] - builder = "heroku/buildpacks:20" + builder = "heroku/builder:24" [env] PORT = "8080" diff --git a/public/index.html b/public/index.html index 3765393..ea2fd40 100644 --- a/public/index.html +++ b/public/index.html @@ -872,9 +872,9 @@

return Math.round(Math.random() * 40) } if (['.', '?', '!'].includes(nextChar)) { - return 300 + Math.round(Math.random() * 500) + return 150 + Math.round(Math.random() * 250) } - return Math.round(Math.random() * 120) + return Math.round(Math.random() * 60) } const stopSpinner = () => { diff --git a/server.js b/server.js index 77a7d45..f39409d 100644 --- a/server.js +++ b/server.js @@ -9,6 +9,10 @@ import PROMPT_QA_EXAMPLES from "./prompt-qa-examples.js"; const PROMPT_INSTRUCTIONS = fs.readFileSync('prompt-instructions.txt', 'utf8'); +if (!process.env.OPENAI_ORGANIZATION || !process.env.OPENAI_API_KEY) { + throw new Error('Missing env variables for OpenAI authentication!') +} + const configuration = new Configuration({ organization: process.env.OPENAI_ORGANIZATION, apiKey: process.env.OPENAI_API_KEY @@ -100,7 +104,7 @@ const constructPromptChatGPT = (PROMPT_INSTRUCTIONS, PROMPT_QA_EXAMPLES, session const smokeTestAPI = async () => { try { - const response = await openai.retrieveModel("text-davinci-003"); + const response = await openai.retrieveModel("gpt-4o-2024-05-13"); } catch (error) { serverStatusGreen = false const errorMessage = error.response ? (error.response.status + error.response.data) : error.message @@ -122,7 +126,7 @@ const getResponse = async (PROMPT_INSTRUCTIONS, PROMPT_QA_EXAMPLES, sessionHisto } try { const response = await openai.createChatCompletion({ - model: "gpt-3.5-turbo-0301", + model: "gpt-4o-2024-05-13", messages: messages, max_tokens: 256, temperature: 0.4