604ed219 by Administrator

Merge branch '32993_CHATBOT_PLUS_GPT4o_MINI' into 'master'

32993_CHATBOT_PLUS_GPT4o_MINI

See merge request !128
2 parents 4b88ee48 50ef50e1
Pipeline #30440 for 604ed219 passed in 16 seconds
......@@ -13,10 +13,12 @@ function App() {
displayTogetherAiResponse()
}, [])
const DEFAULT_MODEL = process.env.REACT_APP_DEFAULT_MODEL || 'gpt-4o-mini';
const [chatInput, setChatInput] = useState("");
const [models, setModels] = useState([]);
const [temperature, setTemperature] = useState(0.7);
const GPTTurbo = "gpt-3.5-turbo";
const GPTTurbo = DEFAULT_MODEL;
const [currentModel, setCurrentModel] = useState(GPTTurbo);
const [chatLog, setChatLog] = useState([{
user: "gpt",
......@@ -148,12 +150,6 @@ function App() {
const data = await response.json();
const parsedData = data.message ? data.message.trim() : "";
if (togetherAiResponse) {
if (currentModel !== 'gpt-3.5-turbo') {
console.log(data)
}
}
if(data.status === 'invalid'){
if(data.limited) {
window.btutil_modalRegisterUpgrade(false);
......
import ExportButton from "./ExportButton";
const DEFAULT_MODEL = process.env.REACT_APP_DEFAULT_MODEL || 'gpt-4o-mini';
const SideMenu = ({
clearChat,
currentModel,
......@@ -33,8 +34,8 @@ const SideMenu = ({
</option>
))
) : (
<option key={"gpt-3.5-turbo"} value={"gpt-3.5-turbo"}>
{"gpt-3.5-turbo"}
<option key={DEFAULT_MODEL} value={DEFAULT_MODEL}>
{DEFAULT_MODEL}
</option>
)}
</select>
......
......@@ -10,6 +10,7 @@ const fetch = require('node-fetch');
const anchorme = require("anchorme").default;
const axios = require('axios');
const { encodingForModel } = require('js-tiktoken');
const DEFAULT_MODEL = process.env.REACT_APP_DEFAULT_MODEL || 'gpt-4o-mini';
const tiktokenModels = [
'text-davinci-003',
'text-davinci-002',
......@@ -45,7 +46,7 @@ const tiktokenModels = [
'gpt-4-0314',
'gpt-4-32k',
'gpt-4-32k-0314',
'gpt-3.5-turbo',
'gpt-4o-mini',
'gpt-3.5-turbo-0301'
];
......@@ -53,7 +54,7 @@ let client;
let filteredModels = {};
const allowedEndpoints = ["openAI", "Opensource", "Llama"];
const allowedModels = [
"gpt-3.5-turbo",
DEFAULT_MODEL,
"google/gemma-2-9b-it",
"meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo"
];
......@@ -157,7 +158,7 @@ app.post('/api', async (req, res) => {
let usage = {};
let enc = null;
try {
enc = encodingForModel(tiktokenModels.includes(currentModel) ? currentModel : 'gpt-3.5-turbo');
enc = encodingForModel(tiktokenModels.includes(currentModel) ? currentModel : DEFAULT_MODEL);
usage.prompt_tokens = (enc.encode(query_prompt)).length;
usage.completion_tokens = (enc.encode(input)).length;
usage.total_tokens = usage.prompt_tokens + usage.completion_tokens;
......@@ -240,7 +241,7 @@ async function runGPTTurbo(req, res) {
let usage = {};
let enc = null;
try {
enc = encodingForModel(tiktokenModels.includes(currentModel) ? currentModel : 'gpt-3.5-turbo');
enc = encodingForModel(tiktokenModels.includes(currentModel) ? currentModel : DEFAULT_MODEL);
usage.prompt_tokens = (enc.encode(query_prompt)).length;
usage.completion_tokens = (enc.encode(input)).length;
usage.total_tokens = usage.prompt_tokens + usage.completion_tokens;
......
......@@ -19,7 +19,7 @@
"dotenv": "^16.0.3",
"express": "^4.18.2",
"express-rate-limit": "^6.7.0",
"js-tiktoken": "1.0.7",
"js-tiktoken": "1.0.15",
"morgan": "^1.10.0",
"node-fetch": "^2.7.0",
"openai": "^3.2.0",
......
Styling with Markdown is supported
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!