diff --git a/src/ketr-chat/src/App.tsx b/src/ketr-chat/src/App.tsx
index ec12d5f..2ebe87e 100644
--- a/src/ketr-chat/src/App.tsx
+++ b/src/ketr-chat/src/App.tsx
@@ -62,11 +62,11 @@ This LLM agent was built by James Ketrenos in order to provide answers to any qu
In addition to being a RAG enabled expert system, the LLM is configured with real-time access to weather, stocks, the current time, and can answer questions about the contents of a website.
You can ask things like:
- *
- *
*
- *
*
+ *
+ *
+ *
Or click the text above to submit that query.
@@ -451,7 +451,7 @@ interface ChatQueryInterface {
}
const ChatQuery = ({ text, submitQuery }: ChatQueryInterface) => {
- return ();
+ return ();
}
const Message = ({ message, submitQuery }: MessageInterface) => {
@@ -532,7 +532,7 @@ const App = () => {
const [lastEvalTPS, setLastEvalTPS] = useState(35);
const [lastPromptTPS, setLastPromptTPS] = useState(430);
const [countdown, setCountdown] = useState(0);
- const [messageHistoryLength, setMessageHistoryLength] = useState(0);
+ const [messageHistoryLength, setMessageHistoryLength] = useState(5);
const timerRef = useRef(null);
diff --git a/src/server.py b/src/server.py
index 133d75d..2587f89 100644
--- a/src/server.py
+++ b/src/server.py
@@ -144,6 +144,7 @@ LOG_LEVEL="info"
USE_TLS=False
WEB_HOST="0.0.0.0"
WEB_PORT=5000
+DEFAULT_HISTORY_LENGTH=5
# %%
# Globals
@@ -391,8 +392,8 @@ class WebServer:
context["context_tokens"] = round(len(str(context["system"])) * 3 / 4) # Estimate context usage
response["context_used"] = context["context_tokens"]
case "message-history-length":
- context["message_history_length"] = 5
- response["message-history-length"] = 5
+ context["message_history_length"] = DEFAULT_HISTORY_LENGTH
+ response["message-history-length"] = DEFAULT_HISTORY_LENGTH
if not response:
return JSONResponse({ "error": "Usage: { reset: rags|tools|history|system-prompt}"})