From 8042cb950d594735e5f5788ca16d29ca2f64bc6d Mon Sep 17 00:00:00 2001 From: HanishKVC Date: Wed, 22 May 2024 18:56:51 +0530 Subject: [PATCH 01/19] SimpleChat: A placeholder system prompt, Use usage msg in code Just have a alert msg wrt needing javascript enabled in html. And have usage message from js file. Update the usage message a bit. So also enable switch session wrt setup_ui call. Add a possible system prompt as a placeholder for the system-input. --- examples/server/public_simplechat/index.html | 7 ++----- examples/server/public_simplechat/simplechat.js | 12 +++++++----- 2 files changed, 9 insertions(+), 10 deletions(-) diff --git a/examples/server/public_simplechat/index.html b/examples/server/public_simplechat/index.html index 1eb390b85a69c..72a91a09bf916 100644 --- a/examples/server/public_simplechat/index.html +++ b/examples/server/public_simplechat/index.html @@ -30,15 +30,12 @@
- +

-

Enter the system prompt above, before entering/submitting any user query.

-

Enter your text to the ai assistant below.

-

Use shift+enter for inserting enter.

-

Refresh the page to start over fresh.

+

You need to have javascript enabled.


diff --git a/examples/server/public_simplechat/simplechat.js b/examples/server/public_simplechat/simplechat.js index 3fc4dbc2026fa..51986ad9c6d21 100644 --- a/examples/server/public_simplechat/simplechat.js +++ b/examples/server/public_simplechat/simplechat.js @@ -14,10 +14,12 @@ class ApiEP { } let gUsageMsg = ` -

Enter the system prompt above, before entering/submitting any user query.

-

Enter your text to the ai assistant below.

-

Use shift+enter for inserting enter.

-

Refresh the page to start over fresh.

+ `; class SimpleChat { @@ -471,7 +473,7 @@ function startme() { for (let cid of gChatIds) { gMuitChat.new_chat_session(cid); } - gMuitChat.setup_ui(gChatIds[0]); + gMuitChat.setup_ui(gChatIds[0], true); gMuitChat.show_sessions(); } From 3c11098d1eb4f4a53194efe00e555b0aa69ebb5f Mon Sep 17 00:00:00 2001 From: HanishKVC Date: Wed, 22 May 2024 22:10:52 +0530 Subject: [PATCH 02/19] SimpleChat:CompletionMode: Allow control of Role: prefix --- examples/server/public_simplechat/simplechat.js | 11 ++++++++--- 1 file changed, 8 insertions(+), 3 deletions(-) diff --git a/examples/server/public_simplechat/simplechat.js b/examples/server/public_simplechat/simplechat.js index 51986ad9c6d21..4209feb6b12ae 100644 --- a/examples/server/public_simplechat/simplechat.js +++ b/examples/server/public_simplechat/simplechat.js @@ -97,11 +97,15 @@ class SimpleChat { /** * Return a string form of json object suitable for /completions + * @param {boolean} bInsertStandardRolePrefix Insert ": " as prefix wrt each role's message */ - request_prompt_jsonstr() { + request_prompt_jsonstr(bInsertStandardRolePrefix) { let prompt = ""; for(const chat of this.xchat) { - prompt += `${chat.role}: ${chat.content}\n`; + if (bInsertStandardRolePrefix) { + prompt += `${chat.role}: `; + } + prompt += `${chat.content}\n`; } let req = { prompt: prompt, @@ -174,6 +178,7 @@ let gChatURL = { 'completion': `${gBaseURL}/completions`, } const gbCompletionFreshChatAlways = true; +let gbCompletionInsertStandardRolePrefix = true; /** @@ -346,7 +351,7 @@ class MultiChatUI { if (apiEP == ApiEP.Chat) { theBody = chat.request_messages_jsonstr(); } else { - theBody = chat.request_prompt_jsonstr(); + theBody = chat.request_prompt_jsonstr(gbCompletionInsertStandardRolePrefix); } this.elInUser.value = "working..."; From 0dba8f8857b87770d74bdf2c493938ea74d93b5d Mon Sep 17 00:00:00 2001 From: HanishKVC Date: Wed, 22 May 2024 22:40:44 +0530 Subject: [PATCH 03/19] SimpleChat:Completion: Avoid Role: prefix; Newline only in between In completion mode * avoid inserting Role: prefix before each role's message * avoid inserting newline at the begin and end of the prompt message. However if there are multiple role messages, then insert newline when going from one role's message to the next role's message. --- examples/server/public_simplechat/simplechat.js | 9 +++++++-- 1 file changed, 7 insertions(+), 2 deletions(-) diff --git a/examples/server/public_simplechat/simplechat.js b/examples/server/public_simplechat/simplechat.js index 4209feb6b12ae..3db050cacaec9 100644 --- a/examples/server/public_simplechat/simplechat.js +++ b/examples/server/public_simplechat/simplechat.js @@ -101,11 +101,16 @@ class SimpleChat { */ request_prompt_jsonstr(bInsertStandardRolePrefix) { let prompt = ""; + let iCnt = 0; for(const chat of this.xchat) { + iCnt += 1; + if (iCnt > 1) { + prompt += "\n"; + } if (bInsertStandardRolePrefix) { prompt += `${chat.role}: `; } - prompt += `${chat.content}\n`; + prompt += `${chat.content}`; } let req = { prompt: prompt, @@ -178,7 +183,7 @@ let gChatURL = { 'completion': `${gBaseURL}/completions`, } const gbCompletionFreshChatAlways = true; -let gbCompletionInsertStandardRolePrefix = true; +let gbCompletionInsertStandardRolePrefix = false; /** From 7a0a42367f02c44276ade44d865c74aca38e1a29 Mon Sep 17 00:00:00 2001 From: HanishKVC Date: Wed, 22 May 2024 22:50:44 +0530 Subject: [PATCH 04/19] SimpleChat:CompletionMode: Update readme/usage, trim textarea newline Readme update wrt completion mode behavior. Usage help updated wrt completion mode behavior. When changing from input to textarea elment wrt user input, the last newline at the end of the user input wrt textarea, was forgotten to be filtered, this is fixed now. However if user wants to have a explicit newline they can using shift+enter to insert a newline, that wont be removed. The extra newline removal logic uses substring and keyup to keep things simple and avoid some previously noted bugs wrt other events in the key path as well as IME composition etal. --- examples/server/public_simplechat/readme.md | 10 ++++++++++ examples/server/public_simplechat/simplechat.js | 6 +++++- 2 files changed, 15 insertions(+), 1 deletion(-) diff --git a/examples/server/public_simplechat/readme.md b/examples/server/public_simplechat/readme.md index 5ac8258f21aca..87fa464b947b0 100644 --- a/examples/server/public_simplechat/readme.md +++ b/examples/server/public_simplechat/readme.md @@ -47,6 +47,16 @@ Open this simple web front end from your local browser Once inside * Select between chat and completion mode. By default it is set to chat mode. + * in completion mode + * the logic by default doesnt insert "THE_ROLE: " prefix wrt each role's message. + If the model requires any prefix wrt user role messages, then the end user has to + explicitly add the needed prefix, when they enter their chat message. + This keeps the logic simple, while still giving flexibility to the end user to + manage any templating/tagging requirement wrt their messages to the model. + * the logic doesnt insert newline at the begining and end wrt the prompt message generated. + However if the chat being sent to completion end point has more than one role's message, + then insert newline when moving from one role's message to the next role's message, so + that it can be clearly identified. * If you want to provide a system prompt, then ideally enter it first, before entering any user query. * if chat.add_system_begin is used * you cant change the system prompt, after it is has been submitted once along with user query. diff --git a/examples/server/public_simplechat/simplechat.js b/examples/server/public_simplechat/simplechat.js index 3db050cacaec9..59710d330925c 100644 --- a/examples/server/public_simplechat/simplechat.js +++ b/examples/server/public_simplechat/simplechat.js @@ -16,8 +16,10 @@ class ApiEP { let gUsageMsg = `
  • Set system prompt above, to try control ai response charactersitic, if model supports same.
  • +
  • + Completion mode normally wont have a system prompt.
  • Enter your query to ai assistant below.
  • -
  • Use shift+enter for inserting enter.
  • +
  • + Completion mode doesnt insert user: prefix implicitly.
  • +
  • + Use shift+enter for inserting enter/newline.
  • Refresh the page to start over fresh.
`; @@ -303,6 +305,8 @@ class MultiChatUI { // allow user to insert enter into their message using shift+enter. // while just pressing enter key will lead to submitting. if ((ev.key === "Enter") && (!ev.shiftKey)) { + let value = this.elInUser.value; + this.elInUser.value = value.substring(0,value.length-1); this.elBtnUser.click(); ev.preventDefault(); } From fe60655e1f87fca01d58e27e6097a1a2d75b9884 Mon Sep 17 00:00:00 2001 From: HanishKVC Date: Thu, 23 May 2024 03:42:02 +0530 Subject: [PATCH 05/19] SimpleChat:SC: Ensure proper clearing/reseting previous logic would have cleared/reset the xchat, without doing the same wrt iLastSys, thus leading to it pointing to a now non existent role-content entry. So if a user set a system prompt and used completion mode, it would have done the half stupid clear, after the model response was got. Inturn when user tries to send a new completion query, it would inturn lead to handle_user_submit trying to add/update system prompt if any, which will fail, bcas iLastSys will be still pointing to a non existant entry. This is fixed now, by having a proper clear helper wrt SC class. --- examples/server/public_simplechat/simplechat.js | 7 ++++++- 1 file changed, 6 insertions(+), 1 deletion(-) diff --git a/examples/server/public_simplechat/simplechat.js b/examples/server/public_simplechat/simplechat.js index 59710d330925c..5bcf9ed173833 100644 --- a/examples/server/public_simplechat/simplechat.js +++ b/examples/server/public_simplechat/simplechat.js @@ -35,6 +35,11 @@ class SimpleChat { this.iLastSys = -1; } + clear() { + this.xchat = []; + this.iLastSys = -1; + } + /** * Add an entry into xchat * @param {string} role @@ -397,7 +402,7 @@ class MultiChatUI { // in a completion mode with multiple user-assistant chat data // from before to be sent/occur once. if ((apiEP == ApiEP.Completion) && (gbCompletionFreshChatAlways)) { - chat.xchat.length = 0; + chat.clear(); } this.ui_reset_userinput(); } From 01594daeb422583e8ccdce56281465318f0d9494 Mon Sep 17 00:00:00 2001 From: HanishKVC Date: Thu, 23 May 2024 05:07:32 +0530 Subject: [PATCH 06/19] SimpleChat: Update usage note and readme a bit --- examples/server/public_simplechat/readme.md | 34 +++++++++++++------ .../server/public_simplechat/simplechat.js | 2 +- 2 files changed, 25 insertions(+), 11 deletions(-) diff --git a/examples/server/public_simplechat/readme.md b/examples/server/public_simplechat/readme.md index 87fa464b947b0..10c49dff20a33 100644 --- a/examples/server/public_simplechat/readme.md +++ b/examples/server/public_simplechat/readme.md @@ -43,20 +43,30 @@ next run this web front end in examples/server/public_simplechat ### using the front end Open this simple web front end from your local browser + * http://127.0.0.1:PORT/index.html Once inside + * Select between chat and completion mode. By default it is set to chat mode. - * in completion mode - * the logic by default doesnt insert "THE_ROLE: " prefix wrt each role's message. - If the model requires any prefix wrt user role messages, then the end user has to - explicitly add the needed prefix, when they enter their chat message. - This keeps the logic simple, while still giving flexibility to the end user to - manage any templating/tagging requirement wrt their messages to the model. - * the logic doesnt insert newline at the begining and end wrt the prompt message generated. - However if the chat being sent to completion end point has more than one role's message, - then insert newline when moving from one role's message to the next role's message, so - that it can be clearly identified. + +* In completion mode + * logic by default doesnt insert any role specific "ROLE: " prefix wrt each role's message. + If the model requires any prefix wrt user role messages, then the end user has to + explicitly add the needed prefix, when they enter their chat message. + Similarly if the model requires any prefix to trigger assistant/ai-model response, + then the end user needs to enter the same. + This keeps the logic simple, while still giving flexibility to the end user to + manage any templating/tagging requirement wrt their messages to the model. + * the logic doesnt insert newline at the begining and end wrt the prompt message generated. + However if the chat being sent to /completions end point has more than one role's message, + then insert newline when moving from one role's message to the next role's message, so + that it can be clearly identified/distinguished. + * given that /completions endpoint normally doesnt add additional chat-templating of its + own, the above ensures that end user can create a custom single/multi message combo with + any tags/special-tokens related chat templating to test out model handshake. Or enduser + can use it just for normal completion related/based query. + * If you want to provide a system prompt, then ideally enter it first, before entering any user query. * if chat.add_system_begin is used * you cant change the system prompt, after it is has been submitted once along with user query. @@ -65,12 +75,16 @@ Once inside * one can change the system prompt any time during chat, by changing the contents of system prompt. * inturn the updated/changed system prompt will be inserted into the chat session. * this allows for the subsequent user chatting to be driven by the new system prompt set above. + * Enter your query and either press enter or click on the submit button. If you want to insert enter (\n) as part of your chat/query to ai model, use shift+enter. + * Wait for the logic to communicate with the server and get the response. * the user is not allowed to enter any fresh query during this time. * the user input box will be disabled and a working message will be shown in it. + * just refresh the page, to reset wrt the chat history and or system prompt and start afresh. + * Using NewChat one can start independent chat sessions. * two independent chat sessions are setup by default. diff --git a/examples/server/public_simplechat/simplechat.js b/examples/server/public_simplechat/simplechat.js index 5bcf9ed173833..ba291a6c63f11 100644 --- a/examples/server/public_simplechat/simplechat.js +++ b/examples/server/public_simplechat/simplechat.js @@ -18,7 +18,7 @@ let gUsageMsg = `
  • Set system prompt above, to try control ai response charactersitic, if model supports same.
  • + Completion mode normally wont have a system prompt.
  • Enter your query to ai assistant below.
  • -
  • + Completion mode doesnt insert user: prefix implicitly.
  • +
  • + Completion mode doesnt insert user/role: prefix implicitly.
  • + Use shift+enter for inserting enter/newline.
  • Refresh the page to start over fresh.
  • From e2164d66e693aa9e5fdd28657c97423cc98a66cc Mon Sep 17 00:00:00 2001 From: HanishKVC Date: Thu, 23 May 2024 05:26:49 +0530 Subject: [PATCH 07/19] SimpleChat:Completion: clear any prev chat history at begining Previously any chat history including model response to a completion query would have got cleared, after showing the same to the user, at the end of handle_user_submit, rather than at the begining. This gave the flexibility that user could switch from chat mode to completion mode and have the chat history till then sent to the ai model, as part of the completion query. However this flow also had the issue that, if user switches between different chat sessions, after getting a completion response, they can no longer see the completion query and its response that they had just got. The new flow changes the clearing of chat history wrt completion mode to the begining of handle_user_submit, so that user doesnt lose the last completion mode query and response, till a new completion mode query is sent to the model, even if they were to switch between the chat sessions. At the same time the loss of flexibility wrt converting previous chat history into being part of the completion query implicitly doesnt matter, because now the end user can enter multiline queries. --- examples/server/public_simplechat/simplechat.js | 15 ++++++++------- 1 file changed, 8 insertions(+), 7 deletions(-) diff --git a/examples/server/public_simplechat/simplechat.js b/examples/server/public_simplechat/simplechat.js index ba291a6c63f11..fa8703948382d 100644 --- a/examples/server/public_simplechat/simplechat.js +++ b/examples/server/public_simplechat/simplechat.js @@ -351,6 +351,14 @@ class MultiChatUI { let chat = this.simpleChats[chatId]; + // In completion mode, if configured, clear any previous chat history. + // So if user wants to simulate a multi-chat based completion query, + // they will have to enter the full thing, as a suitable multiline + // user input/query. + if ((apiEP == ApiEP.Completion) && (gbCompletionFreshChatAlways)) { + chat.clear(); + } + chat.add_system_anytime(this.elInSystem.value, chatId); let content = this.elInUser.value; @@ -397,13 +405,6 @@ class MultiChatUI { } else { console.debug(`DBUG:SimpleChat:MCUI:HandleUserSubmit:ChatId has changed:[${chatId}] [${this.curChatId}]`); } - // Purposefully clear at end rather than begin of this function - // so that one can switch from chat to completion mode and sequece - // in a completion mode with multiple user-assistant chat data - // from before to be sent/occur once. - if ((apiEP == ApiEP.Completion) && (gbCompletionFreshChatAlways)) { - chat.clear(); - } this.ui_reset_userinput(); } From 40fbbeb2f6e0b3a211328e4439f90123940f807e Mon Sep 17 00:00:00 2001 From: HanishKVC Date: Thu, 23 May 2024 14:32:01 +0530 Subject: [PATCH 08/19] SimpleChat:Try read json early, if available For later the server flow doesnt seem to be sending back data early, atleast for the request (inc options) that is currently sent. if able to read json data early on in future, as and when ai model is generating data, then this helper needs to indirectly update the chat div with the recieved data, without waiting for the overall data to be available. --- .../server/public_simplechat/simplechat.js | 24 +++++++++++++++++++ 1 file changed, 24 insertions(+) diff --git a/examples/server/public_simplechat/simplechat.js b/examples/server/public_simplechat/simplechat.js index fa8703948382d..fd834c4e0a689 100644 --- a/examples/server/public_simplechat/simplechat.js +++ b/examples/server/public_simplechat/simplechat.js @@ -342,6 +342,29 @@ class MultiChatUI { } } + /** + * Try read json response early, if available. + * @param {Response} resp + */ + async read_json_early(resp) { + if (!resp.body) { + throw Error("ERRR:SimpleChat:MCUI:ReadJsonEarly:No body..."); + } + let tdUtf8 = new TextDecoder("utf-8"); + let rr = resp.body.getReader(); + let gotBody = ""; + while(true) { + let { value: cur, done: done} = await rr.read(); + let curBody = tdUtf8.decode(cur); + console.debug("DBUG:SC:PART:", curBody); + gotBody += curBody; + if (done) { + break; + } + } + return JSON.parse(gotBody); + } + /** * Handle user query submit request, wrt specified chat session. * @param {string} chatId @@ -388,6 +411,7 @@ class MultiChatUI { }); let respBody = await resp.json(); + //let respBody = await this.read_json_early(resp); console.debug(`DBUG:SimpleChat:MCUI:${chatId}:HandleUserSubmit:RespBody:${JSON.stringify(respBody)}`); let assistantMsg; if (apiEP == ApiEP.Chat) { From 5d84a92d623a7467ace0e3fff6c4fff603cdfffc Mon Sep 17 00:00:00 2001 From: HanishKVC Date: Thu, 23 May 2024 15:23:47 +0530 Subject: [PATCH 09/19] SimpleChat: Rename the half asleep mis-spelled global var --- examples/server/public_simplechat/simplechat.js | 10 +++++----- 1 file changed, 5 insertions(+), 5 deletions(-) diff --git a/examples/server/public_simplechat/simplechat.js b/examples/server/public_simplechat/simplechat.js index fd834c4e0a689..a79553ddaeba6 100644 --- a/examples/server/public_simplechat/simplechat.js +++ b/examples/server/public_simplechat/simplechat.js @@ -508,17 +508,17 @@ class MultiChatUI { } -let gMuitChat; +let gMultiChat; const gChatIds = [ "Default", "Other" ]; function startme() { console.log("INFO:SimpleChat:StartMe:Starting..."); - gMuitChat = new MultiChatUI(); + gMultiChat = new MultiChatUI(); for (let cid of gChatIds) { - gMuitChat.new_chat_session(cid); + gMultiChat.new_chat_session(cid); } - gMuitChat.setup_ui(gChatIds[0], true); - gMuitChat.show_sessions(); + gMultiChat.setup_ui(gChatIds[0], true); + gMultiChat.show_sessions(); } document.addEventListener("DOMContentLoaded", startme); From 073eae6778673b151c38db196868f1f64ae3db02 Mon Sep 17 00:00:00 2001 From: HanishKVC Date: Thu, 23 May 2024 15:32:54 +0530 Subject: [PATCH 10/19] SimpleChat: Common chat request options from a global object --- examples/server/public_simplechat/simplechat.js | 12 ++++++++++-- 1 file changed, 10 insertions(+), 2 deletions(-) diff --git a/examples/server/public_simplechat/simplechat.js b/examples/server/public_simplechat/simplechat.js index a79553ddaeba6..5e7e79fecf9b8 100644 --- a/examples/server/public_simplechat/simplechat.js +++ b/examples/server/public_simplechat/simplechat.js @@ -24,6 +24,11 @@ let gUsageMsg = ` `; +// Add needed fields wrt json object to be sent wrt LLM web services completions endpoint. +let gChatRequestOptions = { + "temperature": 0.7 +}; + class SimpleChat { constructor() { @@ -83,12 +88,15 @@ class SimpleChat { } /** - * Add needed fields wrt json object to be sent wrt LLM web services completions endpoint + * Add needed fields wrt json object to be sent wrt LLM web services completions endpoint. + * The needed fields/options are picked from a global object. * Convert the json into string. * @param {Object} obj */ request_jsonstr(obj) { - obj["temperature"] = 0.7; + for(let k in gChatRequestOptions) { + obj[k] = gChatRequestOptions[k]; + } return JSON.stringify(obj); } From 59f74c7de94361109504a355f1655c72594e1ba1 Mon Sep 17 00:00:00 2001 From: HanishKVC Date: Thu, 23 May 2024 15:43:41 +0530 Subject: [PATCH 11/19] SimpleChat: Update title, usage and readme a bit Keep the title simple so that print file name doesnt have chars that need to be removed. Update readme wrt some of the new helpers and options. Change Usage list to a list of lists, add few items and style it to reduce the margin wrt lists. --- examples/server/public_simplechat/index.html | 2 +- examples/server/public_simplechat/readme.md | 20 +++++++++++++++++++ .../server/public_simplechat/simplechat.css | 7 +++++++ .../server/public_simplechat/simplechat.js | 16 +++++++++++---- 4 files changed, 40 insertions(+), 5 deletions(-) diff --git a/examples/server/public_simplechat/index.html b/examples/server/public_simplechat/index.html index 72a91a09bf916..89b32f3ff3d54 100644 --- a/examples/server/public_simplechat/index.html +++ b/examples/server/public_simplechat/index.html @@ -1,7 +1,7 @@ - SimpleChat (LlamaCPP, ...) + SimpleChat LlamaCppEtal diff --git a/examples/server/public_simplechat/readme.md b/examples/server/public_simplechat/readme.md index 10c49dff20a33..72a428075214d 100644 --- a/examples/server/public_simplechat/readme.md +++ b/examples/server/public_simplechat/readme.md @@ -68,6 +68,8 @@ Once inside can use it just for normal completion related/based query. * If you want to provide a system prompt, then ideally enter it first, before entering any user query. + Normally Completion mode doesnt need system prompt, while Chat mode can generate better/interesting + responses with a suitable system prompt. * if chat.add_system_begin is used * you cant change the system prompt, after it is has been submitted once along with user query. * you cant set a system prompt, after you have submitted any user query @@ -91,6 +93,13 @@ Once inside ## Devel note +gChatRequestOptions maintains the list of options/fields to send along with chat request, +irrespective of whether /chat/completions or /completions endpoint. + + If you want to add additional options/fields to send to the server/ai-model, and or + modify the existing options value, for now you can update this global var using + browser's development-tools/console. + Sometimes the browser may be stuborn with caching of the file, so your updates to html/css/js may not be visible. Also remember that just refreshing/reloading page in browser or for that matter clearing site data, dont directly override site caching in all cases. Worst case you may @@ -103,3 +112,14 @@ not been added, for now. By switching between chat.add_system_begin/anytime, one can control whether one can change the system prompt, anytime during the conversation or only at the beginning. + +read_json_early, is to experiment with reading json response data early on, if available, +so that user can be shown generated data, as and when it is being generated, rather than +at the end when full data is available. + + the server flow doesnt seem to be sending back data early, atleast for request (inc options) + that is currently sent. + + if able to read json data early on in future, as and when ai model is generating data, then + this helper needs to indirectly update the chat div with the recieved data, without waiting + for the overall data to be available. diff --git a/examples/server/public_simplechat/simplechat.css b/examples/server/public_simplechat/simplechat.css index d45f50a957e4c..20c738b12ed6f 100644 --- a/examples/server/public_simplechat/simplechat.css +++ b/examples/server/public_simplechat/simplechat.css @@ -48,6 +48,13 @@ button { flex-direction: column; } +.ul1 { + padding-inline-start: 2vw; +} +.ul2 { + padding-inline-start: 2vw; +} + * { margin: 0.6vmin; } diff --git a/examples/server/public_simplechat/simplechat.js b/examples/server/public_simplechat/simplechat.js index 5e7e79fecf9b8..2ddab0ec2b7f5 100644 --- a/examples/server/public_simplechat/simplechat.js +++ b/examples/server/public_simplechat/simplechat.js @@ -14,13 +14,21 @@ class ApiEP { } let gUsageMsg = ` -
      +
      • Set system prompt above, to try control ai response charactersitic, if model supports same.
      • -
      • + Completion mode normally wont have a system prompt.
      • +
          +
        • Completion mode normally wont have a system prompt.
        • +
      • Enter your query to ai assistant below.
      • -
      • + Completion mode doesnt insert user/role: prefix implicitly.
      • -
      • + Use shift+enter for inserting enter/newline.
      • +
          +
        • Completion mode doesnt insert user/role: prefix implicitly.
        • +
        • Use shift+enter for inserting enter/newline.
        • +
      • Refresh the page to start over fresh.
      • +
          +
        • old chat is not culled when sending to server/ai-model.
        • +
        • either use New CHAT, or refresh if chat getting long.
        • +
      `; From cbd853eda989269140778df147c1430369bc89c1 Mon Sep 17 00:00:00 2001 From: HanishKVC Date: Thu, 23 May 2024 18:51:36 +0530 Subject: [PATCH 12/19] SimpleChat:ChatRequestOptions: max_tokens As some times based on the query from the user, the ai model may get into a run away kind of generation with repeatations etal, so adding max_tokens to try and limit this run away behaviour, if possible. --- examples/server/public_simplechat/simplechat.js | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/examples/server/public_simplechat/simplechat.js b/examples/server/public_simplechat/simplechat.js index 2ddab0ec2b7f5..a93a89105fc75 100644 --- a/examples/server/public_simplechat/simplechat.js +++ b/examples/server/public_simplechat/simplechat.js @@ -34,7 +34,8 @@ let gUsageMsg = ` // Add needed fields wrt json object to be sent wrt LLM web services completions endpoint. let gChatRequestOptions = { - "temperature": 0.7 + "temperature": 0.7, + "max_tokens": 2048 }; class SimpleChat { From 4b29736da588019a0648d6e14d35e56410816028 Mon Sep 17 00:00:00 2001 From: HanishKVC Date: Thu, 23 May 2024 20:55:11 +0530 Subject: [PATCH 13/19] SimpleChat: Reduce max_tokens to be small but still sufficient --- examples/server/public_simplechat/simplechat.js | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/examples/server/public_simplechat/simplechat.js b/examples/server/public_simplechat/simplechat.js index a93a89105fc75..303886b262a4c 100644 --- a/examples/server/public_simplechat/simplechat.js +++ b/examples/server/public_simplechat/simplechat.js @@ -35,7 +35,7 @@ let gUsageMsg = ` // Add needed fields wrt json object to be sent wrt LLM web services completions endpoint. let gChatRequestOptions = { "temperature": 0.7, - "max_tokens": 2048 + "max_tokens": 512 }; class SimpleChat { From f0dd91d550ab5337538e326a1f837573397e030e Mon Sep 17 00:00:00 2001 From: HanishKVC Date: Thu, 23 May 2024 21:18:48 +0530 Subject: [PATCH 14/19] SimpleChat: Consolidate global vars into gMe, Display to user This allows the end user to see the settings used by the logic, as well as allows users to change/update the settings if they want to by using devel-tools/console --- examples/server/public_simplechat/readme.md | 19 +++-- .../server/public_simplechat/simplechat.js | 72 ++++++++++++++----- 2 files changed, 68 insertions(+), 23 deletions(-) diff --git a/examples/server/public_simplechat/readme.md b/examples/server/public_simplechat/readme.md index 72a428075214d..d8818b7383686 100644 --- a/examples/server/public_simplechat/readme.md +++ b/examples/server/public_simplechat/readme.md @@ -93,12 +93,21 @@ Once inside ## Devel note -gChatRequestOptions maintains the list of options/fields to send along with chat request, -irrespective of whether /chat/completions or /completions endpoint. +Me/gMe consolidates the settings which control the behaviour into one object. +One can see the current settings, as well as change/update them using browsers devel-tool/console. - If you want to add additional options/fields to send to the server/ai-model, and or - modify the existing options value, for now you can update this global var using - browser's development-tools/console. + bCompletionFreshChatAlways - whether Completion mode collates completion history when communicating + with the server. + + bCompletionInsertStandardRolePrefix - whether Completion mode inserts role related prefix wrt the + messages that get inserted into prompt field wrt /Completion endpoint. + + chatRequestOptions - maintains the list of options/fields to send along with chat request, + irrespective of whether /chat/completions or /completions endpoint. + + If you want to add additional options/fields to send to the server/ai-model, and or + modify the existing options value, for now you can update this global var using + browser's development-tools/console. Sometimes the browser may be stuborn with caching of the file, so your updates to html/css/js may not be visible. Also remember that just refreshing/reloading page in browser or for that diff --git a/examples/server/public_simplechat/simplechat.js b/examples/server/public_simplechat/simplechat.js index 303886b262a4c..97a300eac0184 100644 --- a/examples/server/public_simplechat/simplechat.js +++ b/examples/server/public_simplechat/simplechat.js @@ -14,6 +14,7 @@ class ApiEP { } let gUsageMsg = ` +

      Usage

      • Set system prompt above, to try control ai response charactersitic, if model supports same.
        • @@ -32,11 +33,6 @@ let gUsageMsg = `
        `; -// Add needed fields wrt json object to be sent wrt LLM web services completions endpoint. -let gChatRequestOptions = { - "temperature": 0.7, - "max_tokens": 512 -}; class SimpleChat { @@ -92,6 +88,7 @@ class SimpleChat { } else { if (bClear) { div.innerHTML = gUsageMsg; + gMe.show_info(div); } } } @@ -103,8 +100,8 @@ class SimpleChat { * @param {Object} obj */ request_jsonstr(obj) { - for(let k in gChatRequestOptions) { - obj[k] = gChatRequestOptions[k]; + for(let k in gMe.chatRequestOptions) { + obj[k] = gMe.chatRequestOptions[k]; } return JSON.stringify(obj); } @@ -206,8 +203,6 @@ let gChatURL = { 'chat': `${gBaseURL}/chat/completions`, 'completion': `${gBaseURL}/completions`, } -const gbCompletionFreshChatAlways = true; -let gbCompletionInsertStandardRolePrefix = false; /** @@ -395,7 +390,7 @@ class MultiChatUI { // So if user wants to simulate a multi-chat based completion query, // they will have to enter the full thing, as a suitable multiline // user input/query. - if ((apiEP == ApiEP.Completion) && (gbCompletionFreshChatAlways)) { + if ((apiEP == ApiEP.Completion) && (gMe.bCompletionFreshChatAlways)) { chat.clear(); } @@ -413,7 +408,7 @@ class MultiChatUI { if (apiEP == ApiEP.Chat) { theBody = chat.request_messages_jsonstr(); } else { - theBody = chat.request_prompt_jsonstr(gbCompletionInsertStandardRolePrefix); + theBody = chat.request_prompt_jsonstr(gMe.bCompletionInsertStandardRolePrefix); } this.elInUser.value = "working..."; @@ -525,17 +520,58 @@ class MultiChatUI { } -let gMultiChat; -const gChatIds = [ "Default", "Other" ]; +class Me { + + constructor() { + this.defaultChatIds = [ "Default", "Other" ]; + this.multiChat = new MultiChatUI(); + this.bCompletionFreshChatAlways = true; + this.bCompletionInsertStandardRolePrefix = false; + // Add needed fields wrt json object to be sent wrt LLM web services completions endpoint. + this.chatRequestOptions = { + "temperature": 0.7, + "max_tokens": 512 + }; + } + + /** + * @param {HTMLDivElement} elDiv + */ + show_info(elDiv) { + + var p = document.createElement("p"); + p.innerText = "Settings (gMe)"; + p.className = "role-system"; + elDiv.appendChild(p); + + var p = document.createElement("p"); + p.innerText = `bCompletionFreshChatAlways:${this.bCompletionFreshChatAlways}`; + elDiv.appendChild(p); + + p = document.createElement("p"); + p.innerText = `bCompletionInsertStandardRolePrefix:${this.bCompletionInsertStandardRolePrefix}`; + elDiv.appendChild(p); + + p = document.createElement("p"); + p.innerText = `chatRequestOptions:${JSON.stringify(this.chatRequestOptions)}`; + elDiv.appendChild(p); + + } + +} + + +/** @type {Me} */ +let gMe; function startme() { console.log("INFO:SimpleChat:StartMe:Starting..."); - gMultiChat = new MultiChatUI(); - for (let cid of gChatIds) { - gMultiChat.new_chat_session(cid); + gMe = new Me(); + for (let cid of gMe.defaultChatIds) { + gMe.multiChat.new_chat_session(cid); } - gMultiChat.setup_ui(gChatIds[0], true); - gMultiChat.show_sessions(); + gMe.multiChat.setup_ui(gMe.defaultChatIds[0], true); + gMe.multiChat.show_sessions(); } document.addEventListener("DOMContentLoaded", startme); From b57aad79a816f823123fe71d4e28ad5329965dad Mon Sep 17 00:00:00 2001 From: HanishKVC Date: Fri, 24 May 2024 01:05:05 +0530 Subject: [PATCH 15/19] SimpleChat:SlidingWindow: iRecentUserMsgCnt to limit context load This is disabled by default. However if enabled, then in addition to latest system message, only the last N user messages, after the latest system message and its reponses from the ai model will be sent to the ai-model, when querying for a new response. This specified N also includes the latest user query. --- examples/server/public_simplechat/readme.md | 43 ++++++++++-- .../server/public_simplechat/simplechat.js | 67 ++++++++++++++++--- 2 files changed, 95 insertions(+), 15 deletions(-) diff --git a/examples/server/public_simplechat/readme.md b/examples/server/public_simplechat/readme.md index d8818b7383686..43d5b127f5aab 100644 --- a/examples/server/public_simplechat/readme.md +++ b/examples/server/public_simplechat/readme.md @@ -14,11 +14,15 @@ own system prompts. The UI follows a responsive web design so that the layout can adapt to available display space in a usable enough manner, in general. +Allows developer/end-user to control some of the behaviour by updating gMe members from browser's devel-tool +console. + NOTE: Given that the idea is for basic minimal testing, it doesnt bother with any model context length and -culling of old messages from the chat. +culling of old messages from the chat by default. However by enabling the sliding window chat logic, a crude +form of old messages culling can be achieved. -NOTE: It doesnt set any parameters other than temperature for now. However if someone wants they can update -the js file as needed. +NOTE: It doesnt set any parameters other than temperature and max_tokens for now. However if someone wants +they can update the js file or equivalent member in gMe as needed. ## usage @@ -96,8 +100,8 @@ Once inside Me/gMe consolidates the settings which control the behaviour into one object. One can see the current settings, as well as change/update them using browsers devel-tool/console. - bCompletionFreshChatAlways - whether Completion mode collates completion history when communicating - with the server. + bCompletionFreshChatAlways - whether Completion mode collates complete/sliding-window history when + communicating with the server or only sends the latest user query/message. bCompletionInsertStandardRolePrefix - whether Completion mode inserts role related prefix wrt the messages that get inserted into prompt field wrt /Completion endpoint. @@ -106,22 +110,42 @@ One can see the current settings, as well as change/update them using browsers d irrespective of whether /chat/completions or /completions endpoint. If you want to add additional options/fields to send to the server/ai-model, and or - modify the existing options value, for now you can update this global var using - browser's development-tools/console. + modify the existing options value or remove them, for now you can update this global var + using browser's development-tools/console. + + iRecentUserMsgCnt - a simple minded SlidingWindow to limit context window load at Ai Model end. + This is disabled by default. However if enabled, then in addition to latest system message, only + the last/latest iRecentUserMsgCnt user messages after the latest system prompt and its responses + from the ai model will be sent to the ai-model, when querying for a new response. IE if enabled, + only user messages after the latest system message/prompt will be considered. + + This specified sliding window user message count also includes the latest user query. + <0 : Send entire chat history to server + 0 : Send only the system message if any to the server + >0 : Send the latest chat history from the latest system prompt, limited to specified cnt. + + +By using gMe's iRecentUserMsgCnt and chatRequestOptions.max_tokens one can try to control the +implications of loading of the ai-model's context window by chat history, wrt chat response to +some extent in a simple crude way. + Sometimes the browser may be stuborn with caching of the file, so your updates to html/css/js may not be visible. Also remember that just refreshing/reloading page in browser or for that matter clearing site data, dont directly override site caching in all cases. Worst case you may have to change port. Or in dev tools of browser, you may be able to disable caching fully. + Concept of multiple chat sessions with different servers, as well as saving and restoring of those across browser usage sessions, can be woven around the SimpleChat/MultiChatUI class and its instances relatively easily, however given the current goal of keeping this simple, it has not been added, for now. + By switching between chat.add_system_begin/anytime, one can control whether one can change the system prompt, anytime during the conversation or only at the beginning. + read_json_early, is to experiment with reading json response data early on, if available, so that user can be shown generated data, as and when it is being generated, rather than at the end when full data is available. @@ -132,3 +156,8 @@ at the end when full data is available. if able to read json data early on in future, as and when ai model is generating data, then this helper needs to indirectly update the chat div with the recieved data, without waiting for the overall data to be available. + + +## At the end + +Also a thank you to all open source and open model developers, who strive for the common good. diff --git a/examples/server/public_simplechat/simplechat.js b/examples/server/public_simplechat/simplechat.js index 97a300eac0184..52b28ad824251 100644 --- a/examples/server/public_simplechat/simplechat.js +++ b/examples/server/public_simplechat/simplechat.js @@ -25,21 +25,23 @@ let gUsageMsg = `
      • Completion mode doesnt insert user/role: prefix implicitly.
      • Use shift+enter for inserting enter/newline.
      -
    • Refresh the page to start over fresh.
    • +
    • If strange responses, Refresh page to start over fresh.
      • -
      • old chat is not culled when sending to server/ai-model.
      • -
      • either use New CHAT, or refresh if chat getting long.
      • +
      • [default] old msgs from chat not culled, when sending to server.
      • +
      • either use New CHAT, or refresh if chat getting long, or
      • +
      • experiment iRecentUserMsgCnt, max_tokens, model ctxt window.
    `; +/** @typedef {{role: string, content: string}[]} ChatMessages */ class SimpleChat { constructor() { /** * Maintain in a form suitable for common LLM web service chat/completions' messages entry - * @type {{role: string, content: string}[]} + * @type {ChatMessages} */ this.xchat = []; this.iLastSys = -1; @@ -50,6 +52,50 @@ class SimpleChat { this.iLastSys = -1; } + /** + * Recent chat messages. + * If iRecentUserMsgCnt < 0 + * Then return the full chat history + * Else + * Return chat messages from latest going back till the last/latest system prompt. + * While keeping track that the number of user queries/messages doesnt exceed iRecentUserMsgCnt. + * @param {number} iRecentUserMsgCnt + */ + recent_chat(iRecentUserMsgCnt) { + if (iRecentUserMsgCnt < 0) { + return this.xchat; + } + if (iRecentUserMsgCnt == 0) { + console.warn("WARN:SimpleChat:SC:RecentChat:iRecentUsermsgCnt of 0 means no user message/query sent"); + } + /** @type{ChatMessages} */ + let rchat = []; + let sysMsg = this.get_system_latest(); + if (sysMsg.length != 0) { + rchat.push({role: Roles.System, content: sysMsg}); + } + let iUserCnt = 0; + let iStart = this.xchat.length; + for(let i=this.xchat.length-1; i > this.iLastSys; i--) { + if (iUserCnt >= iRecentUserMsgCnt) { + break; + } + let msg = this.xchat[i]; + if (msg.role == Roles.User) { + iStart = i; + iUserCnt += 1; + } + } + for(let i = iStart; i < this.xchat.length; i++) { + let msg = this.xchat[i]; + if (msg.role == Roles.System) { + continue; + } + rchat.push({role: msg.role, content: msg.content}); + } + return rchat; + } + /** * Add an entry into xchat * @param {string} role @@ -76,7 +122,7 @@ class SimpleChat { div.replaceChildren(); } let last = undefined; - for(const x of this.xchat) { + for(const x of this.recent_chat(gMe.iRecentUserMsgCnt)) { let entry = document.createElement("p"); entry.className = `role-${x.role}`; entry.innerText = `${x.role}: ${x.content}`; @@ -111,7 +157,7 @@ class SimpleChat { */ request_messages_jsonstr() { let req = { - messages: this.xchat, + messages: this.recent_chat(gMe.iRecentUserMsgCnt), } return this.request_jsonstr(req); } @@ -123,7 +169,7 @@ class SimpleChat { request_prompt_jsonstr(bInsertStandardRolePrefix) { let prompt = ""; let iCnt = 0; - for(const chat of this.xchat) { + for(const chat of this.recent_chat(gMe.iRecentUserMsgCnt)) { iCnt += 1; if (iCnt > 1) { prompt += "\n"; @@ -527,6 +573,7 @@ class Me { this.multiChat = new MultiChatUI(); this.bCompletionFreshChatAlways = true; this.bCompletionInsertStandardRolePrefix = false; + this.iRecentUserMsgCnt = -1; // Add needed fields wrt json object to be sent wrt LLM web services completions endpoint. this.chatRequestOptions = { "temperature": 0.7, @@ -540,7 +587,7 @@ class Me { show_info(elDiv) { var p = document.createElement("p"); - p.innerText = "Settings (gMe)"; + p.innerText = "Settings (devel-tools-console gMe)"; p.className = "role-system"; elDiv.appendChild(p); @@ -552,6 +599,10 @@ class Me { p.innerText = `bCompletionInsertStandardRolePrefix:${this.bCompletionInsertStandardRolePrefix}`; elDiv.appendChild(p); + p = document.createElement("p"); + p.innerText = `iRecentUserMsgCnt:${this.iRecentUserMsgCnt}`; + elDiv.appendChild(p); + p = document.createElement("p"); p.innerText = `chatRequestOptions:${JSON.stringify(this.chatRequestOptions)}`; elDiv.appendChild(p); From 11d2d316220614d3d910ecb8f85fc24192dd8159 Mon Sep 17 00:00:00 2001 From: HanishKVC Date: Fri, 24 May 2024 04:18:04 +0530 Subject: [PATCH 16/19] SimpleChat: placeholder based usage hint for user-in textarea --- examples/server/public_simplechat/index.html | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/examples/server/public_simplechat/index.html b/examples/server/public_simplechat/index.html index 89b32f3ff3d54..1a1a342089ba3 100644 --- a/examples/server/public_simplechat/index.html +++ b/examples/server/public_simplechat/index.html @@ -40,7 +40,7 @@
    - +
    From 8f172b90709fae38749a8c74ed14344c27bb8b73 Mon Sep 17 00:00:00 2001 From: HanishKVC Date: Fri, 24 May 2024 22:53:43 +0530 Subject: [PATCH 17/19] SimpleChat: Try make user experience better, if possible Reduce chat history context sent to the server/ai-model to be just the system-prompt, prev-user-request-and-ai-response and cur-user-request, instead of the previous full chat history. This way if there is any response with garbage/repeatation, it doesnt mess with things beyond the next question, in some ways. Increase max_tokens to 1024, so that a relatively large previous reponse doesnt eat up the space available wrt next query-response. However dont forget that the server when started should also be started with a model context size of 1k or more, to be on safe side. Add frequency and presence penalty fields set to 1.2 to the set of fields sent to server along with the user query. So that the model is partly set to try avoid repeating text in its response. --- examples/server/public_simplechat/readme.md | 23 +++++++++++++++++++ .../server/public_simplechat/simplechat.js | 12 +++++----- 2 files changed, 29 insertions(+), 6 deletions(-) diff --git a/examples/server/public_simplechat/readme.md b/examples/server/public_simplechat/readme.md index 43d5b127f5aab..0a1c28131a2a5 100644 --- a/examples/server/public_simplechat/readme.md +++ b/examples/server/public_simplechat/readme.md @@ -97,6 +97,8 @@ Once inside ## Devel note +### General + Me/gMe consolidates the settings which control the behaviour into one object. One can see the current settings, as well as change/update them using browsers devel-tool/console. @@ -158,6 +160,27 @@ at the end when full data is available. for the overall data to be available. +### Default setup + +By default things are setup to try and make the user experience a bit better, if possible. +However a developer when testing the server of ai-model may want to change these value. + +Using iRecentUserMsgCnt reduce chat history context sent to the server/ai-model to be +just the system-prompt, prev-user-request-and-ai-response and cur-user-request, instead of +full chat history. This way if there is any response with garbage/repeatation, it doesnt +mess with things beyond the next question/request/query, in some ways. + +Set max_tokens to 1024, so that a relatively large previous reponse doesnt eat up the space +available wrt next query-response. However dont forget that the server when started should +also be started with a model context size of 1k or more, to be on safe side. + +Frequency and presence penalty fields are set to 1.2 in the set of fields sent to server +along with the user query. So that the model is partly set to try avoid repeating text in +its response. + +A end-user can change these behaviour by editing gMe from browser's devel-tool/console. + + ## At the end Also a thank you to all open source and open model developers, who strive for the common good. diff --git a/examples/server/public_simplechat/simplechat.js b/examples/server/public_simplechat/simplechat.js index 52b28ad824251..973f6046e9f73 100644 --- a/examples/server/public_simplechat/simplechat.js +++ b/examples/server/public_simplechat/simplechat.js @@ -25,11 +25,9 @@ let gUsageMsg = `
  • Completion mode doesnt insert user/role: prefix implicitly.
  • Use shift+enter for inserting enter/newline.
  • -
  • If strange responses, Refresh page to start over fresh.
  • +
  • Default ContextWindow = [System, Last Query+Resp, Cur Query].
    • -
    • [default] old msgs from chat not culled, when sending to server.
    • -
    • either use New CHAT, or refresh if chat getting long, or
    • -
    • experiment iRecentUserMsgCnt, max_tokens, model ctxt window.
    • +
    • experiment iRecentUserMsgCnt, max_tokens, model ctxt window to expand
    `; @@ -573,11 +571,13 @@ class Me { this.multiChat = new MultiChatUI(); this.bCompletionFreshChatAlways = true; this.bCompletionInsertStandardRolePrefix = false; - this.iRecentUserMsgCnt = -1; + this.iRecentUserMsgCnt = 2; // Add needed fields wrt json object to be sent wrt LLM web services completions endpoint. this.chatRequestOptions = { "temperature": 0.7, - "max_tokens": 512 + "max_tokens": 1024, + "frequency_penalty": 1.2, + "presence_penalty": 1.2, }; } From b3afd6c86a324026a6206fce6d8a890b8696e5cf Mon Sep 17 00:00:00 2001 From: HanishKVC Date: Fri, 24 May 2024 23:16:55 +0530 Subject: [PATCH 18/19] SimpleChat:Add n_predict (equiv max_tokens) for llamacpp server The /completions endpoint of examples/server doesnt take max_tokens, instead it takes the internal n_predict, for now add the same on the client side, maybe later add max_tokens to /completions endpoint handling. --- examples/server/public_simplechat/readme.md | 4 ++++ examples/server/public_simplechat/simplechat.js | 1 + 2 files changed, 5 insertions(+) diff --git a/examples/server/public_simplechat/readme.md b/examples/server/public_simplechat/readme.md index 0a1c28131a2a5..585ece8883cc2 100644 --- a/examples/server/public_simplechat/readme.md +++ b/examples/server/public_simplechat/readme.md @@ -174,6 +174,10 @@ Set max_tokens to 1024, so that a relatively large previous reponse doesnt eat u available wrt next query-response. However dont forget that the server when started should also be started with a model context size of 1k or more, to be on safe side. + The /completions endpoint of examples/server doesnt take max_tokens, instead it takes the + internal n_predict, for now add the same here on the client side, maybe later add max_tokens + to /completions endpoint handling code on server side. + Frequency and presence penalty fields are set to 1.2 in the set of fields sent to server along with the user query. So that the model is partly set to try avoid repeating text in its response. diff --git a/examples/server/public_simplechat/simplechat.js b/examples/server/public_simplechat/simplechat.js index 973f6046e9f73..0c48da8796fd8 100644 --- a/examples/server/public_simplechat/simplechat.js +++ b/examples/server/public_simplechat/simplechat.js @@ -578,6 +578,7 @@ class Me { "max_tokens": 1024, "frequency_penalty": 1.2, "presence_penalty": 1.2, + "n_predict": 1024 }; } From 6d2f3d9d51150c952ff629950fe4faea994cbbc3 Mon Sep 17 00:00:00 2001 From: HanishKVC Date: Sat, 25 May 2024 19:25:19 +0530 Subject: [PATCH 19/19] SimpleChat: Note about trying to keep things simple yet flexible --- examples/server/public_simplechat/readme.md | 11 +++++++++++ 1 file changed, 11 insertions(+) diff --git a/examples/server/public_simplechat/readme.md b/examples/server/public_simplechat/readme.md index 585ece8883cc2..de0dfc99de805 100644 --- a/examples/server/public_simplechat/readme.md +++ b/examples/server/public_simplechat/readme.md @@ -97,6 +97,17 @@ Once inside ## Devel note +### Reason behind this + +The idea is to be easy enough to use for basic purposes, while also being simple and easily discernable +by developers who may not be from web frontend background (so inturn may not be familiar with template / +end-use-specific-language-extensions driven flows) so that they can use it to explore/experiment things. + +And given that the idea is also to help explore/experiment for developers, some flexibility is provided +to change behaviour easily using the devel-tools/console, for now. And skeletal logic has been implemented +to explore some of the end points and ideas/implications around them. + + ### General Me/gMe consolidates the settings which control the behaviour into one object.