diff --git a/Jenkinsfile b/Jenkinsfile index 40200fbe11..df3517fa71 100644 --- a/Jenkinsfile +++ b/Jenkinsfile @@ -256,105 +256,6 @@ pipeline { } } } - - stage('Build-ML') { - steps { - script{ - startTime = currentBuild.duration - Exception ex = null - catchError(buildResult: 'FAILURE', stageResult: 'FAILURE') { - try { - sh ''' - tests/runtests_russian.sh MODE=clean - tests/runtests_multilingual.sh MODE=build - ''' - } - catch (Exception e) { - int duration = (currentBuild.duration - startTime) / 1000 - throw e - } - } - } - } - post { - failure { - script { - sh 'tests/runtests_multilingual.sh MODE=clean' - } - } - success { - script { - int duration = (currentBuild.duration - startTime) / 1000 - } - } - } - } - - stage('Start-ML') { - steps { - script { - startTime = currentBuild.duration - Exception ex = null - catchError(buildResult: 'FAILURE', stageResult: 'FAILURE') { - try { - sh 'tests/runtests_multilingual.sh MODE=clean && tests/runtests_multilingual.sh MODE=start' - } - catch (Exception e) { - int duration = (currentBuild.duration - startTime) / 1000 - throw e - } - } - } - } - post { - failure { - script { - sh 'tests/runtests_multilingual.sh MODE=clean' - } - } - success { - script { - started = true - int duration = (currentBuild.duration - startTime) / 1000 - } - } - aborted { - script { - sh 'tests/runtests_multilingual.sh MODE=clean' - } - } - } - } - - stage('Test skills-ML') { - steps { - script { - startTime = currentBuild.duration - Exception ex = null - catchError(buildResult: 'FAILURE', stageResult: 'FAILURE') { - try { - sh label: 'test skills', script: 'tests/runtests_multilingual.sh MODE=test_skills' - } - catch (Exception e) { - int duration = (currentBuild.duration - startTime) / 1000 - throw e - } - } - } - } - post { - success { - script { - int duration = (currentBuild.duration - startTime) / 1000 - } - } - aborted { - script { - sh 'tests/runtests_multilingual.sh MODE=clean' - } - } - } - } } post { aborted { diff --git a/annotators/relative_persona_extractor/test.py b/annotators/relative_persona_extractor/test.py index 4ab99aa2b1..ad5a87fa6f 100644 --- a/annotators/relative_persona_extractor/test.py +++ b/annotators/relative_persona_extractor/test.py @@ -19,19 +19,19 @@ def main(): } gold = [ { - "max_similarity": 0.695, + "max_similarity": 0.551, "persona": [ "I like Italian food especially pasta and pizza.", + "I like visiting interesting places virtually.", "I like to watch football and basketball on TV.", - "I like watching travel video blogs.", ], }, { - "max_similarity": 0.645, + "max_similarity": 0.474, "persona": [ + "I like visiting interesting places virtually.", "I like watching travel video blogs.", - "I like to watch football and basketball on TV.", - "I like Italian food especially pasta and pizza.", + "I like travelling.", ], }, ] diff --git a/assistant_dists/ai_faq_assistant/docker-compose.override.yml b/assistant_dists/ai_faq_assistant/docker-compose.override.yml index b5131f6d5c..dbff8b7a5e 100644 --- a/assistant_dists/ai_faq_assistant/docker-compose.override.yml +++ b/assistant_dists/ai_faq_assistant/docker-compose.override.yml @@ -133,7 +133,7 @@ services: args: SERVICE_PORT: 8128 SERVICE_NAME: sentence_ranker - PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/bert-base-nli-mean-tokens + PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/all-MiniLM-L6-v2 context: ./services/sentence_ranker/ command: flask run -h 0.0.0.0 -p 8128 environment: @@ -177,7 +177,7 @@ services: GENERATIVE_SERVICE_URL: http://transformers-lm-oasst12b:8158/respond GENERATIVE_SERVICE_CONFIG: default_generative_config.json GENERATIVE_TIMEOUT: 5 - N_UTTERANCES_CONTEXT: 3 + N_UTTERANCES_CONTEXT: 7 context: . dockerfile: ./skills/dff_template_prompted_skill/Dockerfile command: gunicorn --workers=1 server:app -b 0.0.0.0:8143 --reload diff --git a/assistant_dists/da_costa_clothes_assistant/docker-compose.override.yml b/assistant_dists/da_costa_clothes_assistant/docker-compose.override.yml index 6040c97d8c..0e662ce499 100644 --- a/assistant_dists/da_costa_clothes_assistant/docker-compose.override.yml +++ b/assistant_dists/da_costa_clothes_assistant/docker-compose.override.yml @@ -133,7 +133,7 @@ services: args: SERVICE_PORT: 8128 SERVICE_NAME: sentence_ranker - PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/bert-base-nli-mean-tokens + PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/all-MiniLM-L6-v2 context: ./services/sentence_ranker/ command: flask run -h 0.0.0.0 -p 8128 environment: @@ -177,7 +177,7 @@ services: GENERATIVE_SERVICE_URL: http://transformers-lm-oasst12b:8158/respond GENERATIVE_SERVICE_CONFIG: default_generative_config.json GENERATIVE_TIMEOUT: 5 - N_UTTERANCES_CONTEXT: 3 + N_UTTERANCES_CONTEXT: 7 context: . dockerfile: ./skills/dff_template_prompted_skill/Dockerfile command: gunicorn --workers=1 server:app -b 0.0.0.0:8144 --reload diff --git a/assistant_dists/deepy_assistant/docker-compose.override.yml b/assistant_dists/deepy_assistant/docker-compose.override.yml index 0859b8303c..82e981d5e4 100644 --- a/assistant_dists/deepy_assistant/docker-compose.override.yml +++ b/assistant_dists/deepy_assistant/docker-compose.override.yml @@ -133,7 +133,7 @@ services: args: SERVICE_PORT: 8128 SERVICE_NAME: sentence_ranker - PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/bert-base-nli-mean-tokens + PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/all-MiniLM-L6-v2 context: ./services/sentence_ranker/ command: flask run -h 0.0.0.0 -p 8128 environment: @@ -176,7 +176,7 @@ services: GENERATIVE_SERVICE_URL: http://openai-api-davinci3:8131/respond GENERATIVE_SERVICE_CONFIG: openai-text-davinci-003-long.json GENERATIVE_TIMEOUT: 7 - N_UTTERANCES_CONTEXT: 3 + N_UTTERANCES_CONTEXT: 7 ENVVARS_TO_SEND: OPENAI_API_KEY,OPENAI_ORGANIZATION context: . dockerfile: ./skills/dff_template_prompted_skill/Dockerfile diff --git a/assistant_dists/dream/docker-compose.override.yml b/assistant_dists/dream/docker-compose.override.yml index daed4763ec..89e6fc34c1 100644 --- a/assistant_dists/dream/docker-compose.override.yml +++ b/assistant_dists/dream/docker-compose.override.yml @@ -1276,7 +1276,7 @@ services: args: SERVICE_PORT: 8128 SERVICE_NAME: sentence_ranker - PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/bert-base-nli-mean-tokens + PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/all-MiniLM-L6-v2 context: ./services/sentence_ranker/ command: flask run -h 0.0.0.0 -p 8128 environment: diff --git a/assistant_dists/dream_mini/docker-compose.override.yml b/assistant_dists/dream_mini/docker-compose.override.yml index 923fcf776e..af3b4fb5e2 100644 --- a/assistant_dists/dream_mini/docker-compose.override.yml +++ b/assistant_dists/dream_mini/docker-compose.override.yml @@ -177,7 +177,7 @@ services: args: SERVICE_PORT: 8128 SERVICE_NAME: sentence_ranker - PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/bert-base-nli-mean-tokens + PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/all-MiniLM-L6-v2 context: ./services/sentence_ranker/ command: flask run -h 0.0.0.0 -p 8128 environment: diff --git a/assistant_dists/dream_mini_persona_based/docker-compose.override.yml b/assistant_dists/dream_mini_persona_based/docker-compose.override.yml index 50cb7ce157..e2083e35ce 100644 --- a/assistant_dists/dream_mini_persona_based/docker-compose.override.yml +++ b/assistant_dists/dream_mini_persona_based/docker-compose.override.yml @@ -179,7 +179,7 @@ services: args: SERVICE_PORT: 8128 SERVICE_NAME: sentence_ranker - PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/bert-base-nli-mean-tokens + PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/all-MiniLM-L6-v2 context: ./services/sentence_ranker/ command: flask run -h 0.0.0.0 -p 8128 environment: diff --git a/assistant_dists/dream_multimodal/docker-compose.override.yml b/assistant_dists/dream_multimodal/docker-compose.override.yml index f55e0ec73d..3dcd4c7b8e 100644 --- a/assistant_dists/dream_multimodal/docker-compose.override.yml +++ b/assistant_dists/dream_multimodal/docker-compose.override.yml @@ -181,7 +181,7 @@ services: args: SERVICE_PORT: 8128 SERVICE_NAME: sentence_ranker - PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/bert-base-nli-mean-tokens + PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/all-MiniLM-L6-v2 context: ./services/sentence_ranker/ command: flask run -h 0.0.0.0 -p 8128 environment: diff --git a/assistant_dists/dream_persona_openai_prompted/docker-compose.override.yml b/assistant_dists/dream_persona_openai_prompted/docker-compose.override.yml index 1a1056e6ca..6c71ecdd49 100644 --- a/assistant_dists/dream_persona_openai_prompted/docker-compose.override.yml +++ b/assistant_dists/dream_persona_openai_prompted/docker-compose.override.yml @@ -133,7 +133,7 @@ services: args: SERVICE_PORT: 8128 SERVICE_NAME: sentence_ranker - PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/bert-base-nli-mean-tokens + PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/all-MiniLM-L6-v2 context: ./services/sentence_ranker/ command: flask run -h 0.0.0.0 -p 8128 environment: @@ -176,7 +176,7 @@ services: GENERATIVE_SERVICE_URL: http://openai-api-davinci3:8131/respond GENERATIVE_SERVICE_CONFIG: openai-text-davinci-003.json GENERATIVE_TIMEOUT: 7 - N_UTTERANCES_CONTEXT: 3 + N_UTTERANCES_CONTEXT: 7 ENVVARS_TO_SEND: OPENAI_API_KEY,OPENAI_ORGANIZATION context: . dockerfile: ./skills/dff_template_prompted_skill/Dockerfile diff --git a/assistant_dists/dream_persona_prompted/docker-compose.override.yml b/assistant_dists/dream_persona_prompted/docker-compose.override.yml index d3d8abc9a3..c9bc3cfe97 100644 --- a/assistant_dists/dream_persona_prompted/docker-compose.override.yml +++ b/assistant_dists/dream_persona_prompted/docker-compose.override.yml @@ -133,7 +133,7 @@ services: args: SERVICE_PORT: 8128 SERVICE_NAME: sentence_ranker - PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/bert-base-nli-mean-tokens + PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/all-MiniLM-L6-v2 context: ./services/sentence_ranker/ command: flask run -h 0.0.0.0 -p 8128 environment: @@ -177,7 +177,7 @@ services: GENERATIVE_SERVICE_URL: http://transformers-lm-gptj:8130/respond GENERATIVE_SERVICE_CONFIG: default_generative_config.json GENERATIVE_TIMEOUT: 5 - N_UTTERANCES_CONTEXT: 3 + N_UTTERANCES_CONTEXT: 7 context: . dockerfile: ./skills/dff_template_prompted_skill/Dockerfile command: gunicorn --workers=1 server:app -b 0.0.0.0:8134 --reload diff --git a/assistant_dists/empathetic_marketing_assistant/docker-compose.override.yml b/assistant_dists/empathetic_marketing_assistant/docker-compose.override.yml index e9e1941336..0be59eff5b 100644 --- a/assistant_dists/empathetic_marketing_assistant/docker-compose.override.yml +++ b/assistant_dists/empathetic_marketing_assistant/docker-compose.override.yml @@ -133,7 +133,7 @@ services: args: SERVICE_PORT: 8128 SERVICE_NAME: sentence_ranker - PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/bert-base-nli-mean-tokens + PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/all-MiniLM-L6-v2 context: ./services/sentence_ranker/ command: flask run -h 0.0.0.0 -p 8128 environment: @@ -177,7 +177,7 @@ services: GENERATIVE_SERVICE_URL: http://transformers-lm-gptj:8130/respond GENERATIVE_SERVICE_CONFIG: default_generative_config.json GENERATIVE_TIMEOUT: 5 - N_UTTERANCES_CONTEXT: 3 + N_UTTERANCES_CONTEXT: 7 context: . dockerfile: ./skills/dff_template_prompted_skill/Dockerfile command: gunicorn --workers=1 server:app -b 0.0.0.0:8138 --reload diff --git a/assistant_dists/fairytale_assistant/docker-compose.override.yml b/assistant_dists/fairytale_assistant/docker-compose.override.yml index 1adbe95a82..0202e3dead 100644 --- a/assistant_dists/fairytale_assistant/docker-compose.override.yml +++ b/assistant_dists/fairytale_assistant/docker-compose.override.yml @@ -133,7 +133,7 @@ services: args: SERVICE_PORT: 8128 SERVICE_NAME: sentence_ranker - PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/bert-base-nli-mean-tokens + PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/all-MiniLM-L6-v2 context: ./services/sentence_ranker/ command: flask run -h 0.0.0.0 -p 8128 environment: @@ -176,7 +176,7 @@ services: GENERATIVE_SERVICE_URL: http://openai-api-chatgpt:8145/respond GENERATIVE_SERVICE_CONFIG: openai-chatgpt.json GENERATIVE_TIMEOUT: 7 - N_UTTERANCES_CONTEXT: 3 + N_UTTERANCES_CONTEXT: 7 ENVVARS_TO_SEND: OPENAI_API_KEY,OPENAI_ORGANIZATION context: . dockerfile: ./skills/dff_template_prompted_skill/Dockerfile diff --git a/assistant_dists/nutrition_assistant/docker-compose.override.yml b/assistant_dists/nutrition_assistant/docker-compose.override.yml index df99c55ae4..ee4e1a61df 100644 --- a/assistant_dists/nutrition_assistant/docker-compose.override.yml +++ b/assistant_dists/nutrition_assistant/docker-compose.override.yml @@ -133,7 +133,7 @@ services: args: SERVICE_PORT: 8128 SERVICE_NAME: sentence_ranker - PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/bert-base-nli-mean-tokens + PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/all-MiniLM-L6-v2 context: ./services/sentence_ranker/ command: flask run -h 0.0.0.0 -p 8128 environment: @@ -177,7 +177,7 @@ services: GENERATIVE_SERVICE_URL: http://transformers-lm-oasst12b:8158/respond GENERATIVE_SERVICE_CONFIG: default_generative_config.json GENERATIVE_TIMEOUT: 5 - N_UTTERANCES_CONTEXT: 3 + N_UTTERANCES_CONTEXT: 7 context: . dockerfile: ./skills/dff_template_prompted_skill/Dockerfile command: gunicorn --workers=1 server:app -b 0.0.0.0:8141 --reload diff --git a/assistant_dists/rhodes_coaching_assistant/docker-compose.override.yml b/assistant_dists/rhodes_coaching_assistant/docker-compose.override.yml index 1c7789a424..fa1c862762 100644 --- a/assistant_dists/rhodes_coaching_assistant/docker-compose.override.yml +++ b/assistant_dists/rhodes_coaching_assistant/docker-compose.override.yml @@ -133,7 +133,7 @@ services: args: SERVICE_PORT: 8128 SERVICE_NAME: sentence_ranker - PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/bert-base-nli-mean-tokens + PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/all-MiniLM-L6-v2 context: ./services/sentence_ranker/ command: flask run -h 0.0.0.0 -p 8128 environment: @@ -176,7 +176,7 @@ services: GENERATIVE_SERVICE_URL: http://openai-api-davinci3:8131/respond GENERATIVE_SERVICE_CONFIG: openai-text-davinci-003.json GENERATIVE_TIMEOUT: 7 - N_UTTERANCES_CONTEXT: 3 + N_UTTERANCES_CONTEXT: 7 ENVVARS_TO_SEND: OPENAI_API_KEY,OPENAI_ORGANIZATION context: . dockerfile: ./skills/dff_template_prompted_skill/Dockerfile diff --git a/assistant_dists/universal_prompted_assistant/docker-compose.override.yml b/assistant_dists/universal_prompted_assistant/docker-compose.override.yml index ba99cea0e4..4773cf165d 100644 --- a/assistant_dists/universal_prompted_assistant/docker-compose.override.yml +++ b/assistant_dists/universal_prompted_assistant/docker-compose.override.yml @@ -114,7 +114,7 @@ services: args: SERVICE_PORT: 8128 SERVICE_NAME: sentence_ranker - PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/bert-base-nli-mean-tokens + PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/all-MiniLM-L6-v2 context: ./services/sentence_ranker/ command: flask run -h 0.0.0.0 -p 8128 environment: @@ -216,7 +216,7 @@ services: SERVICE_PORT: 8147 SERVICE_NAME: dff_universal_prompted_skill GENERATIVE_TIMEOUT: 7 - N_UTTERANCES_CONTEXT: 3 + N_UTTERANCES_CONTEXT: 7 context: . dockerfile: ./skills/dff_universal_prompted_skill/Dockerfile command: gunicorn --workers=1 server:app -b 0.0.0.0:8147 --reload diff --git a/services/openai_api_lm/server.py b/services/openai_api_lm/server.py index 7d973c9511..0acd1484e4 100644 --- a/services/openai_api_lm/server.py +++ b/services/openai_api_lm/server.py @@ -71,8 +71,10 @@ def generate_responses(context, openai_api_key, openai_org, prompt, generation_p elif isinstance(response, str): outputs = [response.strip()] - outputs = [GENERATIVE_ROBOT_TEMPLATE.sub("\n", resp).strip() for resp in outputs] - outputs = [resp.split("\n")[0] for resp in outputs] + if PRETRAINED_MODEL_NAME_OR_PATH != "gpt-3.5-turbo": + # post-processing of the responses by all models except of ChatGPT + outputs = [GENERATIVE_ROBOT_TEMPLATE.sub("\n", resp).strip() for resp in outputs] + outputs = [resp.split("\n")[0] for resp in outputs] return outputs diff --git a/services/sentence_ranker/component.yml b/services/sentence_ranker/component.yml index 9ab8ebb94a..1bbfa10eb4 100644 --- a/services/sentence_ranker/component.yml +++ b/services/sentence_ranker/component.yml @@ -8,8 +8,8 @@ sentence-ranker: author: DeepPavlov description: ranking model given as `PRETRAINED_MODEL_NAME_OR_PATH` which for a pair os sentences returns a float score of correspondence - ram_usage: 1.2G - gpu_usage: 2.1G + ram_usage: 1.5G + gpu_usage: 800M port: 8128 endpoints: - group: candidate_annotators @@ -19,7 +19,7 @@ sentence-ranker: build_args: SERVICE_PORT: 8128 SERVICE_NAME: sentence_ranker - PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/bert-base-nli-mean-tokens + PRETRAINED_MODEL_NAME_OR_PATH: sentence-transformers/all-MiniLM-L6-v2 date_created: '2023-03-16T09:45:32' sentence-ranker-multilingual: name: sentence_ranker diff --git a/services/sentence_ranker/test.py b/services/sentence_ranker/test.py index bf5346f474..edceccd95d 100644 --- a/services/sentence_ranker/test.py +++ b/services/sentence_ranker/test.py @@ -5,19 +5,19 @@ def test_respond(): url = "http://0.0.0.0:8128/respond" sentence_pairs = [ - ["Привет! Как дела?", "хорошо. а у тебя как дела?"], - ["Привет! Как дела?", "какой твой любимый фильм?"], - ["Какой твой любимый фильм?", "Гордость и предубеждение"], - ["Какой твой любимый фильм?", "пересматриваю Гордость и предубеждение иногда."], - ["Какой твой любимый фильм?", "я люблю играть в компьютерные игры."], + ["Hi! How are you?", "Good. How are you?"], + ["Hi! How are you?", "what's your favorite movie?"], + ["What's your favorite movie?", "Pride and Prejudice"], + ["What's your favorite movie?", "I watch Pride and Prejudice sometimes."], + ["What's your favorite movie?", "I like to play computer games."], ] - gold = [0.8988315, 0.62241143, 0.65046525, 0.54038674, 0.48419473] + gold = [0.7672719955444336, 0.15072298049926758, 0.5039470791816711, 0.4652850031852722, 0.23436866700649261] request_data = {"sentence_pairs": sentence_pairs} result = requests.post(url, json=request_data).json()[0]["batch"] for i, score in enumerate(result): - assert score != 0.0, f"Expected:{gold[i]}\tGot\n{score}" + assert round(score, 2) == round(gold[i], 2), f"Expected:{gold[i]}\tGot\n{score}" print("Success!") diff --git a/skills/dff_universal_prompted_skill/scenario/response.py b/skills/dff_universal_prompted_skill/scenario/response.py index 466fd42dc3..c4b9a7324a 100644 --- a/skills/dff_universal_prompted_skill/scenario/response.py +++ b/skills/dff_universal_prompted_skill/scenario/response.py @@ -21,6 +21,7 @@ FIX_PUNCTUATION = re.compile(r"\s(?=[\.,:;])") DEFAULT_CONFIDENCE = 0.9 LOW_CONFIDENCE = 0.7 +DEFAULT_PROMPT = "Respond like a friendly chatbot." CONSIDERED_LM_SERVICES = { "GPT-J 6B": { @@ -55,6 +56,21 @@ def compose_data_for_model(ctx, actor): if context: context = [re.sub(FIX_PUNCTUATION, "", x) for x in context] + + # drop the dialog history when prompt changes + last_uttr = int_ctx.get_last_human_utterance(ctx, actor) + # get prompt from the current utterance attributes + given_prompt = last_uttr.get("attributes", {}).get("prompt", DEFAULT_PROMPT) + history = int_ctx.get_utterances(ctx, actor) + + for i in range(1, len(history) + 1, 2): + curr_prompt = history[-i].get("attributes", {}).get("prompt", DEFAULT_PROMPT) + # checking only user utterances + if curr_prompt != given_prompt: + # cut context on the last user utterance utilizing the current prompt + context = context[-i + 2 :] + break + return context @@ -79,7 +95,7 @@ def gathering_responses(reply, confidence, human_attr, bot_attr, attr): dialog_context = compose_data_for_model(ctx, actor) logger.info(f"dialog_context: {dialog_context}") last_uttr = int_ctx.get_last_human_utterance(ctx, actor) - prompt = last_uttr.get("attributes", {}).get("prompt", "Respond like a friendly chatbot.") + prompt = last_uttr.get("attributes", {}).get("prompt", DEFAULT_PROMPT) logger.info(f"prompt: {prompt}") lm_service = last_uttr.get("attributes", {}).get("lm_service", "GPT-J 6B") logger.info(f"lm_service: {lm_service}") diff --git a/state_formatters/dp_formatters.py b/state_formatters/dp_formatters.py index 21494b8ab0..f1eca4ef9c 100755 --- a/state_formatters/dp_formatters.py +++ b/state_formatters/dp_formatters.py @@ -1003,6 +1003,7 @@ def dff_ai_faq_prompted_skill_formatter(dialog): return utils.dff_formatter( dialog, "dff_ai_faq_prompted_skill", + bot_last_turns=5, types_utterances=["human_utterances", "bot_utterances", "utterances"], ) @@ -1011,6 +1012,7 @@ def dff_da_costa_clothes_prompted_skill_formatter(dialog): return utils.dff_formatter( dialog, "dff_da_costa_clothes_prompted_skill", + bot_last_turns=5, types_utterances=["human_utterances", "bot_utterances", "utterances"], ) @@ -1019,6 +1021,7 @@ def dff_dream_persona_prompted_skill_formatter(dialog): return utils.dff_formatter( dialog, "dff_dream_persona_prompted_skill", + bot_last_turns=5, types_utterances=["human_utterances", "bot_utterances", "utterances"], ) @@ -1027,6 +1030,7 @@ def dff_empathetic_marketing_prompted_skill_formatter(dialog): return utils.dff_formatter( dialog, "dff_empathetic_marketing_prompted_skill", + bot_last_turns=5, types_utterances=["human_utterances", "bot_utterances", "utterances"], ) @@ -1035,6 +1039,7 @@ def dff_fairytale_prompted_skill_formatter(dialog): return utils.dff_formatter( dialog, "dff_fairytale_prompted_skill", + bot_last_turns=5, types_utterances=["human_utterances", "bot_utterances", "utterances"], ) @@ -1043,6 +1048,7 @@ def dff_nutrition_prompted_skill_formatter(dialog): return utils.dff_formatter( dialog, "dff_nutrition_prompted_skill", + bot_last_turns=5, types_utterances=["human_utterances", "bot_utterances", "utterances"], ) @@ -1051,6 +1057,7 @@ def dff_rhodes_coaching_prompted_skill_formatter(dialog): return utils.dff_formatter( dialog, "dff_rhodes_coaching_prompted_skill", + bot_last_turns=5, types_utterances=["human_utterances", "bot_utterances", "utterances"], ) @@ -1059,6 +1066,7 @@ def dff_deepy_prompted_skill_formatter(dialog): return utils.dff_formatter( dialog, "dff_deepy_prompted_skill", + bot_last_turns=5, types_utterances=["human_utterances", "bot_utterances", "utterances"], ) @@ -1067,6 +1075,7 @@ def dff_prompted_skill_formatter(dialog, skill_name=None): return utils.dff_formatter( dialog, skill_name, + bot_last_turns=5, types_utterances=["human_utterances", "bot_utterances", "utterances"], ) @@ -1075,6 +1084,7 @@ def dff_universal_prompted_skill_formatter(dialog, skill_name=None): return utils.dff_formatter( dialog, "dff_universal_prompted_skill", + bot_last_turns=5, types_utterances=["human_utterances", "bot_utterances", "utterances"], wanted_keys=["text", "annotations", "active_skill", "user", "attributes"], )