From 88eeb0d7e6a800680ec9891e91087461026ffcad Mon Sep 17 00:00:00 2001 From: Letong Han <106566639+letonghan@users.noreply.github.com> Date: Thu, 8 Aug 2024 09:14:38 +0800 Subject: [PATCH] Remove LangSmith from Examples (#545) Signed-off-by: letonghan --- ChatQnA/docker/aipc/compose.yaml | 14 ++------------ ChatQnA/docker/gaudi/README.md | 19 ------------------- ChatQnA/docker/gaudi/compose.yaml | 12 ------------ ChatQnA/docker/gaudi/compose_guardrails.yaml | 14 +------------- ChatQnA/docker/gaudi/compose_vllm.yaml | 10 +--------- ChatQnA/docker/gaudi/compose_vllm_ray.yaml | 10 +--------- ChatQnA/docker/gpu/README.md | 19 ------------------- ChatQnA/docker/gpu/compose.yaml | 12 ------------ ChatQnA/docker/xeon/README.md | 19 ------------------- ChatQnA/docker/xeon/README_qdrant.md | 19 ------------------- ChatQnA/docker/xeon/compose.yaml | 12 ------------ .../docker/xeon/docker_compose_qdrant.yaml | 1 + ChatQnA/docker/xeon/docker_compose_vllm.yaml | 1 + CodeGen/docker/gaudi/README.md | 19 ------------------- CodeGen/docker/gaudi/compose.yaml | 3 --- CodeGen/docker/xeon/README.md | 19 ------------------- CodeGen/docker/xeon/compose.yaml | 3 --- CodeTrans/docker/gaudi/README.md | 19 ------------------- CodeTrans/docker/gaudi/compose.yaml | 3 --- CodeTrans/docker/xeon/README.md | 19 ------------------- CodeTrans/docker/xeon/compose.yaml | 3 --- DocSum/docker/gaudi/README.md | 19 ------------------- DocSum/docker/gaudi/compose.yaml | 3 --- DocSum/docker/xeon/README.md | 19 ------------------- DocSum/docker/xeon/compose.yaml | 3 --- FaqGen/docker/gaudi/README.md | 19 ------------------- FaqGen/docker/gaudi/compose.yaml | 3 --- FaqGen/docker/xeon/README.md | 19 ------------------- FaqGen/docker/xeon/compose.yaml | 3 --- SearchQnA/docker/gaudi/compose.yaml | 9 --------- SearchQnA/docker/xeon/compose.yaml | 9 --------- Translation/docker/xeon/compose.yaml | 3 --- 32 files changed, 7 insertions(+), 352 deletions(-) diff --git a/ChatQnA/docker/aipc/compose.yaml b/ChatQnA/docker/aipc/compose.yaml index b7df95b24..52afbc8bf 100644 --- a/ChatQnA/docker/aipc/compose.yaml +++ b/ChatQnA/docker/aipc/compose.yaml @@ -23,6 +23,8 @@ services: https_proxy: ${https_proxy} REDIS_URL: ${REDIS_URL} INDEX_NAME: ${INDEX_NAME} + TEI_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} + HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} tei-embedding-service: image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.5 container_name: tei-embedding-server @@ -49,9 +51,6 @@ services: http_proxy: ${http_proxy} https_proxy: ${https_proxy} TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-embedding-service" restart: unless-stopped retriever: image: opea/retriever-redis:latest @@ -68,9 +67,6 @@ services: REDIS_URL: ${REDIS_URL} INDEX_NAME: ${INDEX_NAME} TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-retriever-service" restart: unless-stopped tei-reranking-service: image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.5 @@ -104,9 +100,6 @@ services: HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_ENABLE_HF_TRANSFER: 0 - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-reranking-service" restart: unless-stopped llm: image: opea/llm-ollama @@ -122,9 +115,6 @@ services: HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_ENABLE_HF_TRANSFER: 0 - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-llm-service" OLLAMA_ENDPOINT: ${OLLAMA_ENDPOINT} chaqna-aipc-backend-server: image: opea/chatqna:latest diff --git a/ChatQnA/docker/gaudi/README.md b/ChatQnA/docker/gaudi/README.md index 20319876b..058dcd7ed 100644 --- a/ChatQnA/docker/gaudi/README.md +++ b/ChatQnA/docker/gaudi/README.md @@ -399,25 +399,6 @@ curl http://${host_ip}:9090/v1/guardrails\ -H 'Content-Type: application/json' ``` -## Enable LangSmith for Monotoring Application (Optional) - -LangSmith offers tools to debug, evaluate, and monitor language models and intelligent agents. It can be used to assess benchmark data for each microservice. Before launching your services with `docker compose -f compose.yaml up -d`, you need to enable LangSmith tracing by setting the `LANGCHAIN_TRACING_V2` environment variable to true and configuring your LangChain API key. - -Here's how you can do it: - -1. Install the latest version of LangSmith: - -```bash -pip install -U langsmith -``` - -2. Set the necessary environment variables: - -```bash -export LANGCHAIN_TRACING_V2=true -export LANGCHAIN_API_KEY=ls_... -``` - ## 🚀 Launch the UI To access the frontend, open the following URL in your browser: http://{host_ip}:5173. By default, the UI runs on port 5173 internally. If you prefer to use a different host port to access the frontend, you can modify the port mapping in the `compose.yaml` file as shown below: diff --git a/ChatQnA/docker/gaudi/compose.yaml b/ChatQnA/docker/gaudi/compose.yaml index 912430eaf..ead0b7db3 100644 --- a/ChatQnA/docker/gaudi/compose.yaml +++ b/ChatQnA/docker/gaudi/compose.yaml @@ -63,9 +63,6 @@ services: http_proxy: ${http_proxy} https_proxy: ${https_proxy} TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-embedding-service" restart: unless-stopped retriever: image: opea/retriever-redis:latest @@ -81,9 +78,6 @@ services: https_proxy: ${https_proxy} REDIS_URL: ${REDIS_URL} INDEX_NAME: ${INDEX_NAME} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-retriever-service" restart: unless-stopped tei-reranking-service: image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.5 @@ -117,9 +111,6 @@ services: HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_ENABLE_HF_TRANSFER: 0 - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-reranking-service" restart: unless-stopped tgi-service: image: ghcr.io/huggingface/tgi-gaudi:2.0.1 @@ -158,9 +149,6 @@ services: HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_ENABLE_HF_TRANSFER: 0 - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-llm-service" restart: unless-stopped chaqna-gaudi-backend-server: image: opea/chatqna:latest diff --git a/ChatQnA/docker/gaudi/compose_guardrails.yaml b/ChatQnA/docker/gaudi/compose_guardrails.yaml index f340d7858..b3c759064 100644 --- a/ChatQnA/docker/gaudi/compose_guardrails.yaml +++ b/ChatQnA/docker/gaudi/compose_guardrails.yaml @@ -28,6 +28,7 @@ services: REDIS_URL: ${REDIS_URL} INDEX_NAME: ${INDEX_NAME} TEI_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} + HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} tgi-guardrails-service: image: ghcr.io/huggingface/tgi-gaudi:2.0.1 container_name: tgi-guardrails-server @@ -61,7 +62,6 @@ services: https_proxy: ${https_proxy} SAFETY_GUARD_MODEL_ID: ${SAFETY_GUARD_MODEL_ID} SAFETY_GUARD_ENDPOINT: ${SAFETY_GUARD_ENDPOINT} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} restart: unless-stopped tei-embedding-service: @@ -96,9 +96,6 @@ services: http_proxy: ${http_proxy} https_proxy: ${https_proxy} TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-embedding-service" restart: unless-stopped retriever: image: opea/retriever-redis:latest @@ -114,9 +111,6 @@ services: https_proxy: ${https_proxy} REDIS_URL: ${REDIS_URL} INDEX_NAME: ${INDEX_NAME} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-retriever-service" restart: unless-stopped tei-reranking-service: image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.5 @@ -150,9 +144,6 @@ services: HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_ENABLE_HF_TRANSFER: 0 - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-reranking-service" restart: unless-stopped tgi-service: image: ghcr.io/huggingface/tgi-gaudi:2.0.1 @@ -191,9 +182,6 @@ services: HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_ENABLE_HF_TRANSFER: 0 - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-llm-service" restart: unless-stopped chaqna-gaudi-backend-server: image: opea/chatqna-guardrails:latest diff --git a/ChatQnA/docker/gaudi/compose_vllm.yaml b/ChatQnA/docker/gaudi/compose_vllm.yaml index 2e05a46f9..2f29a932f 100644 --- a/ChatQnA/docker/gaudi/compose_vllm.yaml +++ b/ChatQnA/docker/gaudi/compose_vllm.yaml @@ -28,6 +28,7 @@ services: REDIS_URL: ${REDIS_URL} INDEX_NAME: ${INDEX_NAME} TEI_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} + HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} tei-embedding-service: image: opea/tei-gaudi:latest container_name: tei-embedding-gaudi-server @@ -60,9 +61,6 @@ services: http_proxy: ${http_proxy} https_proxy: ${https_proxy} TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-embedding-service" restart: unless-stopped retriever: image: opea/retriever-redis:latest @@ -78,9 +76,6 @@ services: https_proxy: ${https_proxy} REDIS_URL: ${REDIS_URL} INDEX_NAME: ${INDEX_NAME} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-retriever-service" restart: unless-stopped tei-reranking-service: image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.2 @@ -114,9 +109,6 @@ services: HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_ENABLE_HF_TRANSFER: 0 - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-reranking-service" restart: unless-stopped vllm-service: image: vllm:hpu diff --git a/ChatQnA/docker/gaudi/compose_vllm_ray.yaml b/ChatQnA/docker/gaudi/compose_vllm_ray.yaml index ec898518d..9d954a0b9 100644 --- a/ChatQnA/docker/gaudi/compose_vllm_ray.yaml +++ b/ChatQnA/docker/gaudi/compose_vllm_ray.yaml @@ -28,6 +28,7 @@ services: REDIS_URL: ${REDIS_URL} INDEX_NAME: ${INDEX_NAME} TEI_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} + HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} tei-embedding-service: image: opea/tei-gaudi:latest container_name: tei-embedding-gaudi-server @@ -60,9 +61,6 @@ services: http_proxy: ${http_proxy} https_proxy: ${https_proxy} TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-embedding-service" restart: unless-stopped retriever: image: opea/retriever-redis:latest @@ -78,9 +76,6 @@ services: https_proxy: ${https_proxy} REDIS_URL: ${REDIS_URL} INDEX_NAME: ${INDEX_NAME} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-retriever-service" restart: unless-stopped tei-reranking-service: image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.2 @@ -114,9 +109,6 @@ services: HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_ENABLE_HF_TRANSFER: 0 - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-reranking-service" restart: unless-stopped vllm-ray-service: image: vllm_ray:habana diff --git a/ChatQnA/docker/gpu/README.md b/ChatQnA/docker/gpu/README.md index 9bcc5b42f..eff77ee17 100644 --- a/ChatQnA/docker/gpu/README.md +++ b/ChatQnA/docker/gpu/README.md @@ -243,25 +243,6 @@ curl -X POST "http://${host_ip}:6009/v1/dataprep/delete_file" \ -H "Content-Type: application/json" ``` -## Enable LangSmith for Monotoring Application (Optional) - -LangSmith offers tools to debug, evaluate, and monitor language models and intelligent agents. It can be used to assess benchmark data for each microservice. Before launching your services with `docker compose -f compose.yaml up -d`, you need to enable LangSmith tracing by setting the `LANGCHAIN_TRACING_V2` environment variable to true and configuring your LangChain API key. - -Here's how you can do it: - -1. Install the latest version of LangSmith: - -```bash -pip install -U langsmith -``` - -2. Set the necessary environment variables: - -```bash -export LANGCHAIN_TRACING_V2=true -export LANGCHAIN_API_KEY=ls_... -``` - ## 🚀 Launch the UI To access the frontend, open the following URL in your browser: http://{host_ip}:5173. By default, the UI runs on port 5173 internally. If you prefer to use a different host port to access the frontend, you can modify the port mapping in the `compose.yaml` file as shown below: diff --git a/ChatQnA/docker/gpu/compose.yaml b/ChatQnA/docker/gpu/compose.yaml index 036f64ba0..65682d384 100644 --- a/ChatQnA/docker/gpu/compose.yaml +++ b/ChatQnA/docker/gpu/compose.yaml @@ -64,9 +64,6 @@ services: http_proxy: ${http_proxy} https_proxy: ${https_proxy} TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-embedding-service" restart: unless-stopped retriever: image: opea/retriever-redis:latest @@ -82,9 +79,6 @@ services: https_proxy: ${https_proxy} REDIS_URL: ${REDIS_URL} INDEX_NAME: ${INDEX_NAME} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-retriever-service" restart: unless-stopped tei-reranking-service: image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.5 @@ -125,9 +119,6 @@ services: HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_ENABLE_HF_TRANSFER: 0 - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-reranking-service" restart: unless-stopped tgi-service: image: ghcr.io/huggingface/text-generation-inference:2.0 @@ -168,9 +159,6 @@ services: HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_ENABLE_HF_TRANSFER: 0 - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-llm-service" restart: unless-stopped chaqna-backend-server: image: opea/chatqna:latest diff --git a/ChatQnA/docker/xeon/README.md b/ChatQnA/docker/xeon/README.md index 63eef7370..8be8828a5 100644 --- a/ChatQnA/docker/xeon/README.md +++ b/ChatQnA/docker/xeon/README.md @@ -375,25 +375,6 @@ curl -X POST "http://${host_ip}:6009/v1/dataprep/delete_file" \ -H "Content-Type: application/json" ``` -## Enable LangSmith for Monotoring Application (Optional) - -LangSmith offers tools to debug, evaluate, and monitor language models and intelligent agents. It can be used to assess benchmark data for each microservice. Before launching your services with `docker compose -f compose.yaml up -d`, you need to enable LangSmith tracing by setting the `LANGCHAIN_TRACING_V2` environment variable to true and configuring your LangChain API key. - -Here's how you can do it: - -1. Install the latest version of LangSmith: - -```bash -pip install -U langsmith -``` - -2. Set the necessary environment variables: - -```bash -export LANGCHAIN_TRACING_V2=true -export LANGCHAIN_API_KEY=ls_... -``` - ## 🚀 Launch the UI To access the frontend, open the following URL in your browser: http://{host_ip}:5173. By default, the UI runs on port 5173 internally. If you prefer to use a different host port to access the frontend, you can modify the port mapping in the `compose.yaml` file as shown below: diff --git a/ChatQnA/docker/xeon/README_qdrant.md b/ChatQnA/docker/xeon/README_qdrant.md index 3b5b848f8..4fd3b4e8a 100644 --- a/ChatQnA/docker/xeon/README_qdrant.md +++ b/ChatQnA/docker/xeon/README_qdrant.md @@ -339,25 +339,6 @@ curl -X POST "http://${host_ip}:6009/v1/dataprep/delete_file" \ -H "Content-Type: application/json" ``` -## Enable LangSmith for Monotoring Application (Optional) - -LangSmith offers tools to debug, evaluate, and monitor language models and intelligent agents. It can be used to assess benchmark data for each microservice. Before launching your services with `docker compose -f compose.yaml up -d`, you need to enable LangSmith tracing by setting the `LANGCHAIN_TRACING_V2` environment variable to true and configuring your LangChain API key. - -Here's how you can do it: - -1. Install the latest version of LangSmith: - -```bash -pip install -U langsmith -``` - -2. Set the necessary environment variables: - -```bash -export LANGCHAIN_TRACING_V2=true -export LANGCHAIN_API_KEY=ls_... -``` - ## 🚀 Launch the UI To access the frontend, open the following URL in your browser: http://{host_ip}:5173. By default, the UI runs on port 5173 internally. If you prefer to use a different host port to access the frontend, you can modify the port mapping in the `compose.yaml` file as shown below: diff --git a/ChatQnA/docker/xeon/compose.yaml b/ChatQnA/docker/xeon/compose.yaml index 2417a97dd..b865ab38e 100644 --- a/ChatQnA/docker/xeon/compose.yaml +++ b/ChatQnA/docker/xeon/compose.yaml @@ -56,9 +56,6 @@ services: http_proxy: ${http_proxy} https_proxy: ${https_proxy} TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-embedding-service" restart: unless-stopped retriever: image: opea/retriever-redis:latest @@ -75,9 +72,6 @@ services: REDIS_URL: ${REDIS_URL} INDEX_NAME: ${INDEX_NAME} TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-retriever-service" restart: unless-stopped tei-reranking-service: image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.5 @@ -111,9 +105,6 @@ services: HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_ENABLE_HF_TRANSFER: 0 - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-reranking-service" restart: unless-stopped tgi-service: image: ghcr.io/huggingface/text-generation-inference:2.1.0 @@ -147,9 +138,6 @@ services: HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_ENABLE_HF_TRANSFER: 0 - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-llm-service" restart: unless-stopped chaqna-xeon-backend-server: image: opea/chatqna:latest diff --git a/ChatQnA/docker/xeon/docker_compose_qdrant.yaml b/ChatQnA/docker/xeon/docker_compose_qdrant.yaml index ee1c6ece9..e3ed77f4e 100644 --- a/ChatQnA/docker/xeon/docker_compose_qdrant.yaml +++ b/ChatQnA/docker/xeon/docker_compose_qdrant.yaml @@ -27,6 +27,7 @@ services: QDRANT_PORT: 6333 COLLECTION_NAME: ${INDEX_NAME} TEI_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} + HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} tei-embedding-service: image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.2 container_name: tei-embedding-server diff --git a/ChatQnA/docker/xeon/docker_compose_vllm.yaml b/ChatQnA/docker/xeon/docker_compose_vllm.yaml index 2caa891f0..4791b530f 100644 --- a/ChatQnA/docker/xeon/docker_compose_vllm.yaml +++ b/ChatQnA/docker/xeon/docker_compose_vllm.yaml @@ -28,6 +28,7 @@ services: REDIS_URL: ${REDIS_URL} INDEX_NAME: ${INDEX_NAME} TEI_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} + HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} tei-embedding-service: image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.2 container_name: tei-embedding-server diff --git a/CodeGen/docker/gaudi/README.md b/CodeGen/docker/gaudi/README.md index 7a236d17e..a3ffc8a72 100644 --- a/CodeGen/docker/gaudi/README.md +++ b/CodeGen/docker/gaudi/README.md @@ -134,25 +134,6 @@ curl http://${host_ip}:7778/v1/codegen -H "Content-Type: application/json" -d '{ }' ``` -## Enable LangSmith to Monitor Application (Optional) - -LangSmith offers tools to debug, evaluate, and monitor language models and intelligent agents. It can be used to assess benchmark data for each microservice. Before launching your services with `docker compose -f compose.yaml up -d`, you need to enable LangSmith tracing by setting the `LANGCHAIN_TRACING_V2` environment variable to true and configuring your LangChain API key. - -Here's how you can do it: - -1. Install the latest version of LangSmith: - -```bash -pip install -U langsmith -``` - -2. Set the necessary environment variables: - -```bash -export LANGCHAIN_TRACING_V2=true -export LANGCHAIN_API_KEY=ls_... -``` - ## 🚀 Launch the Svelte Based UI To access the frontend, open the following URL in your browser: `http://{host_ip}:5173`. By default, the UI runs on port 5173 internally. If you prefer to use a different host port to access the frontend, you can modify the port mapping in the `compose.yaml` file as shown below: diff --git a/CodeGen/docker/gaudi/compose.yaml b/CodeGen/docker/gaudi/compose.yaml index 477fd7c9a..074743015 100644 --- a/CodeGen/docker/gaudi/compose.yaml +++ b/CodeGen/docker/gaudi/compose.yaml @@ -38,9 +38,6 @@ services: https_proxy: ${https_proxy} TGI_LLM_ENDPOINT: ${TGI_LLM_ENDPOINT} HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-llm-service" restart: unless-stopped codegen-gaudi-backend-server: image: opea/codegen:latest diff --git a/CodeGen/docker/xeon/README.md b/CodeGen/docker/xeon/README.md index 4a8ae4ed4..29efec031 100644 --- a/CodeGen/docker/xeon/README.md +++ b/CodeGen/docker/xeon/README.md @@ -137,25 +137,6 @@ curl http://${host_ip}:7778/v1/codegen -H "Content-Type: application/json" -d '{ }' ``` -## Enable LangSmith for Monitoring Application (Optional) - -LangSmith offers tools to debug, evaluate, and monitor language models and intelligent agents. It can be used to assess benchmark data for each microservice. Before launching your services with `docker compose -f compose.yaml up -d`, you need to enable LangSmith tracing by setting the `LANGCHAIN_TRACING_V2` environment variable to true and configuring your LangChain API key. - -Here's how you can do it: - -1. Install the latest version of LangSmith: - -```bash -pip install -U langsmith -``` - -2. Set the necessary environment variables: - -```bash -export LANGCHAIN_TRACING_V2=true -export LANGCHAIN_API_KEY=ls_... -``` - ## 🚀 Launch the UI To access the frontend, open the following URL in your browser: `http://{host_ip}:5173`. By default, the UI runs on port 5173 internally. If you prefer to use a different host port to access the frontend, you can modify the port mapping in the `compose.yaml` file as shown below: diff --git a/CodeGen/docker/xeon/compose.yaml b/CodeGen/docker/xeon/compose.yaml index 935447d3d..b002a3984 100644 --- a/CodeGen/docker/xeon/compose.yaml +++ b/CodeGen/docker/xeon/compose.yaml @@ -33,9 +33,6 @@ services: https_proxy: ${https_proxy} TGI_LLM_ENDPOINT: ${TGI_LLM_ENDPOINT} HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-llm-service" restart: unless-stopped codegen-xeon-backend-server: image: opea/codegen:latest diff --git a/CodeTrans/docker/gaudi/README.md b/CodeTrans/docker/gaudi/README.md index 64458cb56..ab9c661ae 100755 --- a/CodeTrans/docker/gaudi/README.md +++ b/CodeTrans/docker/gaudi/README.md @@ -93,25 +93,6 @@ curl http://${host_ip}:7777/v1/codetrans \ -d '{"language_from": "Golang","language_to": "Python","source_code": "package main\n\nimport \"fmt\"\nfunc main() {\n fmt.Println(\"Hello, World!\");\n}"}' ``` -## Enable LangSmith to Monitor an Application (Optional) - -LangSmith offers tools to debug, evaluate, and monitor language models and intelligent agents. It can be used to assess benchmark data for each microservice. Before launching your services with `docker compose -f compose.yaml up -d`, you need to enable LangSmith tracing by setting the `LANGCHAIN_TRACING_V2` environment variable to true and configuring your LangChain API key. - -Here's how you can do it: - -1. Install the latest version of LangSmith: - -```bash -pip install -U langsmith -``` - -2. Set the necessary environment variables: - -```bash -export LANGCHAIN_TRACING_V2=true -export LANGCHAIN_API_KEY=ls_... -``` - ## 🚀 Launch the UI Open this URL `http://{host_ip}:5173` in your browser to access the frontend. diff --git a/CodeTrans/docker/gaudi/compose.yaml b/CodeTrans/docker/gaudi/compose.yaml index dd7a60d9f..88ba9dc72 100644 --- a/CodeTrans/docker/gaudi/compose.yaml +++ b/CodeTrans/docker/gaudi/compose.yaml @@ -36,9 +36,6 @@ services: https_proxy: ${https_proxy} TGI_LLM_ENDPOINT: ${TGI_LLM_ENDPOINT} HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-llm-service" restart: unless-stopped codetrans-gaudi-backend-server: image: opea/codetrans:latest diff --git a/CodeTrans/docker/xeon/README.md b/CodeTrans/docker/xeon/README.md index 88dcf051b..270733f34 100755 --- a/CodeTrans/docker/xeon/README.md +++ b/CodeTrans/docker/xeon/README.md @@ -100,22 +100,3 @@ curl http://${host_ip}:7777/v1/codetrans \ -H "Content-Type: application/json" \ -d '{"language_from": "Golang","language_to": "Python","source_code": "package main\n\nimport \"fmt\"\nfunc main() {\n fmt.Println(\"Hello, World!\");\n}"}' ``` - -## Enable LangSmith to Monitor an Application (Optional) - -LangSmith offers tools to debug, evaluate, and monitor language models and intelligent agents. It can be used to assess benchmark data for each microservice. Before launching your services with `docker compose -f compose.yaml up -d`, you need to enable LangSmith tracing by setting the `LANGCHAIN_TRACING_V2` environment variable to true and configuring your LangChain API key. - -Here's how you can do it: - -1. Install the latest version of LangSmith: - -```bash -pip install -U langsmith -``` - -2. Set the necessary environment variables: - -```bash -export LANGCHAIN_TRACING_V2=true -export LANGCHAIN_API_KEY=ls_... -``` diff --git a/CodeTrans/docker/xeon/compose.yaml b/CodeTrans/docker/xeon/compose.yaml index 5e9da8823..a236b8876 100644 --- a/CodeTrans/docker/xeon/compose.yaml +++ b/CodeTrans/docker/xeon/compose.yaml @@ -31,9 +31,6 @@ services: https_proxy: ${https_proxy} TGI_LLM_ENDPOINT: ${TGI_LLM_ENDPOINT} HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-llm-service" restart: unless-stopped codetrans-xeon-backend-server: image: opea/codetrans:latest diff --git a/DocSum/docker/gaudi/README.md b/DocSum/docker/gaudi/README.md index d61eb99be..d7a3202e5 100644 --- a/DocSum/docker/gaudi/README.md +++ b/DocSum/docker/gaudi/README.md @@ -117,25 +117,6 @@ curl http://${host_ip}:8888/v1/docsum -H "Content-Type: application/json" -d '{ }' ``` -## Enable LangSmith to Monitor an Application (Optional) - -LangSmith offers a suite of tools to debug, evaluate, and monitor language models and intelligent agents. It can be used to assess benchmark data for each microservice. Before launching your services with `docker compose up -d`, you need to enable LangSmith tracing by setting the `LANGCHAIN_TRACING_V2` environment variable to true and configuring your LangChain API key. - -Here's how you can do it: - -1. Install the latest version of LangSmith: - -```bash -pip install -U langsmith -``` - -2. Set the necessary environment variables: - -```bash -export LANGCHAIN_TRACING_V2=true -export LANGCHAIN_API_KEY=ls_... -``` - ## 🚀 Launch the Svelte UI Open this URL `http://{host_ip}:5173` in your browser to access the frontend. diff --git a/DocSum/docker/gaudi/compose.yaml b/DocSum/docker/gaudi/compose.yaml index e2f797ed8..2a2a8dbaa 100644 --- a/DocSum/docker/gaudi/compose.yaml +++ b/DocSum/docker/gaudi/compose.yaml @@ -37,9 +37,6 @@ services: https_proxy: ${https_proxy} TGI_LLM_ENDPOINT: ${TGI_LLM_ENDPOINT} HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-llm-service" restart: unless-stopped docsum-gaudi-backend-server: image: opea/docsum:latest diff --git a/DocSum/docker/xeon/README.md b/DocSum/docker/xeon/README.md index e17e8ce17..c272d1619 100644 --- a/DocSum/docker/xeon/README.md +++ b/DocSum/docker/xeon/README.md @@ -126,25 +126,6 @@ curl http://${host_ip}:8888/v1/docsum -H "Content-Type: application/json" -d '{ Following the validation of all aforementioned microservices, we are now prepared to construct a mega-service. -## Enable LangSmith to Monitor an Application (Optional) - -LangSmith offers tools to debug, evaluate, and monitor language models and intelligent agents. It can be used to assess benchmark data for each microservice. Before launching your services with `docker compose -f compose.yaml up -d`, you need to enable LangSmith tracing by setting the `LANGCHAIN_TRACING_V2` environment variable to true and configuring your LangChain API key. - -Here's how you can do it: - -1. Install the latest version of LangSmith: - -```bash -pip install -U langsmith -``` - -2. Set the necessary environment variables: - -```bash -export LANGCHAIN_TRACING_V2=true -export LANGCHAIN_API_KEY=ls_... -``` - ## 🚀 Launch the UI Open this URL `http://{host_ip}:5173` in your browser to access the svelte based frontend. diff --git a/DocSum/docker/xeon/compose.yaml b/DocSum/docker/xeon/compose.yaml index e8571c146..facd70463 100644 --- a/DocSum/docker/xeon/compose.yaml +++ b/DocSum/docker/xeon/compose.yaml @@ -34,9 +34,6 @@ services: https_proxy: ${https_proxy} TGI_LLM_ENDPOINT: ${TGI_LLM_ENDPOINT} HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-llm-service" restart: unless-stopped docsum-xeon-backend-server: image: opea/docsum:latest diff --git a/FaqGen/docker/gaudi/README.md b/FaqGen/docker/gaudi/README.md index 7d487b2a3..12124cd1c 100644 --- a/FaqGen/docker/gaudi/README.md +++ b/FaqGen/docker/gaudi/README.md @@ -117,25 +117,6 @@ curl http://${host_ip}:8888/v1/faqgen -H "Content-Type: application/json" -d '{ }' ``` -## Enable LangSmith to Monitor an Application (Optional) - -LangSmith offers a suite of tools to debug, evaluate, and monitor language models and intelligent agents. It can be used to assess benchmark data for each microservice. Before launching your services with `docker compose -f compose.yaml up -d`, you need to enable LangSmith tracing by setting the `LANGCHAIN_TRACING_V2` environment variable to true and configuring your LangChain API key. - -Here's how you can do it: - -1. Install the latest version of LangSmith: - -```bash -pip install -U langsmith -``` - -2. Set the necessary environment variables: - -```bash -export LANGCHAIN_TRACING_V2=true -export LANGCHAIN_API_KEY=ls_... -``` - ## 🚀 Launch the UI Open this URL `http://{host_ip}:5173` in your browser to access the frontend. diff --git a/FaqGen/docker/gaudi/compose.yaml b/FaqGen/docker/gaudi/compose.yaml index e504a06b2..6a5ec4db6 100644 --- a/FaqGen/docker/gaudi/compose.yaml +++ b/FaqGen/docker/gaudi/compose.yaml @@ -37,9 +37,6 @@ services: https_proxy: ${https_proxy} TGI_LLM_ENDPOINT: ${TGI_LLM_ENDPOINT} HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-llm-service" restart: unless-stopped faqgen-gaudi-backend-server: image: opea/faqgen:latest diff --git a/FaqGen/docker/xeon/README.md b/FaqGen/docker/xeon/README.md index 19fc7f980..e86e6480b 100644 --- a/FaqGen/docker/xeon/README.md +++ b/FaqGen/docker/xeon/README.md @@ -118,25 +118,6 @@ curl http://${host_ip}:8888/v1/faqgen -H "Content-Type: application/json" -d '{ Following the validation of all aforementioned microservices, we are now prepared to construct a mega-service. -## Enable LangSmith to Monitor an Application (Optional) - -LangSmith offers tools to debug, evaluate, and monitor language models and intelligent agents. It can be used to assess benchmark data for each microservice. Before launching your services with `docker compose -f compose.yaml up -d`, you need to enable LangSmith tracing by setting the `LANGCHAIN_TRACING_V2` environment variable to true and configuring your LangChain API key. - -Here's how you can do it: - -1. Install the latest version of LangSmith: - -```bash -pip install -U langsmith -``` - -2. Set the necessary environment variables: - -```bash -export LANGCHAIN_TRACING_V2=true -export LANGCHAIN_API_KEY=ls_... -``` - ## 🚀 Launch the UI Open this URL `http://{host_ip}:5173` in your browser to access the frontend. diff --git a/FaqGen/docker/xeon/compose.yaml b/FaqGen/docker/xeon/compose.yaml index d5e29f778..b8f775962 100644 --- a/FaqGen/docker/xeon/compose.yaml +++ b/FaqGen/docker/xeon/compose.yaml @@ -33,9 +33,6 @@ services: https_proxy: ${https_proxy} TGI_LLM_ENDPOINT: ${TGI_LLM_ENDPOINT} HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-llm-service" restart: unless-stopped faqgen-xeon-backend-server: image: opea/faqgen:latest diff --git a/SearchQnA/docker/gaudi/compose.yaml b/SearchQnA/docker/gaudi/compose.yaml index eb9e38cef..dc191f5b8 100644 --- a/SearchQnA/docker/gaudi/compose.yaml +++ b/SearchQnA/docker/gaudi/compose.yaml @@ -39,9 +39,6 @@ services: http_proxy: ${http_proxy} https_proxy: ${https_proxy} TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-embedding-service" restart: unless-stopped web-retriever: image: opea/web-retriever-chroma:latest @@ -84,9 +81,6 @@ services: https_proxy: ${https_proxy} TEI_RERANKING_ENDPOINT: ${TEI_RERANKING_ENDPOINT} HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-reranking-service" restart: unless-stopped tgi-service: image: ghcr.io/huggingface/tgi-gaudi:2.0.1 @@ -125,9 +119,6 @@ services: HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} HF_HUB_DISABLE_PROGRESS_BARS: 1 HF_HUB_ENABLE_HF_TRANSFER: 0 - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-llm-service" restart: unless-stopped searchqna-gaudi-backend-server: image: opea/searchqna:latest diff --git a/SearchQnA/docker/xeon/compose.yaml b/SearchQnA/docker/xeon/compose.yaml index 0f81a93fe..37c4232c3 100644 --- a/SearchQnA/docker/xeon/compose.yaml +++ b/SearchQnA/docker/xeon/compose.yaml @@ -31,9 +31,6 @@ services: http_proxy: ${http_proxy} https_proxy: ${https_proxy} TEI_EMBEDDING_ENDPOINT: ${TEI_EMBEDDING_ENDPOINT} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-embedding-service" HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} restart: unless-stopped web-retriever: @@ -77,9 +74,6 @@ services: https_proxy: ${https_proxy} TEI_RERANKING_ENDPOINT: ${TEI_RERANKING_ENDPOINT} HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-reranking-service" restart: unless-stopped tgi-service: image: ghcr.io/huggingface/text-generation-inference:1.4 @@ -109,9 +103,6 @@ services: https_proxy: ${https_proxy} TGI_LLM_ENDPOINT: ${TGI_LLM_ENDPOINT} HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-llm-service" restart: unless-stopped searchqna-xeon-backend-server: image: opea/searchqna:latest diff --git a/Translation/docker/xeon/compose.yaml b/Translation/docker/xeon/compose.yaml index bfbdf0117..697c90a80 100644 --- a/Translation/docker/xeon/compose.yaml +++ b/Translation/docker/xeon/compose.yaml @@ -42,9 +42,6 @@ services: https_proxy: ${https_proxy} TGI_LLM_ENDPOINT: ${TGI_LLM_ENDPOINT} HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN} - LANGCHAIN_API_KEY: ${LANGCHAIN_API_KEY} - LANGCHAIN_TRACING_V2: ${LANGCHAIN_TRACING_V2} - LANGCHAIN_PROJECT: "opea-llm-service" restart: unless-stopped translation-xeon-backend-server: image: opea/translation:latest