Skip to content

Commit

Permalink
refine readme for reorg (#782)
Browse files Browse the repository at this point in the history
Signed-off-by: Xinyao Wang <[email protected]>
Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
  • Loading branch information
XinyaoWa and pre-commit-ci[bot] authored Sep 11, 2024
1 parent d97882e commit d2bab99
Show file tree
Hide file tree
Showing 28 changed files with 60 additions and 60 deletions.
4 changes: 2 additions & 2 deletions AudioQnA/kubernetes/intel/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,14 +7,14 @@
## Deploy On Xeon
```
cd GenAIExamples/AudioQnA/kubernetes/manifests/xeon
cd GenAIExamples/AudioQnA/kubernetes/intel/cpu/xeon/manifests
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" audioqna.yaml
kubectl apply -f audioqna.yaml
```
## Deploy On Gaudi
```
cd GenAIExamples/AudioQnA/kubernetes/manifests/gaudi
cd GenAIExamples/AudioQnA/kubernetes/intel/hpu/gaudi/manifests
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" audioqna.yaml
kubectl apply -f audioqna.yaml
Expand Down
12 changes: 6 additions & 6 deletions ChatQnA/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -123,7 +123,7 @@ Currently we support two ways of deploying ChatQnA services with docker compose:
docker pull opea/chatqna-conversation-ui:latest
```

2. Using the docker images `built from source`: [Guide](docker/xeon/README.md)
2. Using the docker images `built from source`: [Guide](docker_compose/intel/cpu/xeon/README.md)

> Note: The **opea/chatqna-without-rerank:latest** docker image has not been published yet, users need to build this docker image from source.

Expand All @@ -139,7 +139,7 @@ By default, the embedding, reranking and LLM models are set to a default value a

Change the `xxx_MODEL_ID` in `docker/xxx/set_env.sh` for your needs.

For customers with proxy issues, the models from [ModelScope](https://www.modelscope.cn/models) are also supported in ChatQnA. Refer to [this readme](docker/xeon/README.md) for details.
For customers with proxy issues, the models from [ModelScope](https://www.modelscope.cn/models) are also supported in ChatQnA. Refer to [this readme](docker_compose/intel/cpu/xeon/README.md) for details.

### Setup Environment Variable

Expand Down Expand Up @@ -202,19 +202,19 @@ Refer to the [Xeon Guide](./docker_compose/intel/cpu/xeon/README.md) for more in
### Deploy ChatQnA on NVIDIA GPU

```bash
cd GenAIExamples/ChatQnA/docker/gpu/
cd GenAIExamples/ChatQnA/docker_compose/nvidia/gpu/
docker compose up -d
```

Refer to the [NVIDIA GPU Guide](./docker/gpu/README.md) for more instructions on building docker images from source.
Refer to the [NVIDIA GPU Guide](./docker_compose/nvidia/gpu/README.md) for more instructions on building docker images from source.

### Deploy ChatQnA into Kubernetes on Xeon & Gaudi with GMC

Refer to the [Kubernetes Guide](./kubernetes/intel/README_gmc.md) for instructions on deploying ChatQnA into Kubernetes on Xeon & Gaudi with GMC.

### Deploy ChatQnA into Kubernetes on Xeon & Gaudi without GMC

Refer to the [Kubernetes Guide](./kubernetes/kubernetes/intel/README.md) for instructions on deploying ChatQnA into Kubernetes on Xeon & Gaudi without GMC.
Refer to the [Kubernetes Guide](./kubernetes/intel/README.md) for instructions on deploying ChatQnA into Kubernetes on Xeon & Gaudi without GMC.

### Deploy ChatQnA into Kubernetes using Helm Chart

Expand All @@ -224,7 +224,7 @@ Refer to the [ChatQnA helm chart](https://github.com/opea-project/GenAIInfra/tre

### Deploy ChatQnA on AI PC

Refer to the [AI PC Guide](./docker/aipc/README.md) for instructions on deploying ChatQnA on AI PC.
Refer to the [AI PC Guide](./docker_compose/intel/cpu/aipc/README.md) for instructions on deploying ChatQnA on AI PC.

### Deploy ChatQnA on Red Hat OpenShift Container Platform (RHOCP)

Expand Down
2 changes: 1 addition & 1 deletion ChatQnA/docker_compose/intel/cpu/aipc/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -159,7 +159,7 @@ Note: Please replace with `host_ip` with you external IP address, do not use loc
> Before running the docker compose command, you need to be in the folder that has the docker compose yaml file
```bash
cd GenAIExamples/ChatQnA/docker/aipc/
cd GenAIExamples/ChatQnA/docker_compose/intel/cpu/aipc/
docker compose up -d

# let ollama service runs
Expand Down
2 changes: 1 addition & 1 deletion ChatQnA/docker_compose/intel/cpu/xeon/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -147,7 +147,7 @@ cd ..
Build frontend Docker image via below command:

```bash
cd GenAIExamples/ChatQnA/
cd GenAIExamples/ChatQnA/ui
docker build --no-cache -t opea/chatqna-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile .
cd ../../../..
```
Expand Down
4 changes: 2 additions & 2 deletions ChatQnA/docker_compose/intel/cpu/xeon/README_qdrant.md
Original file line number Diff line number Diff line change
Expand Up @@ -85,7 +85,7 @@ docker build --no-cache -t opea/retriever-qdrant:latest --build-arg https_proxy=
### 3. Build Rerank Image

```bash
docker build --no-cache -t opea/reranking-tei:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/reranks/tei/Dockerfile .
docker build --no-cache -t opea/reranking-tei:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/reranks/tei/Dockerfile .`
```

### 4. Build LLM Image
Expand Down Expand Up @@ -117,7 +117,7 @@ cd ../../..
Build frontend Docker image via below command:

```bash
cd GenAIExamples/ChatQnA/
cd GenAIExamples/ChatQnA/ui
docker build --no-cache -t opea/chatqna-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile .
cd ../../../..
```
Expand Down
4 changes: 2 additions & 2 deletions ChatQnA/docker_compose/intel/hpu/gaudi/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -128,7 +128,7 @@ cd ../..
Construct the frontend Docker image using the command below:

```bash
cd GenAIExamples/ChatQnA/
cd GenAIExamples/ChatQnA/ui
docker build --no-cache -t opea/chatqna-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile .
cd ../../../..
```
Expand All @@ -150,7 +150,7 @@ cd ../../../..
To fortify AI initiatives in production, Guardrails microservice can secure model inputs and outputs, building Trustworthy, Safe, and Secure LLM-based Applications.

```bash
cd GenAIExamples/ChatQnA/docker
cd GenAIComps
docker build -t opea/guardrails-tgi:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f comps/guardrails/llama_guard/langchain/Dockerfile .
cd ../../..
```
Expand Down
4 changes: 2 additions & 2 deletions ChatQnA/docker_compose/nvidia/gpu/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -59,7 +59,7 @@ cd ../../..
Construct the frontend Docker image using the command below:

```bash
cd GenAIExamples/ChatQnA/
cd GenAIExamples/ChatQnA/ui
docker build --no-cache -t opea/chatqna-ui:latest --build-arg https_proxy=$https_proxy --build-arg http_proxy=$http_proxy -f ./docker/Dockerfile .
cd ../../../..
```
Expand Down Expand Up @@ -132,7 +132,7 @@ Note: Please replace with `host_ip` with you external IP address, do **NOT** use
### Start all the services Docker Containers

```bash
cd GenAIExamples/ChatQnA/docker/gpu/
cd GenAIExamples/ChatQnA/docker_compose/nvidia/gpu/
docker compose up -d
```

Expand Down
4 changes: 2 additions & 2 deletions ChatQnA/kubernetes/intel/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@
## Deploy On Xeon

```
cd GenAIExamples/ChatQnA/kubernetes/manifests/xeon
cd GenAIExamples/ChatQnA/kubernetes/intel/cpu/xeon/manifests
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" chatqna.yaml
kubectl apply -f chatqna.yaml
Expand All @@ -20,7 +20,7 @@ kubectl apply -f chatqna.yaml
## Deploy On Gaudi

```
cd GenAIExamples/ChatQnA/kubernetes/manifests/gaudi
cd GenAIExamples/ChatQnA/kubernetes/intel/hpu/gaudi/manifests
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" chatqna.yaml
kubectl apply -f chatqna.yaml
Expand Down
2 changes: 1 addition & 1 deletion ChatQnA/tests/test_manifest_on_gaudi.sh
Original file line number Diff line number Diff line change
Expand Up @@ -166,7 +166,7 @@ case "$1" in
if [ $ret -ne 0 ]; then
exit $ret
fi
pushd ChatQnA/kubernetes/manifests/gaudi
pushd ChatQnA/kubernetes/intel/hpu/gaudi/manifests
set +e
install_and_validate_chatqna_guardrail
popd
Expand Down
2 changes: 1 addition & 1 deletion ChatQnA/tests/test_manifest_on_xeon.sh
Original file line number Diff line number Diff line change
Expand Up @@ -166,7 +166,7 @@ case "$1" in
if [ $ret -ne 0 ]; then
exit $ret
fi
pushd ChatQnA/kubernetes/manifests/xeon
pushd ChatQnA/kubernetes/intel/cpu/xeon/manifests
set +e
install_and_validate_chatqna_guardrail
popd
Expand Down
4 changes: 2 additions & 2 deletions CodeGen/kubernetes/intel/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,7 @@
## Deploy On Xeon

```
cd GenAIExamples/CodeGen/kubernetes/manifests/xeon
cd GenAIExamples/CodeGen/kubernetes/intel/cpu/xeon/manifests
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
export MODEL_ID="meta-llama/CodeLlama-7b-hf"
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" codegen.yaml
Expand All @@ -23,7 +23,7 @@ kubectl apply -f codegen.yaml
## Deploy On Gaudi

```
cd GenAIExamples/CodeGen/kubernetes/manifests/gaudi
cd GenAIExamples/CodeGen/kubernetes/intel/hpu/gaudi/manifests
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" codegen.yaml
kubectl apply -f codegen.yaml
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@ Before deploying the react-codegen.yaml file, ensure that you have the following
```
# You may set the HUGGINGFACEHUB_API_TOKEN via method:
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
cd GenAIExamples/CodeGen/kubernetes/manifests/xeon/ui/
cd GenAIExamples/CodeGen/kubernetes/intel/cpu/xeon/manifests/ui/
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" react-codegen.yaml
```
b. Set the proxies based on your network configuration
Expand Down
4 changes: 2 additions & 2 deletions CodeTrans/kubernetes/intel/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ Change the `MODEL_ID` in `codetrans.yaml` for your needs.
## Deploy On Xeon

```bash
cd GenAIExamples/CodeTrans/kubernetes/manifests/xeon
cd GenAIExamples/CodeTrans/kubernetes/intel/cpu/xeon/manifests
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" codetrans.yaml
kubectl apply -f codetrans.yaml
Expand All @@ -30,7 +30,7 @@ kubectl apply -f codetrans.yaml
## Deploy On Gaudi

```bash
cd GenAIExamples/CodeTrans/kubernetes/manifests/gaudi
cd GenAIExamples/CodeTrans/kubernetes/intel/hpu/gaudi/manifests
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" codetrans.yaml
kubectl apply -f codetrans.yaml
Expand Down
2 changes: 1 addition & 1 deletion DocSum/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ Currently we support two ways of deploying Document Summarization services with
docker pull opea/docsum:latest
```

2. Start services using the docker images `built from source`: [Guide](./docker)
2. Start services using the docker images `built from source`: [Guide](./docker_compose)

### Required Models

Expand Down
4 changes: 2 additions & 2 deletions DocSum/kubernetes/intel/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@
## Deploy On Xeon

```
cd GenAIExamples/DocSum/kubernetes/manifests/xeon
cd GenAIExamples/DocSum/kubernetes/intel/cpu/xeon/manifests
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" docsum.yaml
kubectl apply -f docsum.yaml
Expand All @@ -20,7 +20,7 @@ kubectl apply -f docsum.yaml
## Deploy On Gaudi

```
cd GenAIExamples/DocSum/kubernetes/manifests/gaudi
cd GenAIExamples/DocSum/kubernetes/intel/hpu/gaudi/manifests
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" docsum.yaml
kubectl apply -f docsum.yaml
Expand Down
2 changes: 1 addition & 1 deletion DocSum/kubernetes/intel/cpu/xeon/manifest/ui/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,7 @@ Before deploying the react-docsum.yaml file, ensure that you have the following
```
# You may set the HUGGINGFACEHUB_API_TOKEN via method:
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
cd GenAIExamples/DocSum/kubernetes/manifests/xeon/ui/
cd GenAIExamples/DocSum/kubernetes/intel/cpu/xeon/manifests/ui/
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" react-docsum.yaml
```
b. Set the proxies based on your network configuration
Expand Down
4 changes: 2 additions & 2 deletions FaqGen/kubernetes/intel/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@ If use gated models, you also need to provide [huggingface token](https://huggin
## Deploy On Xeon

```
cd GenAIExamples/FaqGen/kubernetes/manifests/xeon
cd GenAIExamples/FaqGen/kubernetes/intel/cpu/xeon/manifests
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" faqgen.yaml
kubectl apply -f faqgen.yaml
Expand All @@ -26,7 +26,7 @@ kubectl apply -f faqgen.yaml
## Deploy On Gaudi

```
cd GenAIExamples/FaqGen/kubernetes/manifests/gaudi
cd GenAIExamples/FaqGen/kubernetes/intel/hpu/gaudi/manifests
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" faqgen.yaml
kubectl apply -f faqgen.yaml
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,7 @@ Before deploying the react-faqgen.yaml file, ensure that you have the following
```
# You may set the HUGGINGFACEHUB_API_TOKEN via method:
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
cd GenAIExamples/FaqGen/kubernetes/manifests/xeon/ui/
cd GenAIExamples/FaqGen/kubernetes/intel/cpu/xeon/manifests/ui/
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" react-faqgen.yaml
```
b. Set the proxies based on your network configuration
Expand Down
2 changes: 1 addition & 1 deletion ProductivitySuite/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -20,4 +20,4 @@ Refer to the [Keycloak Configuration Guide](./docker_compose/intel/cpu/xeon/keyc

Refer to the [Xeon Guide](./docker_compose/intel/cpu/xeon/README.md) for more instructions on building docker images from source and running the application via docker compose.

Refer to the [Xeon Kubernetes Guide](./kubernetes/manifests/README.md) for more instruction on deploying the application via kubernetes.
Refer to the [Xeon Kubernetes Guide](./kubernetes/intel/README.md) for more instruction on deploying the application via kubernetes.
4 changes: 2 additions & 2 deletions ProductivitySuite/kubernetes/intel/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,7 @@ To begin with, ensure that you have following prerequisites in place:
```
# You may set the HUGGINGFACEHUB_API_TOKEN via method:
export HUGGINGFACEHUB_API_TOKEN="YourOwnToken"
cd GenAIExamples/ProductivitySuite/kubernetes/manifests/xeon/
cd GenAIExamples/ProductivitySuite/kubernetes/intel/cpu/xeon/manifests/
sed -i "s/insert-your-huggingface-token-here/${HUGGINGFACEHUB_API_TOKEN}/g" *.yaml
```

Expand All @@ -48,7 +48,7 @@ To begin with, ensure that you have following prerequisites in place:
## Deploying ProductivitySuite
You can use yaml files in xeon folder to deploy ProductivitySuite with reactUI.
```
cd GenAIExamples/ProductivitySuite/kubernetes/manifests/xeon/
cd GenAIExamples/ProductivitySuite/kubernetes/intel/cpu/xeon/manifests/
kubectl apply -f *.yaml
```

Expand Down
Loading

0 comments on commit d2bab99

Please sign in to comment.