forked from microsoft/DeepSpeed
-
Notifications
You must be signed in to change notification settings - Fork 4
82 lines (65 loc) · 2.56 KB
/
autotp.yml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
name: autotp
on:
workflow_dispatch:
merge_group:
branches: [ master, gma/test_cpu_branch ]
concurrency:
group: ${{ github.workflow }}-${{ github.ref }}
cancel-in-progress: true
jobs:
tests:
runs-on: [self-hosted]
steps:
- uses: actions/checkout@v3
- id: setup-venv
uses: ./.github/workflows/setup-venv
- name: Install essential packages
run: |
sudo apt-get install -y git cmake numactl iputils-ping
- name: Check python version
run: |
python --version
- name: Install oneCCL Bindings for PyTorch
run: |
python -m pip install intel_extension_for_pytorch
wget https://intel-extension-for-pytorch.s3.amazonaws.com/ipex_stable/cpu/oneccl_bind_pt-2.1.0%2Bcpu-cp310-cp310-linux_x86_64.whl
pip install oneccl_bind_pt-2.1.0+cpu-cp310-cp310-linux_x86_64.whl
#python -m pip install oneccl_bind_pt -f https://developer.intel.com/ipex-whl-stable-cpu
- name: Install oneCCL
run: |
git clone https://github.com/oneapi-src/oneCCL
cd oneCCL
mkdir build
cd build
cmake ..
make -j install
- name: Install transformers
run: |
git clone https://github.com/huggingface/transformers
cd transformers
git rev-parse --short HEAD
pip install .
- name: Install deepspeed
run: |
# check why the host does not have AVX2 support
pip install .[dev,1bit,autotuning,inf]
ds_report
- name: Python environment
run: |
pip list
- name: Download DeepSpeedExamples
run: |
git clone https://github.com/foin6/DeepSpeedExamples --branch dev
- name: Sanity check minimal
run: |
source oneCCL/build/_install/env/setvars.sh
python -c "import torch;import intel_extension_for_pytorch as ipex;print(ipex._C._has_xpu())"
cd DeepSpeedExamples/scripts
deepspeed --num_gpus 2 --bind_cores_to_rank check-license.py
- name: AutoTP tests
run: |
source oneCCL/build/_install/env/setvars.sh
cd DeepSpeedExamples/inference/huggingface/text-generation
# test auto tp with meta tensor
TRANSFORMERS_CACHE=/tmp/transformers_cache/ deepspeed --num_gpus 2 inference-test.py --model bigscience/bloom-3b --dtype float32 --use_meta_tensor
TRANSFORMERS_CACHE=/tmp/transformers_cache/ deepspeed --num_gpus 2 inference-test.py --model facebook/opt-1.3b --dtype float32 --use_meta_tensor