From af07b7e5481dc8ec320b6041e70c61ffd83f2194 Mon Sep 17 00:00:00 2001 From: liujie44 Date: Fri, 15 Mar 2024 15:07:21 +0800 Subject: [PATCH 1/2] add skip_memory_metrics --- .../hybrid_parallelism/ce_gpt/benchmark_common/run_benchmark.sh | 1 + 1 file changed, 1 insertion(+) diff --git a/tests/test_tipc/dygraph/hybrid_parallelism/ce_gpt/benchmark_common/run_benchmark.sh b/tests/test_tipc/dygraph/hybrid_parallelism/ce_gpt/benchmark_common/run_benchmark.sh index 3fa75dadd6bc..fa05504c743b 100755 --- a/tests/test_tipc/dygraph/hybrid_parallelism/ce_gpt/benchmark_common/run_benchmark.sh +++ b/tests/test_tipc/dygraph/hybrid_parallelism/ce_gpt/benchmark_common/run_benchmark.sh @@ -145,6 +145,7 @@ function _train(){ --scale_loss 32768\ --per_device_train_batch_size ${micro_batch_size}\ --do_train \ + --skip_memory_metrics 0 \ --recompute ${use_recompute}" From 146f558c2da299fbd19e4d47b1f9310e120b3034 Mon Sep 17 00:00:00 2001 From: liujie44 Date: Fri, 15 Mar 2024 17:18:51 +0800 Subject: [PATCH 2/2] add skip_memory_metrics for autotuner --- tests/test_tipc/auto_tuner/autoconfig/llama7b_lora_params.json | 1 + .../auto_tuner/autoconfig/llama7b_pretrain_params.json | 3 ++- tests/test_tipc/auto_tuner/autoconfig/llama7b_sft_params.json | 1 + 3 files changed, 4 insertions(+), 1 deletion(-) diff --git a/tests/test_tipc/auto_tuner/autoconfig/llama7b_lora_params.json b/tests/test_tipc/auto_tuner/autoconfig/llama7b_lora_params.json index 585b13aa6133..9c5acd00ab6f 100644 --- a/tests/test_tipc/auto_tuner/autoconfig/llama7b_lora_params.json +++ b/tests/test_tipc/auto_tuner/autoconfig/llama7b_lora_params.json @@ -34,5 +34,6 @@ "sharding": "stage3", "recompute_granularity": "full_attn", "autotuner_benchmark": 1, + "skip_memory_metrics": 0, "benchmark": 1 } \ No newline at end of file diff --git a/tests/test_tipc/auto_tuner/autoconfig/llama7b_pretrain_params.json b/tests/test_tipc/auto_tuner/autoconfig/llama7b_pretrain_params.json index 71a02ed1a33f..b9c7cdc84b89 100644 --- a/tests/test_tipc/auto_tuner/autoconfig/llama7b_pretrain_params.json +++ b/tests/test_tipc/auto_tuner/autoconfig/llama7b_pretrain_params.json @@ -38,5 +38,6 @@ "recompute_granularity": "full", "save_total_limit": 2, "sharding_parallel_degree": 1, - "autotuner_benchmark": 1 + "autotuner_benchmark": 1, + "skip_memory_metrics": 0 } diff --git a/tests/test_tipc/auto_tuner/autoconfig/llama7b_sft_params.json b/tests/test_tipc/auto_tuner/autoconfig/llama7b_sft_params.json index c8f24041ec02..89d175664f95 100644 --- a/tests/test_tipc/auto_tuner/autoconfig/llama7b_sft_params.json +++ b/tests/test_tipc/auto_tuner/autoconfig/llama7b_sft_params.json @@ -33,5 +33,6 @@ "sharding": "stage2", "recompute_granularity": "full", "autotuner_benchmark": 1, + "skip_memory_metrics": 0, "benchmark": 1 } \ No newline at end of file