WARNING: THIS SITE IS A MIRROR OF GITHUB.COM / IT CANNOT LOGIN OR REGISTER ACCOUNTS / THE CONTENTS ARE PROVIDED AS-IS / THIS SITE ASSUMES NO RESPONSIBILITY FOR ANY DISPLAYED CONTENT OR LINKS / IF YOU FOUND SOMETHING MAY NOT GOOD FOR EVERYONE, CONTACT ADMIN AT ilovescratch@foxmail.com
Skip to content

Commit bb37125

Browse files
committed
modify
Signed-off-by: MrZ20 <[email protected]>
1 parent d1e8ee8 commit bb37125

9 files changed

+15
-15
lines changed

tests/e2e/multicard/test_aclgraph_capture_replay.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -134,7 +134,7 @@ def _run_worker_process(
134134
@pytest.mark.parametrize("model", MODELS)
135135
@pytest.mark.parametrize("max_tokens", [4, 36])
136136
@patch.dict(os.environ, {"ASCEND_RT_VISIBLE_DEVICES": "0,1"})
137-
def test_dp2_aclgraph_capture_replay_metrics(
137+
def test_aclgraph_capture_replay_metrics_dp2(
138138
model: str,
139139
max_tokens: int,
140140
monkeypatch: pytest.MonkeyPatch,

tests/e2e/multicard/test_data_parallel.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,7 @@
3838
@pytest.mark.parametrize("model", MODELS)
3939
@pytest.mark.parametrize("max_tokens", [32])
4040
@patch.dict(os.environ, {"ASCEND_RT_VISIBLE_DEVICES": "0,1"})
41-
def test_dp_inference_multi_model_configs(model, max_tokens):
41+
def test_qwen_inference_dp2(model, max_tokens):
4242
moe_models = ["Qwen/Qwen3-30B-A3B", "vllm-ascend/Qwen3-30B-A3B-W8A8"]
4343
quantization_models = ["vllm-ascend/Qwen3-30B-A3B-W8A8"]
4444
script = "examples/offline_data_parallel.py"

tests/e2e/multicard/test_data_parallel_tp2.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@
1515
@pytest.mark.parametrize("model", MODELS)
1616
@pytest.mark.parametrize("max_tokens", [32])
1717
@patch.dict(os.environ, {"ASCEND_RT_VISIBLE_DEVICES": "0,1,2,3"})
18-
def test_dp2_tp2_hybrid_inference(model, max_tokens):
18+
def test_qwen_inference_dp2_tp2(model, max_tokens):
1919
script = "examples/offline_data_parallel.py"
2020

2121
env = os.environ.copy()

tests/e2e/multicard/test_expert_parallel

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ from tests.e2e.model_utils import check_outputs_equal
55

66

77
@pytest.mark.parametrize("model_name", ["deepseek-ai/DeepSeek-V2-Lite-Chat"])
8-
def test_ep_correctness_against_tp(model_name):
8+
def test_deepseek_correctness_ep(model_name):
99
example_prompts = [
1010
"Hello, my name is",
1111
"The president of the United States is",

tests/e2e/multicard/test_external_launcher.py

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,7 @@
3737

3838
@pytest.mark.parametrize("model", MODELS)
3939
@patch.dict(os.environ, {"HCCL_BUFFSIZE": "500"})
40-
def test_external_launcher(model):
40+
def test_qwen_external_launcher(model):
4141
script = Path(
4242
__file__
4343
).parent.parent.parent.parent / "examples" / "offline_external_launcher.py"
@@ -78,7 +78,7 @@ def test_external_launcher(model):
7878

7979

8080
@pytest.mark.parametrize("model", MOE_MODELS)
81-
def test_external_launcher_moe_ep_tp2(model):
81+
def test_qwen_moe_external_launcher_ep(model):
8282
script = Path(
8383
__file__
8484
).parent.parent.parent.parent / "examples" / "offline_external_launcher.py"
@@ -109,7 +109,7 @@ def test_external_launcher_moe_ep_tp2(model):
109109

110110

111111
@patch.dict(os.environ, {"VLLM_ASCEND_ENABLE_NZ": "0"})
112-
def test_external_launcher_with_sleepmode_level1():
112+
def test_qwen_external_launcher_with_sleepmode():
113113
script = Path(
114114
__file__
115115
).parent.parent.parent.parent / "examples" / "offline_external_launcher.py"
@@ -154,7 +154,7 @@ def test_external_launcher_with_sleepmode_level1():
154154

155155

156156
@patch.dict(os.environ, {"VLLM_ASCEND_ENABLE_NZ": "0"})
157-
def test_external_launcher_with_sleepmode_level2():
157+
def test_qwen_external_launcher_with_sleepmode_level2():
158158
script = Path(
159159
__file__
160160
).parent.parent.parent.parent / "examples" / "offline_external_launcher.py"
@@ -210,7 +210,7 @@ def test_external_launcher_with_sleepmode_level2():
210210
"VLLM_ASCEND_ENABLE_MATMUL_ALLREDUCE": "1",
211211
"HCCL_BUFFSIZE": "500"
212212
})
213-
def test_external_launcher_with_matmul_allreduce(model):
213+
def test_qwen_external_launcher_with_matmul_allreduce(model):
214214
script = Path(
215215
__file__
216216
).parent.parent.parent.parent / "examples" / "offline_external_launcher.py"

tests/e2e/multicard/test_full_graph_mode.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,7 @@
2929
from tests.e2e.model_utils import check_outputs_equal
3030

3131

32-
def test_models_distributed_Qwen3_MOE_TP2_WITH_FULL_DECODE_ONLY():
32+
def test_qwen_moe_with_full_decode_only():
3333
if 'HCCL_OP_EXPANSION_MODE' in os.environ:
3434
del os.environ['HCCL_OP_EXPANSION_MODE']
3535
prompts = [
@@ -75,7 +75,7 @@ def test_models_distributed_Qwen3_MOE_TP2_WITH_FULL_DECODE_ONLY():
7575
)
7676

7777

78-
def test_models_distributed_Qwen3_MOE_TP2_WITH_FULL():
78+
def test_qwen_moe_with_full():
7979
if 'HCCL_OP_EXPANSION_MODE' in os.environ:
8080
del os.environ['HCCL_OP_EXPANSION_MODE']
8181
prompts = [

tests/e2e/multicard/test_fused_moe_allgather_ep.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -41,7 +41,7 @@
4141
"TASK_QUEUE_ENABLE": "1",
4242
"VLLM_ENABLE_FUSED_EXPERTS_ALLGATHER_EP": "1"
4343
})
44-
def test_generate_with_allgather():
44+
def test_deepseek_moe_fused_allgather_ep():
4545
example_prompts = ["Hello, my name is"]
4646
sampling_params = SamplingParams(max_tokens=100, temperature=0.0)
4747

@@ -62,7 +62,7 @@ def test_generate_with_allgather():
6262
"VLLM_WORKER_MULTIPROC_METHOD": "spawn",
6363
"TASK_QUEUE_ENABLE": "1"
6464
})
65-
def test_generate_with_alltoall():
65+
def test_deepseek_moe_fused_alltoall_ep():
6666
example_prompts = ["Hello, my name is"]
6767
sampling_params = SamplingParams(max_tokens=100, temperature=0.0)
6868

tests/e2e/multicard/test_offline_weight_load.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -31,7 +31,7 @@
3131

3232
@pytest.mark.parametrize("model", MODELS)
3333
@patch.dict(os.environ, {"VLLM_ASCEND_ENABLE_NZ": "0"})
34-
def test_offline_weight_load_and_sleepmode(model):
34+
def test_qwen_offline_weight_load_and_sleepmode(model):
3535
script = Path(
3636
__file__
3737
).parent.parent.parent.parent / "examples" / "offline_external_launcher.py"

tests/e2e/multicard/test_pipeline_parallel.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,7 @@
3737
@pytest.mark.parametrize("tp_size", TENSOR_PARALLELS)
3838
@pytest.mark.parametrize("pp_size", PIPELINE_PARALLELS)
3939
@pytest.mark.parametrize("distributed_executor_backend", DIST_EXECUTOR_BACKEND)
40-
def test_pp_tp_dist(model: str, tp_size: int, pp_size: int,
40+
def test_models_pp2(model: str, tp_size: int, pp_size: int,
4141
distributed_executor_backend: str) -> None:
4242
with VllmRunner(model,
4343
tensor_parallel_size=tp_size,

0 commit comments

Comments
 (0)