Skip to content

Commit 90ccbff

Browse files
committed
Fixing test with overlap
Signed-off-by: Patrice Castonguay <55748270+pcastonguay@users.noreply.github.com>
1 parent 45a3689 commit 90ccbff

File tree

2 files changed

+29
-23
lines changed

2 files changed

+29
-23
lines changed

tests/unittest/llmapi/test_llm.py

Lines changed: 7 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -1808,7 +1808,6 @@ def validate_stats(results,
18081808
assert results
18091809
assert len(results) == max_tokens if pytorch_backend else max_tokens + 1
18101810
for iter, result in enumerate(results):
1811-
print(result)
18121811
ifbStats = result["inflightBatchingStats"]
18131812
expected_num_scheduled = 1 if (iter < max_tokens) else 0
18141813
assert ifbStats["numScheduledRequests"] == expected_num_scheduled
@@ -1906,13 +1905,11 @@ def llm_get_stats_test_harness(tp_size: int = 1,
19061905

19071906

19081907
@pytest.mark.parametrize("return_context_logits", [True, False])
1909-
@pytest.mark.parametrize("pytorch_backend", [True, False])
19101908
@pytest.mark.parametrize("enable_iter_req_stats", [True, False])
1911-
def test_llm_get_stats(return_context_logits, pytorch_backend,
1912-
enable_iter_req_stats):
1909+
def test_llm_get_stats(return_context_logits, enable_iter_req_stats):
19131910
llm_get_stats_test_harness(tp_size=1,
19141911
return_context_logits=return_context_logits,
1915-
pytorch_backend=pytorch_backend,
1912+
pytorch_backend=False,
19161913
enable_iter_req_stats=enable_iter_req_stats)
19171914

19181915

@@ -1977,8 +1974,9 @@ async def task1():
19771974
results.append(stats)
19781975

19791976
assert results
1980-
validate_stats(results, pytorch_backend, max_tokens,
1981-
enable_iter_req_stats)
1977+
if not use_overlap:
1978+
validate_stats(results, pytorch_backend, max_tokens,
1979+
enable_iter_req_stats)
19821980

19831981
async def main():
19841982
for i in range(2): # test recurrent usage
@@ -1988,14 +1986,12 @@ async def main():
19881986

19891987

19901988
@pytest.mark.parametrize("return_context_logits", [True, False])
1991-
@pytest.mark.parametrize("pytorch_backend", [True, False])
19921989
@pytest.mark.parametrize("enable_iter_req_stats", [True, False])
1993-
def test_llm_get_stats_async(return_context_logits, pytorch_backend,
1994-
enable_iter_req_stats):
1990+
def test_llm_get_stats_async(return_context_logits, enable_iter_req_stats):
19951991
llm_get_stats_async_test_harness(
19961992
tp_size=1,
19971993
return_context_logits=return_context_logits,
1998-
pytorch_backend=pytorch_backend,
1994+
pytorch_backend=False,
19991995
enable_iter_req_stats=enable_iter_req_stats)
20001996

20011997

tests/unittest/llmapi/test_llm_pytorch.py

Lines changed: 22 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -29,27 +29,37 @@ def test_tinyllama_logits_processor():
2929
tinyllama_logits_processor_test_harness(backend="pytorch")
3030

3131

32-
@pytest.mark.parametrize("return_context_logits, use_overlap", [
33-
(False, False),
34-
(False, True),
35-
])
36-
def test_llm_get_stats(return_context_logits, use_overlap):
32+
@pytest.mark.parametrize(
33+
"return_context_logits, use_overlap, enable_iter_req_stats", [
34+
(False, False, False),
35+
(False, False, True),
36+
(False, True, False),
37+
(False, True, True),
38+
])
39+
def test_llm_get_stats(return_context_logits, use_overlap,
40+
enable_iter_req_stats):
3741
llm_get_stats_test_harness(tp_size=1,
3842
return_context_logits=return_context_logits,
3943
pytorch_backend=True,
40-
use_overlap=use_overlap)
44+
use_overlap=use_overlap,
45+
enable_iter_req_stats=enable_iter_req_stats)
4146

4247

43-
@pytest.mark.parametrize("return_context_logits, use_overlap", [
44-
(False, False),
45-
(False, True),
46-
])
47-
def test_llm_get_stats_async(return_context_logits, use_overlap):
48+
@pytest.mark.parametrize(
49+
"return_context_logits, use_overlap, enable_iter_req_stats", [
50+
(False, False, False),
51+
(False, False, True),
52+
(False, True, False),
53+
(False, True, True),
54+
])
55+
def test_llm_get_stats_async(return_context_logits, use_overlap,
56+
enable_iter_req_stats):
4857
llm_get_stats_async_test_harness(
4958
tp_size=1,
5059
return_context_logits=return_context_logits,
5160
pytorch_backend=True,
52-
use_overlap=use_overlap)
61+
use_overlap=use_overlap,
62+
enable_iter_req_stats=enable_iter_req_stats)
5363

5464

5565
@force_ampere

0 commit comments

Comments
 (0)