File tree Expand file tree Collapse file tree 1 file changed +3
-3
lines changed
backends/gaudi/server/text_generation_server/models Expand file tree Collapse file tree 1 file changed +3
-3
lines changed Original file line number Diff line number Diff line change @@ -232,7 +232,7 @@ def __init__(
232
232
self .prefilling = prefilling
233
233
234
234
@property
235
- def token_idx (self ):
235
+ def token_idx (self ): # noqa: F811
236
236
if self .prefilling :
237
237
# no right padding for prefill
238
238
token_idx_scalar = self .attention_mask .shape [- 1 ] - 1
@@ -1534,8 +1534,8 @@ def warmup(
1534
1534
1535
1535
except Exception :
1536
1536
raise RuntimeError (
1537
- f "Not enough memory to handle following prefill and decode warmup."
1538
- f "You need to decrease `--max-batch-prefill-tokens`"
1537
+ "Not enough memory to handle following prefill and decode warmup."
1538
+ "You need to decrease `--max-batch-prefill-tokens`"
1539
1539
)
1540
1540
1541
1541
mem_stats = get_hpu_memory_stats (self .device )
You can’t perform that action at this time.
0 commit comments