Skip to content

Commit 92cb48c

Browse files
committed
ExLlamaV3: Fix regression in max_seq_len limit
1 parent 0d1a8ba commit 92cb48c

1 file changed

Lines changed: 1 addition & 1 deletion

File tree

backends/exllamav3/model.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -264,7 +264,7 @@ async def create(cls, model_directory: pathlib.Path, hf_model: HFModel, **kwargs
264264
)
265265
cache_size = cache_size_default
266266

267-
if max_seq_len < cache_size:
267+
if max_seq_len > cache_size:
268268
logger.warning(
269269
f"The given max_seq_len ({max_seq_len}) is larger than the cache size "
270270
f"and will be limited to {cache_size} tokens."

0 commit comments

Comments
 (0)