Skip to content

Commit a81fb4e

Browse files
authored
changed max_seq_len 1024 to 2048
The models support a 2048 context window. This is not well advertised and people are getting confused. No sense having a smaller size here, as it just adds to the confusion.
1 parent 2845af9 commit a81fb4e

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

llama/model.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@ class ModelArgs:
2727
norm_eps: float = 1e-5
2828

2929
max_batch_size: int = 32
30-
max_seq_len: int = 1024
30+
max_seq_len: int = 2048
3131

3232

3333
class RMSNorm(torch.nn.Module):

0 commit comments

Comments
 (0)