Skip to content

Commit

Permalink
Change the llama max_batch_size larger than default eval batchsize
Browse files Browse the repository at this point in the history
  • Loading branch information
leslie-fang-intel committed Jun 7, 2024
1 parent 2d8999b commit e242597
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion torchbenchmark/models/llama/model.py
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@ class ModelArgs:
multiple_of: int = 256 # make SwiGLU hidden layer size multiple of large power of 2
norm_eps: float = 1e-5

max_batch_size: int = 32 # From the paper they use a batch size of 4M for training
max_batch_size: int = 64 # From the paper they use a batch size of 4M for training
max_seq_len: int = 1024

device: Optional[str] = None
Expand Down

0 comments on commit e242597

Please sign in to comment.