Replies: 1 comment
-
Hello, have you solved this problem? I have the same problem |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I'd like to train llama3.1 model with variable sequence length as mentioned in Meta's paper. That is, to start the training with one sequence length and after a number of consumed tokens switch to a larger sequence length and continue training.
Several approaches in my opinion:
Any thoughts?
Beta Was this translation helpful? Give feedback.
All reactions