You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
As part of Kubeflow Training V2 work, we should design and implement custom Trainer to fine-tune LLMs that we are planning to support via TrainingRuntimes in Kubeflow upstream.
We should discuss whether we should use native PyTorch APIs or HuggingFace Transformers in the LLM Trainer implementation.
The Trainer should allow users to configure LoRA, QLoRA, FSDP, and other important configurations.
@andreyvelich: GitHub didn't allow me to assign the following users: saileshd1402.
Note that only kubeflow members with read permissions, repo collaborators and people who have commented on this issue/PR can be assigned. Additionally, issues/PRs can only have 10 assignees at the same time.
For more information please see the contributor guide
We are experimenting with some PyTorch-native and Transformers APIs to design this Trainer.
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.
As part of Kubeflow Training V2 work, we should design and implement custom Trainer to fine-tune LLMs that we are planning to support via TrainingRuntimes in Kubeflow upstream.
We should discuss whether we should use native PyTorch APIs or HuggingFace Transformers in the LLM Trainer implementation.
The Trainer should allow users to configure LoRA, QLoRA, FSDP, and other important configurations.
Useful resources:
Part of: #2170
Design Doc
Initial design doc from @Electronic-Waste where we can brainstorm ideas: https://docs.google.com/document/d/1a4xWGVWZo43QKv8tIomoK_XHzBMC_byXBnDb0104htQ/edit?tab=t.0
cc @saileshd1402 @deepanker13 @kubeflow/wg-training-leads
Love this feature?
Give it a 👍 We prioritize the features with most 👍
The text was updated successfully, but these errors were encountered: