Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Finetuning general LLM models from hugging face #521

Closed
bkhanal-11 opened this issue May 17, 2024 · 2 comments
Closed

Finetuning general LLM models from hugging face #521

bkhanal-11 opened this issue May 17, 2024 · 2 comments

Comments

@bkhanal-11
Copy link

Thank you for your continuous support to the LLM open-source community.

I was wondering if we use AutoModel, AutoConfig and AutoTokenizer instead of LlamaForCausalLM, LlamaConfig and LlamaTokenizer for general LLM fine-tuning, will the fine-tuning pipeline/recipe fail?

@mreso
Copy link
Contributor

mreso commented May 20, 2024

Hi @bkhanal-11 for the single gpu use case this might work, but for the advances use cases with FSDP you'll need to different wrapping policy and there might be other things to consider and adapt.

@bkhanal-11
Copy link
Author

That makes sense, thanks.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants