You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Thank you for your continuous support to the LLM open-source community.
I was wondering if we use AutoModel, AutoConfig and AutoTokenizer instead of LlamaForCausalLM, LlamaConfig and LlamaTokenizer for general LLM fine-tuning, will the fine-tuning pipeline/recipe fail?
The text was updated successfully, but these errors were encountered:
Hi @bkhanal-11 for the single gpu use case this might work, but for the advances use cases with FSDP you'll need to different wrapping policy and there might be other things to consider and adapt.
Thank you for your continuous support to the LLM open-source community.
I was wondering if we use
AutoModel
,AutoConfig
andAutoTokenizer
instead ofLlamaForCausalLM
,LlamaConfig
andLlamaTokenizer
for general LLM fine-tuning, will the fine-tuning pipeline/recipe fail?The text was updated successfully, but these errors were encountered: