Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

any support plan for llava-llama3-70b/12b? any guide for Module optimization ? #689

Open
ztfmars opened this issue May 15, 2024 · 2 comments
Assignees

Comments

@ztfmars
Copy link

ztfmars commented May 15, 2024

  1. support plan

when it will release the version for supporting llava-llama3-70b?

meainwhile, will it will consider of supporting unofficial version like, using llm of llama3-120b?
huggingface link: mlabonne/Meta-Llama-3-120B-Instruct

  1. Module optimization

i think it's very important to enhance the visual encoder funciton, so how can i change the visual encoder instead of just clip vit, suach as other clip or mamba visual version may be better?
or can i just add some personal adapter(just simple MLP may not be a good idea) for visual encoder and finetuning on my on datasets? looking farward to the guidline and hook methods.
i really to want to have a try on personal creativaty!

look forward to your reply, thx!

@hhaAndroid
Copy link
Collaborator

@ztfmars Thank you for your feedback

  1. llava-llama3-70b We will support this in the near future, but for 120b we don't have that much computing power.
  2. Regarding the plan to experiment with replacing model components, we are already working on it. We hope to complete the refactoring by the end of this month.

@ztfmars
Copy link
Author

ztfmars commented May 17, 2024

@ztfmars Thank you for your feedback

  1. llava-llama3-70b We will support this in the near future, but for 120b we don't have that much computing power.
  2. Regarding the plan to experiment with replacing model components, we are already working on it. We hope to complete the refactoring by the end of this month.

hope you can give us some detail guides on personnal method later. or will the intern campus VLM part have an advanced lesson for llm engineer?
that will be great if someone post it in zhihu, bilibili or other. it must be an excellent influent on mmllm and everyone wants to take part in get better on own models.
i really can't wait any more!
thx

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants