Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

qwen 14b 不微调的情况下,问相同的问题,模型输出也不太一致,是为什么?温度已经设置成0了 #1249

Closed
Wangqi12138 opened this issue May 14, 2024 · 2 comments

Comments

@Wangqi12138
Copy link

          > > 我也遇到了,好像确实都不行,同求解

好像不微调也会,没搞清楚怎么引起的

想问下这个问题解决了吗?

Originally posted by @Elissa0723 in #1025 (comment)

@Wangqi12138 Wangqi12138 changed the title > > 我也遇到了,好像确实都不行,同求解 qwen 14b 不微调的情况下,问相同的问题,模型输出也不太一致,是为什么?温度已经设置成0了 May 14, 2024
@jklj077
Copy link
Contributor

jklj077 commented May 20, 2024

Hi!

It appears that the original author was using a quantized model. Unfortunately, most efficient kernels for GPTQ or AWQ use non-deterministic algorithms, such that the results may be slight different even when do_sample is set to False. In addition, transformers does not accept temperature=0 in recent versions and it will ask you to use do_sample=False.

However, it is unclear that whether your issue is the same with the one you have referenced. Please describe with more details, such as which model were you using and which framework were you using.

@jklj077
Copy link
Contributor

jklj077 commented May 22, 2024

As Qwen1.0 is no longer actively maintained, we kindly ask to you migrate to Qwen1.5 and direct your related question there. Thanks for you cooperation.

@jklj077 jklj077 closed this as completed May 22, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants