Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

请教一下Chinese-Alpaca-Plus-13B继续做peft微调需要多少显存? #31

Open
bash99 opened this issue May 15, 2023 · 4 comments
Labels
question Further information is requested wontfix This will not be worked on

Comments

@bash99
Copy link

bash99 commented May 15, 2023

有没有什么量化方法在对微调效果影响不大的情况下,降低显存需求

如果是4个没有NVLink的V100(32G显存),是否实际上只能当成4倍速的32G显存来微调?

@bash99 bash99 added the question Further information is requested label May 15, 2023
@shibing624
Copy link
Owner

是。使用torch.distributed.launch调用的多卡,本质是模型会加载到每个gpu卡里面,训练数据切分为4份,所以是小于等于4倍速,数据通信还要花点时间。

@bash99
Copy link
Author

bash99 commented May 16, 2023

是。使用torch.distributed.launch调用的多卡,本质是模型会加载到每个gpu卡里面,训练数据切分为4份,所以是小于等于4倍速,数据通信还要花点时间。

那么13B用32G显存能微调吗?

@shibing624
Copy link
Owner

能。

Copy link

stale bot commented Dec 27, 2023

This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions.(由于长期不活动,机器人自动关闭此问题,如果需要欢迎提问)

@stale stale bot added the wontfix This will not be worked on label Dec 27, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested wontfix This will not be worked on
Projects
None yet
Development

No branches or pull requests

2 participants