We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
在说明里能不能增加一下显卡的要求啊? 比如说,哪种数据量级的数据微调时,m3e-small base large 对显卡显存的要求是什么? 4080 16G、3090 24G这些卡 单卡能跑吗? 穷人手里没有48G 80G这样的卡。 非常感谢大佬们的答复。
The text was updated successfully, but these errors were encountered:
16G 这种级别的卡就都够用,需要注意的是,batch_size 不要设置的太大
Sorry, something went wrong.
您好,我的单卡是32G,但是最大只能跑batch size32;设置8卡来跑,也跑不通batch size 128,还有什么地方需要配置么?
srun -p src-12xv100-32g --workspace-id src -f pt -r N1lS.Ib.I20.8 -N 8 -d AllReduce bash finetune.sh
uniem 的显存瓶颈主要在激活上,并且依赖于 In Batch 的负采样,所以 DDP 或者 ZeRO 的方式也没有办法提升 Batch Size...
wangyuxinwhy
No branches or pull requests
🚀 The feature
在说明里能不能增加一下显卡的要求啊?
比如说,哪种数据量级的数据微调时,m3e-small base large 对显卡显存的要求是什么?
4080 16G、3090 24G这些卡 单卡能跑吗?
穷人手里没有48G 80G这样的卡。
非常感谢大佬们的答复。
The text was updated successfully, but these errors were encountered: