#用8bit adam优化器节省显存
# Initialize the optimizer
if cfg.solver.use_8bit_adam:
try:
import bitsandbytes as bnb
except ImportError:
raise ImportError(
"Please install bitsandbytes to use 8-bit Adam. You can do so by running `pip install bitsandbytes`"
)
optimizer_cls = bnb.optim.AdamW8bit
else:
optimizer_cls = torch.optim.AdamW
optimizer = optimizer_cls(
trainable_params,
lr=learning_rate,
betas=(cfg.solver.adam_beta1, cfg.solver.adam_beta2),
weight_decay=cfg.solver.adam_weight_decay,
eps=cfg.solver.adam_epsilon,
)
GPU显存不足可用8bit adam优化器节省显存
最新推荐文章于 2024-09-24 14:22:37 发布
本文介绍了如何通过引入8-bitAdam优化器(如AdamW8bit)来节省深度学习训练中的显存消耗,同时提到了安装bitsandbytes库的必要性。
2169

被折叠的 条评论
为什么被折叠?



