微调指令order

这个链接

CUDA_VISIBLE_DEVICES=0 python src/train_bash.py
–stage sft
–model_name_or_path /home/rkwork/work_place/project/ChatGLM2-6B/models/chatglm2-6b-int4
–do_train
–dataset /home/rkwork/work_place/project/ChatGLM2-6B/ChatGLM-Efficient-Tuning/data/self_cognition.json
–finetuning_type lora
–output_dir path_to_sft_checkpoint
–per_device_train_batch_size 4
–gradient_accumulation_steps 4
–lr_scheduler_type cosine
–logging_steps 10
–save_steps 1000
–learning_rate 5e-5
–num_train_epochs 3.0
–plot_loss
–fp16

CUDA_VISIBLE_DEVICES=0 python src/train_sft.py \
--do_train \
--use_v2 \
--dataset self_cognition \
--finetuning_type lora \
--lora_rank 32 \
--output_dir cognition \
--overwrite_cache \
--per_device_train_batch_size 2 \
--gradient_accumulation_steps 2 \
--lr_scheduler_type cosine \
--logging_steps 10 \
--save_steps 1000 \
--warmup_steps 0 \
--learning_rate 1e-3 \
--num_train_epochs 10.0 \
--fp16

python src/cli_demo.py
–model_name_or_path /home/rkwork/work_place/project/ChatGLM2-6B/models/chatglm2-6b
–finetuning_type lora
–checkpoint_dir /home/rkwork/work_place/project/ChatGLM2-6B/ChatGLM-Efficient-Tuning/cognition/checkpoint-9900

CUDA_VISIBLE_DEVICES=0 python src/train_bash.py \
--stage sft \
--model_name_or_path /home/rkwork/work_place/project/ChatGLM2-6B/models/chatglm2-6b \
--do_train \
--dataset self_cognition \
--finetuning_type lora \
--output_dir path_to_sft_checkpoint \
--per_device_train_batch_size 4 \
--gradient_accumulation_steps 4 \
--lr_scheduler_type cosine \
--logging_steps 10 \
--save_steps 1000 \
--learning_rate 5e-5 \
--num_train_epochs 3.0 \
--plot_loss \
--fp16

训练 10000step fail!
CUDA_VISIBLE_DEVICES=0 python src/train_bash.py
–stage sft
–model_name_or_path /home/rkwork/work_place/project/ChatGLM2-6B/models/chatglm2-6b
–do_train
–dataset self_cognition
–finetuning_type lora
–output_dir cognition
–overwrite_cache
–per_device_train_batch_size 2
–gradient_accumulation_steps 2
–lr_scheduler_type cosine
–logging_steps 10
–save_steps 100
–warmup_steps 0
–learning_rate 1e-3
–num_train_epochs 500.0
–fp16

测试

python src/cli_demo.py
–model_name_or_path /home/rkwork/work_place/project/ChatGLM2-6B/models/chatglm2-6b
–finetuning_type lora
–checkpoint_dir /home/rkwork/work_place/project/ChatGLM2-6B/ChatGLM-Efficient-Tuning/cognition/checkpoint-9900

训练step 200成功!
CUDA_VISIBLE_DEVICES=0 python src/train_bash.py
–stage sft
–do_train
–model_name_or_path /home/rkwork/work_place/project/ChatGLM2-6B/models/chatglm2-6b
–dataset self_cognition
–finetuning_type lora
–output_dir cognition
–overwrite_cache
–per_device_train_batch_size 2
–gradient_accumulation_steps 2
–lr_scheduler_type cosine
–logging_steps 10
–save_steps 1000
–warmup_steps 0
–learning_rate 1e-3
–num_train_epochs 10.0
–fp16

 测试

python src/cli_demo.py
–model_name_or_path /home/rkwork/work_place/project/ChatGLM2-6B/models/chatglm2-6b
–finetuning_type lora
–checkpoint_dir /home/rkwork/work_place/project/ChatGLM2-6B/ChatGLM-Efficient-Tuning/cognition

python src/web_demo.py
–model_name_or_path /home/rkwork/work_place/project/ChatGLM2-6B/models/chatglm2-6b
–finetuning_type lora
–checkpoint_dir /home/rkwork/work_place/project/ChatGLM2-6B/ChatGLM-Efficient-Tuning/cognition

导出
python src/export_model.py \
--model_name_or_path /home/rkwork/work_place/project/ChatGLM2-6B/models/chatglm2-6b \
--finetuning_type lora \
--checkpoint_dir /home/rkwork/work_place/project/ChatGLM2-6B/ChatGLM-Efficient-Tuning/cognition\
--output_dir muziAI

aifeng_start
训练step 200成功!
CUDA_VISIBLE_DEVICES=0 python src/train_bash.py
–stage sft
–do_train
–model_name_or_path /home/rkwork/work_place/project/ChatGLM2-6B/models/chatglm2-6b
–dataset self_cognition_cp
–finetuning_type lora
–output_dir /home/rkwork/work_place/project/ChatGLM2-6B/ChatGLM-Efficient-Tuning/my_data/output_checkpoint
–overwrite_cache
–per_device_train_batch_size 2
–gradient_accumulation_steps 2
–lr_scheduler_type cosine
–logging_steps 10
–save_steps 1000
–warmup_steps 0
–learning_rate 1e-3
–num_train_epochs 10.0
–fp16

python src/cli_demo.py
–model_name_or_path /home/rkwork/work_place/project/ChatGLM2-6B/models/chatglm2-6b
–finetuning_type lora
–checkpoint_dir /home/rkwork/work_place/project/ChatGLM2-6B/ChatGLM-Efficient-Tuning/my_data/output_checkpoint

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 打赏
    打赏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包

打赏作者

aifeng-pro

你的鼓励将是我创作的最大动力

¥1 ¥2 ¥4 ¥6 ¥10 ¥20
扫码支付:¥1
获取中
扫码支付

您的余额不足,请更换扫码支付或充值

打赏作者

实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值