在本地加载Lora之后的llama2

import os
import transformers
from transformers import AutoModel,AutoTokenizer
import torch
os.environ['HF_HOME'] = '/root/autodl-tmp/cache/' #将数据存储在数据盘,默认路径会把模型存在系统盘/root/cache
os.environ['HF_ENDPOINT'] = 'https://hf-mirror.com'
from transformers import AutoTokenizer, AutoModelForCausalLM
model_path="/root/autodl-tmp/autodl-fs/LLaMA-Factory/models/llama2_lora_sft"

tokenizer = AutoTokenizer.from_pretrained(model_path)
model = AutoModelForCausalLM.from_pretrained(model_path,device_map="auto")#映射到cuda上,在cpu会被killed

prompt ="""interaction sequence:[\"Guess Who's Coming to Dinner\", 'Gladiator', \"Soldier's Story, A\", 'Key Largo', 'Verdict, The', 'Like Water for Chocolate (Como agua para chocolate)', 'Children of a Lesser God', 'Dances with Wolves', 'Day the Earth Stood Still, The', 'Stand and Deliver', 'Manhattan', 'Starman', 'Kramer Vs. Kramer', \"Miller's Crossing\", 'Simon Birch']\
    user feedback:Actor:I don't like Woody Allen, I prefer Ian Michael Smith."""
inputs = tokenizer(prompt,return_tensors="pt")#返回张量
inputs.input_ids = inputs.input_ids.to('cuda')
# Generate
generate_ids = model.generate(inputs.input_ids, max_length=1024)
res = tokenizer.batch_decode(generate_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False)[0]
print(res)

评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值