8G 显存玩转书生大模型 Demo
关卡任务
本关任务主要包括:
-
InternLM2-Chat-1.8B 模型的部署(基础任务)
-
InternLM-XComposer2-VL-1.8B 模型的部署(进阶任务)
-
InternVL2-2B 模型的部署(进阶任务)
任务详情参见 8G 显存玩转书生大模型 Demo 任务。
创建开发机
选择12.2 切记!
环境配置
我们已经在 /root/share/pre_envs
中配置好了预置环境 icamp3_demo
可以通过如下指令进行激活:
conda activate /root/share/pre_envs/icamp3_demo
这里可以激活内置环境 也可以自己创建一个环境,我这边因为前面做了camp2的实验,想尝试用一下camp2时的环境 省点时间
conda env list#可以看自己开发机的环境
这里我选择demo环境
conda activate demo
Cli Demo 部署 InternLM2-Chat-1.8B 模型
首先,我们创建一个目录,用于存放我们的代码。并创建一个 cli_demo.py
。
mkdir -p /root/demo touch /root/demo/cli_demo.py
然后,我们将下面的代码复制到 cli_demo.py
中。
import torch from transformers import AutoTokenizer, AutoModelForCausalLM model_name_or_path = "/root/share/new_models/Shanghai_AI_Laboratory/internlm2-chat-1_8b" tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, trust_remote_code=True, device_map='cuda:0') model = AutoModelForCausalLM.from_pretrained(model_name_or_path, trust_remote_code=True, torch_dtype=torch.bfloat16, device_map='cuda:0') model = model.eval() system_prompt = """You are an AI assistant whose name is InternLM (书生·浦语). - InternLM (书生·浦语) is a conversational language model that is developed by Shanghai AI Laboratory (上海人工智能实验室). It is designed to be helpful, honest, and harmless. - InternLM (书生·浦语) can understand and communicate fluently in the language chosen by the user such as English and 中文. """ messages = [(system_prompt, '')] print("=============Welcome to InternLM chatbot, type 'exit' to exit.=============") while True: input_text = input("\nUser >>> ") input_text = input_text.replace(' ', '') if input_text == "exit": break length = 0 for response, _ in model.stream_chat(tokenizer, input_text, messages): if response is not None: print(response[length:], flush=True, end="") length = len(response)
接下来,我们便可以通过 python /root/demo/cli_demo.py
来启动我们的 Demo。
效果如下图所示:
输入exit 可以退出