12345678910111213141516171819202122 |
- from peft import PeftModel
- from transformers import LlamaForCausalLM
- def load_model(model_name, quantization):
- model = LlamaForCausalLM.from_pretrained(
- model_name,
- return_dict=True,
- load_in_8bit=quantization,
- device_map="auto",
- low_cpu_mem_usage=True,
- )
- return model
- def load_peft_model(model, peft_model):
- peft_model = PeftModel.from_pretrained(model, peft_model)
- return peft_model
|