Ubuntu – How to load a fine-tuned peft/lora model based on llama with Huggingface transformers?
I've followed this tutorial (colab notebook) in order to finetune my model. Trying to load my locally saved model model = AutoModelForCausalLM.from_pretrained("finetuned_model") yields Killed. Trying to load model from hub: yields import torch from peft import PeftModel, PeftConfig from transformers…