I was following a tutorial in Youtube, when wanted to load Llama3 8B:
model_name = "meta-llama/Meta-Llama-3-8B-Instruct"
tokenizer = AutoTokenizer.from_pretrained(model_name, use_auth_token=hugging_face_key)
model = AutoModelForCausalLM.from_pretrained(model_name, use_auth_token=hugging_face_key)
Got: “Your session has failed because all available RAM has been used”
Try: model = AutoModelForCausalLM.from_pretrained(model_name, use_auth_token=hugging_face_key, low_cpu_mem_usage=True)
But again the same error
New contributor
matteo is a new contributor to this site. Take care in asking for clarification, commenting, and answering.
Check out our Code of Conduct.