from transformers import AutoModelForCausalLM, AutoTokenizer model_name = "mlabonne/llama-2-7b-guanaco" try: model = AutoModelForCausalLM.from_pretrained(model_name, use_auth_token=False, device='cpu', memory_limit=1024*16) except Exception as e: print(f"Error loading the model: {e}")