from transformers import LLaMAForConditionalGeneration, LLaMATokenizer # Load pre-trained model and tokenizer model = LLaMAForConditionalGeneration.from_pretrained('decapoda-research/llama-3.2-hf') tokenizer = LLaMATokenizer.from_pretrained('decapoda-res