| from unsloth import FastLanguageModel | |
| model, tokenizer = FastLanguageModel.from_pretrained( | |
| model_name = "aidando73/llama-3.3-70b-instruct-code-agent-fine-tune-v1", | |
| max_seq_length = 2048, | |
| dtype = "float16", | |
| load_in_4bit = True, | |
| ) | |
| FastLanguageModel.for_inference(model) | |