Daily Tasks Fine Tuned LLaMA 3.2 1B β Weekly & Daily Task Planner
Model Description
This is a fine-tuned LLaMA 3.2 1B model designed to generate structured weekly and daily plans. It can produce:
- Workout routines
- Study schedules
- Meal plans
- Other daily task setups
Fine-tuning was done using PEFT LoRA with float16 precision for efficient training on GPU.
Intended Use
This model is intended for personal productivity, fitness planning, and educational scheduling. It is not meant for medical, legal, or critical decision-making.
Usage
from transformers import pipeline, AutoTokenizer, AutoModelForCausalLM
model_name = "your-username/daily_tasks_fine_tuned_llama3_2_1b"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)
generator = pipeline(
"text-generation",
model=model,
tokenizer=tokenizer,
device=0 # Use -1 for CPU
)
prompt = "Plan a 7-day workout routine for cardiovascular health."
output = generator(prompt, max_new_tokens=600)
print(output[0]['generated_text'])