File size: 3,633 Bytes
e1e77b6 635e0cd 79b4a6a 7b85b83 79b4a6a 635e0cd 7b85b83 635e0cd 79b4a6a 635e0cd 79b4a6a 635e0cd 79b4a6a 635e0cd 79b4a6a 7b85b83 635e0cd 7b85b83 79b4a6a 7b85b83 635e0cd 79b4a6a 7bf5d0d d4f832d 79b4a6a 635e0cd 79b4a6a 7b85b83 79b4a6a 635e0cd 79b4a6a 635e0cd d4f832d 79b4a6a d4f832d 635e0cd 7bf5d0d 635e0cd 7bf5d0d 635e0cd |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 |
---
license: llama3.1
library_name: peft
base_model: meta-llama/Llama-3.1-8B
tags:
- llama
- lora
- qlora
- fine-tuned
- robotics
- task-planning
- construction
- dart-llm
language:
- en
pipeline_tag: text-generation
---
# Llama 3.1 8B - DART LLM Robot Task Planning (QLoRA Fine-tuned)
This model is a QLoRA fine-tuned version of **meta-llama/Llama-3.1-8B** specialized for **robot task planning** in construction environments.
The model converts natural language commands into structured task sequences for construction robots including excavators and dump trucks.
## Model Details
- **Base Model**: meta-llama/Llama-3.1-8B
- **Fine-tuning Method**: QLoRA (4-bit quantization + LoRA)
- **LoRA Rank**: 16-32 (optimized per model size)
- **LoRA Alpha**: 16-32
- **Target Modules**: q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj
- **Dataset**: YongdongWang/dart_llm_tasks_pretty
- **Training Domain**: Construction robotics
## Usage
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
from peft import PeftModel
import torch
# Load tokenizer and base model
tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-3.1-8B")
base_model = AutoModelForCausalLM.from_pretrained(
"meta-llama/Llama-3.1-8B",
torch_dtype=torch.float16,
device_map="auto"
)
# Load LoRA adapter
model = PeftModel.from_pretrained(base_model, "YongdongWang/llama-3.1-8b-lora-qlora-dart-llm")
# Generate robot task sequence
instruction = "Deploy Excavator 1 to Soil Area 1 for excavation"
prompt = f"### Instruction:\n{instruction}\n\n### Response:\n"
inputs = tokenizer(prompt, return_tensors="pt")
with torch.no_grad():
outputs = model.generate(
**inputs,
max_new_tokens=512,
do_sample=False,
pad_token_id=tokenizer.eos_token_id
)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)
```
## Training Details
- **Training Data**: DART LLM Tasks - Robot command and task planning dataset
- **Domain**: Construction robotics (excavators, dump trucks, soil/rock areas)
- **Training Epochs**: 6-12 (optimized per model size)
- **Batch Size**: 1 (with gradient accumulation)
- **Learning Rate**: 1e-4 to 3e-4 (scaled by model size)
- **Optimizer**: paged_adamw_8bit or adamw_torch
## Capabilities
- **Multi-robot coordination**: Handle multiple excavators and dump trucks
- **Task dependencies**: Generate proper task sequences with dependencies
- **Spatial reasoning**: Understand soil areas, rock areas, puddles, and navigation
- **Action planning**: Convert commands to structured JSON task definitions
## Example Output
The model generates structured task sequences in JSON format for robot execution:
```json
{
"tasks": [
{
"instruction_function": {
"dependencies": [],
"name": "target_area_for_specific_robots",
"object_keywords": ["soil_area_1"],
"robot_ids": ["robot_excavator_01"],
"robot_type": null
},
"task": "target_area_for_specific_robots_1"
}
]
}
```
## Limitations
This model is specifically trained for construction robotics scenarios and may not generalize to other domains without additional fine-tuning.
## Citation
```bibtex
@misc{llama_3.1_8b_lora_qlora_dart_llm,
title={Llama 3.1 8B Fine-tuned with QLoRA for DART LLM Tasks},
author={YongdongWang},
year={2024},
publisher={Hugging Face},
url={https://huggingface.co/YongdongWang/llama-3.1-8b-lora-qlora-dart-llm}
}
```
## Model Card Authors
YongdongWang
## Model Card Contact
For questions or issues, please open an issue in the repository or contact the model author.
|