| base_model: | |
| - Qwen/Qwen2.5-Coder-7B-Instruct | |
| datasets: | |
| - SWE-bench/SWE-smith | |
| language: | |
| - en | |
| library_name: transformers | |
| license: apache-2.0 | |
| pipeline_tag: text-generation | |
| tags: | |
| - agent | |
| - software engineering | |
| <div align="center"> | |
| <img src="https://swe-agent.com/latest/assets/swe-agent.svg" alt="Logo" width="200"> | |
| <h1 align="center">SWE-agent LM</h1> | |
| </div> | |
| <p align="center"> | |
| <a href="https://github.com/SWE-bench/SWE-smith">Code</a> | |
| • | |
| <a href="https://arxiv.org/abs/2504.21798">Paper</a> | |
| • | |
| <a href="https://swesmith.com/">Site</a> | |
| </p> | |
| SWE-agent-LM-7B is a Language Model for Software Engineering trained using the [SWE-smith](https://github.com/SWE-bench/SWE-smith) toolkit. | |
| We introduce this model as part of our work: [SWE-smith: Scaling Data for Software Engineering Agents](https://swesmith.com). | |
| SWE-agent-LM-7B is 100% open source. | |
| Training this model was simple - we fine-tuned Qwen 2.5 Coder Instruct on 5k trajectories generated by SWE-agent + Claude 3.7 Sonnet. | |
| The dataset can be found [here](https://huggingface.co/datasets/SWE-bench/SWE-smith-trajs-250429). | |
| SWE-agent-LM-7B is compatible with [SWE-agent](https://github.com/SWE-agent/SWE-agent). | |
| Running this model locally only takes a few steps! | |
| Check [here]() for more instructions on how to do so. | |
| If you found this work exciting and want to push SWE-agents further, please feel free to connect with us (the [SWE-bench team](https://swe-bench.github.io/)) more! | |