--- license: cc-by-nc-sa-4.0 language: - en - zh base_model: - Qwen/Qwen2.5-7B-Instruct tags: - machine tranlsation - O1-like model - Chat pipeline_tag: text-generation --- [![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory) # QuantFactory/DRT-o1-7B-GGUF This is quantized version of [Krystalan/DRT-o1-7B](https://huggingface.co/Krystalan/DRT-o1-7B) created using llama.cpp # Original Model Card # DRT-o1
🤗 DRT-o1-7B   |   🤗 DRT-o1-8B   |   🤗 DRT-o1-14B   |    📑 Paper
This repository contains the resources for our paper ["DRT-o1: Optimized Deep Reasoning Translation via Long Chain-of-Thought"](https://arxiv.org/abs/2412.17498) ### Updates: - *2024.12.31*: We updated [our paper](https://arxiv.org/abs/2412.17498) with more detals and analyses. Check it out! - *2024.12.31*: We released the testing set of our work, please refer to `data/test.jsonl` - *2024.12.30*: We released a new model checkpoint using Llama-3.1-8B-Instruct as the backbone, i.e., 🤗 DRT-o1-8B - *2024.12.24*: We released [our paper](https://arxiv.org/abs/2412.17498). Check it out! - *2024.12.23*: We released our model checkpoints. 🤗 DRT-o1-7B and 🤗 DRT-o1-14B. If you find this work is useful, please consider cite our paper: ``` @article{wang2024drt, title={DRT-o1: Optimized Deep Reasoning Translation via Long Chain-of-Thought}, author={Wang, Jiaan and Meng, Fandong and Liang, Yunlong and Zhou, Jie}, journal={arXiv preprint arXiv:2412.17498}, year={2024} } ``` ## Quick Links - [Introduction](#introduction) - [Models](#models) - [Model Access](#model-access) - [Model Performance](#model-performance) - [Model Prompts](#model-prompts) - [Quickstart](#quickstart) - [Translation Cases](#translation-cases) - [Data](#data) - [License](#license) ## Introduction In this work, we introduce DRT-o1, an attempt to bring the success of long thought reasoning to neural machine translation (MT). To this end, - 🌟 We mine English sentences with similes or metaphors from existing literature books, which are suitable for translation via long thought. - 🌟 We propose a designed multi-agent framework with three agents (i.e., a translator, an advisor and an evaluator) to synthesize the MT samples with long thought. There are 22,264 synthesized samples in total. - 🌟 We train DRT-o1-8B, DRT-o1-7B and DRT-o1-14B using Llama-3.1-8B-Instruct, Qwen2.5-7B-Instruct and Qwen2.5-14B-Instruct as backbones. > Our goal is not to achieve competitive performance with OpenAI’s O1 in neural machine translation (MT). Instead, we explore technical routes to bring the success of long thought to MT. To this end, we introduce DRT-o1, *a byproduct of our exploration*, and we hope it could facilitate the corresponding research in this direction. ## Models ### Model Access | | Backbone | Model Access | | :--: | :--: | :--: | | DRT-o1-7B | 🤗 Qwen2.5-7B-Instruct | 🤗 DRT-o1-7B | | DRT-o1-8B | 🤗 Llama-3.1-8B-Instruct | 🤗 DRT-o1-8B | | DRT-o1-14B | 🤗 Qwen2.5-14B-Instruct | 🤗 DRT-o1-14B | ### Model Performance | | GRF | CometKiwi | GRB | BLEU | CometScore | | :--: | :--: | :--: | :--: | :--: | :--: | | Llama-3.1-8B-Instruct | 79.25 | 70.14 | 73.30 | 18.55 | 74.58 | | Qwen2.5-7B-Instruct | 81.53 | 70.36 | 77.92 | 27.02 | 76.78 | | Qwen2.5-14B-Instruct | 84.74 | 72.01 | 80.85 | 30.23 | 78.84 | | Marco-o1-7B | 82.41 | 71.62 | 77.50 | 29.48 | 77.41 | | QwQ-32B-preview | 86.31 | 71.48 | 83.08 | 27.46 | 78.68 | | DRT-o1-8B | 84.49 | 70.85 | 80.80 | 32.67 | 78.81 | | DRT-o1-7B | 85.57 | 71.78 | 82.38 | 35.54 | 80.19 | | DRT-o1-14B | **87.19** | **72.11** | **83.20** | **36.46** | **80.64** | ### Model Prompts During model inference, please use the following prompts: - System prompt: `You are a philosopher skilled in deep thinking, accustomed to exploring complex problems with profound insight.` - User prompt: `Please translate the following text from English to Chinese:\n[An English text]` DRT-o1 models will first generate the thought and then provide the final translation, with the following format: ```