|
--- |
|
library_name: transformers |
|
license: apache-2.0 |
|
pipeline_tag: translation |
|
language: |
|
- bg |
|
- ca |
|
- cs |
|
- cy |
|
- da |
|
- de |
|
- el |
|
- en |
|
- es |
|
- et |
|
- eu |
|
- fi |
|
- fr |
|
- ga |
|
- gl |
|
- hr |
|
- hu |
|
- it |
|
- lt |
|
- lv |
|
- mt |
|
- nl |
|
- nb |
|
- 'no' |
|
- nn |
|
- oc |
|
- pl |
|
- pt |
|
- ro |
|
- ru |
|
- sl |
|
- sk |
|
- sr |
|
- sv |
|
- uk |
|
- ast |
|
- an |
|
base_model: |
|
- BSC-LT/salamandraTA-7b-instruct |
|
--- |
|
 |
|
|
|
# SalamandraTA-7B-instruct-GGUF Model Card |
|
|
|
This model is the GGUF-quantized version of [SalamandraTA-7b-instruct](https://huggingface.co/BSC-LT/salamandraTA-7b-instruct). |
|
|
|
The model weights are quantized from FP16 to Q4_K_M quantization Q8_0 (8-bit quantization), (4-bit weights with K-means clustering quantization) and Q3_K_M (3-but weights with K-means clustering quantization) using the [Llama.cpp](https://github.com/ggml-org/llama.cpp) framework. |
|
Inferencing with this model can be done using [VLLM](https://docs.vllm.ai/en/stable/models/engine_args.html). |
|
|
|
SalamandraTA-7b-instruct is a translation LLM that has been instruction-tuned from SalamandraTA-7b-base. |
|
The base model results from continually pre-training [Salamandra-7b](https://huggingface.co/BSC-LT/salamandra-7b) on parallel data and has not been published, but is reserved for internal use. |
|
SalamandraTA-7b-instruct is proficent in 37 european languages and supports translation-related tasks, namely: sentence-level-translation, paragraph-level-translation, document-level-translation, automatic post-editing, grammar checking, machine translation evaluation, alternative translations, named-entity-recognition and context-aware translation. |
|
|
|
> [!WARNING] |
|
> **DISCLAIMER:** This version of Salamandra is tailored exclusively for translation tasks. It lacks chat capabilities and has not been trained with any chat instructions. |
|
|
|
|
|
--- |
|
|
|
The entire Salamandra family is released under a permissive [Apache 2.0 license]((https://www.apache.org/licenses/LICENSE-2.0)). |
|
|
|
## How to Use |
|
|
|
The following example code works under ``Python 3.10.4``, ``vllm==0.7.3``, ``torch==2.5.1`` and ``torchvision==0.20.1``, though it should run on |
|
any current version of the libraries. This is an example of translation using the model: |
|
|
|
``` |
|
from huggingface_hub import snapshot_download |
|
from vllm import LLM, SamplingParams |
|
|
|
model_dir = snapshot_download(repo_id="BSC-LT/salamandraTA-7B-instruct-GGUF", revision="main") |
|
model_name = "salamandrata_7b_inst_q4.gguf" |
|
|
|
llm = LLM(model=model_dir + '/' + model_name, tokenizer=model_dir) |
|
|
|
source = "Spanish" |
|
target = "English" |
|
sentence = "Ayer se fue, tomó sus cosas y se puso a navegar. Una camisa, un pantalón vaquero y una canción, dónde irá, dónde irá. Se despidió, y decidió batirse en duelo con el mar. Y recorrer el mundo en su velero. Y navegar, nai-na-na, navegar." |
|
|
|
prompt = f"Translate the following text from {source} into {target}.\\n{source}: {sentence} \\n{target}:" |
|
messages = [{'role': 'user', 'content': prompt}] |
|
|
|
outputs = llm.chat(messages, |
|
sampling_params=SamplingParams( |
|
temperature=0.1, |
|
stop_token_ids=[5], |
|
max_tokens=200) |
|
)[0].outputs |
|
|
|
print(outputs[0].text) |
|
|
|
|
|
``` |
|
|
|
## Additional information |
|
|
|
### Author |
|
The Language Technologies Unit from Barcelona Supercomputing Center. |
|
|
|
### Contact |
|
For further information, please send an email to <[email protected]>. |
|
|
|
### Copyright |
|
Copyright(c) 2025 by Language Technologies Unit, Barcelona Supercomputing Center. |
|
|
|
### Funding |
|
This work has been promoted and financed by the Government of Catalonia through the [Aina Project](https://projecteaina.cat/). |
|
|
|
This work is funded by the _Ministerio para la Transformación Digital y de la Función Pública_ - Funded by EU – NextGenerationEU |
|
within the framework of [ILENIA Project](https://proyectoilenia.es/) with reference 2022/TL22/00215337. |
|
|
|
### Acknowledgements |
|
|
|
The success of this project has been made possible thanks to the invaluable contributions of our partners in the [ILENIA Project](https://proyectoilenia.es/): |
|
[HiTZ](http://hitz.ehu.eus/es), and [CiTIUS](https://citius.gal/es/). |
|
Their efforts have been instrumental in advancing our work, and we sincerely appreciate their help and support. |
|
|
|
|
|
|
|
### Disclaimer |
|
|
|
|
|
### Disclaimer |
|
Be aware that the model may contain biases or other unintended distortions. |
|
When third parties deploy systems or provide services based on this model, or use the model themselves, |
|
they bear the responsibility for mitigating any associated risks and ensuring compliance with applicable regulations, |
|
including those governing the use of Artificial Intelligence. |
|
|
|
The Barcelona Supercomputing Center, as the owner and creator of the model, shall not be held liable for any outcomes resulting from third-party use. |
|
|
|
### License |
|
[Apache License, Version 2.0](https://www.apache.org/licenses/LICENSE-2.0) |
|
|
|
|