---
library_name: transformers
tags: []
---
# ***Mol-MoE***: Performant & Steerable Multi-Objective RLHF in Drug Design
*Diego Calanzone (1, 2), Pierluca D'Oro (2), Pierre-Luc Bacon (1, 2)*
*(1) Universite de Montreal, (2) Mila Quebec AI Institute*
## How to use this model
This LM is fine-tuned to generate molecules in the SMILES format wrt. desired properties.
For unconditioned SMILES generation, use the BOS token ``. For conditioned generation, you can target the following properties: `JNK3, DRD2, GSK3B, CYP2D6, CYP2C19`.
```
prompt:
```
An example of the generation pipeline:
```
from transformers import AutoTokenizer, AutoModelForCausalLM
import re
# Setup
device = "cuda"
tokenizer = AutoTokenizer.from_pretrained("ddidacus/Mol-MoE-6x1b")
model = AutoModelForCausalLM.from_pretrained("ddidacus/Mol-MoE-6x1b")
generation_kwargs = {
"max_new_tokens": 128,
"min_length": -1,
"top_k": 0.0,
"top_p": 0.9,
"do_sample": True,
"pad_token_id": tokenizer.eos_token_id,
"temperature": 1.0
}
# Inference
query = ""
toks = tokenizer([query], return_tensors="pt")["input_ids"].to(device)
output = model.generate(toks, **generation_kwargs)
output = tokenizer.batch_decode(output)
# Parsing
filter = r'(.*?)'
molecule = re.findall(filter, output[0], re.DOTALL)
```
### Model Description
This model is a fine-tuned version of LLaMa 3.2 1B through two stages:
1. Fine-tuning on ~3.5M molecules extracted from: ZINC 250K, MOSES, CHEMBL
2. RLHF-tuning using RLOO on 5 distinct reward functions from PyTDC [1]
The six LLaMa models trained in (1) and (2) are merged into mixtral blocks using MergeKit [2].
- **Developed by:** Diego Calanzone (diego.calanzone@mila.quebec)
- **Model type:** Mixtral Mixture of Experts
- **Finetuned from model [optional]:** LLaMA 3.2 1B
Read the paper for further details.
### Sources
[1] https://tdcommons.ai/single_pred_tasks/overview
[2] https://github.com/arcee-ai/mergekit