File size: 1,192 Bytes
50d9cb4
 
 
 
 
 
 
 
 
 
 
 
 
c37e67e
f034eb7
 
 
 
 
 
 
 
 
c37e67e
 
f034eb7
 
c37e67e
f034eb7
 
c37e67e
f034eb7
c37e67e
f034eb7
 
c37e67e
f034eb7
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
---
license: mit
datasets:
- de-Rodrigo/merit
language:
- en
- es
base_model:
- HuggingFaceM4/idefics2-8b
pipeline_tag: image-text-to-text
---

# IDEFICS2 Merit

<a href="https://x.com/nearcyan/status/1706914605262684394">
  <div style="text-align: center;">
    <picture>
      <source media="(prefers-color-scheme: dark)" srcset="https://huggingface.co/de-Rodrigo/donut-merit/resolve/main/assets/dragon_huggingface.png">
      <source media="(prefers-color-scheme: light)" srcset="https://huggingface.co/de-Rodrigo/donut-merit/resolve/main/assets/dragon_huggingface.png">
      <img alt="DragonHuggingFace" src="https://huggingface.co/de-Rodrigo/donut-merit/resolve/main/assets/dragon_huggingface.png" style="width: 200px;">
    </picture>
  </div>
</a>


## Model Architecture
**This model is based on the Donut architecture and fine-tuned on the Merit dataset for form understanding tasks.**

- Backbone: [Idefics2](https://huggingface.co/HuggingFaceM4/idefics2-8b)
- Training Data: [Merit](https://huggingface.co/datasets/de-Rodrigo/merit)

## Example Usage

```python
from transformers import AutoModel

model = AutoModel.from_pretrained("de-Rodrigo/idefics2-merit")
```
**WIP** 🛠️