|
--- |
|
license: apache-2.0 |
|
--- |
|
# Skimformer |
|
|
|
A collaboration between [reciTAL](https://recital.ai/en/) & [MLIA](https://mlia.lip6.fr/) (ISIR, Sorbonne Université) |
|
|
|
## Model description |
|
Skimformer is a two-stage Transformer that replaces self-attention with Skim-Attention, a self-attention module that computes attention solely based on the 2D positions of tokens in the page. The model adopts a two-step approach: first, the skim-attention scores are computed once and only once using layout information alone; then, these attentions are used in every layer of a text-based Transformer encoder. For more details, please refer to our paper: |
|
|
|
[Skim-Attention: Learning to Focus via Document Layout](https://arxiv.org/abs/2109.01078) |
|
Laura Nguyen, Thomas Scialom, Jacopo Staiano, Benjamin Piwowarski, [EMNLP 2021](https://2021.emnlp.org/papers) |
|
|
|
## Citation |
|
|
|
``` latex |
|
@article{nguyen2021skimattention, |
|
title={Skim-Attention: Learning to Focus via Document Layout}, |
|
author={Laura Nguyen and Thomas Scialom and Jacopo Staiano and Benjamin Piwowarski}, |
|
journal={arXiv preprint arXiv:2109.01078} |
|
year={2021}, |
|
} |
|
``` |