nglaura commited on
Commit
15a959f
·
1 Parent(s): 52274d3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -3,14 +3,14 @@ license: apache-2.0
3
  ---
4
  # Skimformer
5
 
 
 
6
  ## Model description
7
  Skimformer is a two-stage Transformer that replaces self-attention with Skim-Attention, a self-attention module that computes attention solely based on the 2D positions of tokens in the page. The model adopts a two-step approach: first, the skim-attention scores are computed once and only once using layout information alone; then, these attentions are used in every layer of a text-based Transformer encoder. For more details, please refer to our paper:
8
 
9
  [Skim-Attention: Learning to Focus via Document Layout](https://arxiv.org/abs/2109.01078)
10
  Laura Nguyen, Thomas Scialom, Jacopo Staiano, Benjamin Piwowarski, [EMNLP 2021](https://2021.emnlp.org/papers)
11
 
12
- A collaboration between [reciTAL](https://recital.ai/en/) & [MLIA](https://mlia.lip6.fr/) (ISIR, Sorbonne Université)
13
-
14
  ## Citation
15
 
16
  ``` latex
 
3
  ---
4
  # Skimformer
5
 
6
+ A collaboration between [reciTAL](https://recital.ai/en/) & [MLIA](https://mlia.lip6.fr/) (ISIR, Sorbonne Université)
7
+
8
  ## Model description
9
  Skimformer is a two-stage Transformer that replaces self-attention with Skim-Attention, a self-attention module that computes attention solely based on the 2D positions of tokens in the page. The model adopts a two-step approach: first, the skim-attention scores are computed once and only once using layout information alone; then, these attentions are used in every layer of a text-based Transformer encoder. For more details, please refer to our paper:
10
 
11
  [Skim-Attention: Learning to Focus via Document Layout](https://arxiv.org/abs/2109.01078)
12
  Laura Nguyen, Thomas Scialom, Jacopo Staiano, Benjamin Piwowarski, [EMNLP 2021](https://2021.emnlp.org/papers)
13
 
 
 
14
  ## Citation
15
 
16
  ``` latex