Update README.md

#6
by sergiopaniego HF Staff - opened
Files changed (1) hide show
  1. README.md +3 -1
README.md CHANGED
@@ -1,5 +1,7 @@
1
  ---
2
  license: apache-2.0
 
 
3
  ---
4
 
5
  # vllm-flash-attn3
@@ -89,4 +91,4 @@ This will automatically resolve and download the appropriate code for your archi
89
 
90
  - [Tri Dao](https://huggingface.co/tridao) and team for Flash Attention and [Flash Attention 3](https://tridao.me/blog/2024/flash3/).
91
  - The [vLLM team](https://huggingface.co/vllm-project) for their implementation and their contribution of attention sinks.
92
- - The [transformers team](https://huggingface.co/transformers-community) for packaging, testing, building and making it available for use with the [kernels library](https://github.com/huggingface/kernels).
 
1
  ---
2
  license: apache-2.0
3
+ tags:
4
+ - kernel
5
  ---
6
 
7
  # vllm-flash-attn3
 
91
 
92
  - [Tri Dao](https://huggingface.co/tridao) and team for Flash Attention and [Flash Attention 3](https://tridao.me/blog/2024/flash3/).
93
  - The [vLLM team](https://huggingface.co/vllm-project) for their implementation and their contribution of attention sinks.
94
+ - The [transformers team](https://huggingface.co/transformers-community) for packaging, testing, building and making it available for use with the [kernels library](https://github.com/huggingface/kernels).