Tom Aarsen commited on
Commit
b94cf64
·
1 Parent(s): 5c66f04

Notify users that flash_attn is recommended

Browse files
Files changed (1) hide show
  1. README.md +8 -1
README.md CHANGED
@@ -35,7 +35,14 @@ The `gte-modernbert` models demonstrates competitive performance in several text
35
 
36
  ## Usage
37
 
38
- Use with `Transformers`
 
 
 
 
 
 
 
39
 
40
  ```python
41
  # Requires transformers>=4.48.0
 
35
 
36
  ## Usage
37
 
38
+ > [!TIP]
39
+ > For `transformers` and `sentence-transformers`, if your GPU supports it, the efficient Flash Attention 2 will be used automatically if you have `flash_attn` installed. It is not mandatory.
40
+ >
41
+ > ```bash
42
+ > pip install flash_attn
43
+ > ```
44
+
45
+ Use with `transformers`
46
 
47
  ```python
48
  # Requires transformers>=4.48.0