Update README.md
Browse files
README.md
CHANGED
|
@@ -77,11 +77,15 @@ European Championship => competitions
|
|
| 77 |
|
| 78 |
If you want to use flash attention or increase sequence length, please, check the following code:
|
| 79 |
```python
|
|
|
|
|
|
|
|
|
|
| 80 |
model = GLiNER.from_pretrained("knowledgator/gliner-llama-1B-v1.0",
|
| 81 |
_attn_implementation = 'flash_attention_2',
|
| 82 |
-
max_len = 2048).to('cuda:0')
|
| 83 |
```
|
| 84 |
|
|
|
|
| 85 |
### Benchmarks
|
| 86 |
Below you can see the table with benchmarking results on various named entity recognition datasets:
|
| 87 |
|
|
|
|
| 77 |
|
| 78 |
If you want to use flash attention or increase sequence length, please, check the following code:
|
| 79 |
```python
|
| 80 |
+
from gliner import GLiNER
|
| 81 |
+
import torch
|
| 82 |
+
|
| 83 |
model = GLiNER.from_pretrained("knowledgator/gliner-llama-1B-v1.0",
|
| 84 |
_attn_implementation = 'flash_attention_2',
|
| 85 |
+
max_len = 2048).to('cuda:0', dtype=torch.float16)
|
| 86 |
```
|
| 87 |
|
| 88 |
+
|
| 89 |
### Benchmarks
|
| 90 |
Below you can see the table with benchmarking results on various named entity recognition datasets:
|
| 91 |
|