Add limitation due to RoBERTa
Browse files
README.md
CHANGED
@@ -74,6 +74,20 @@ The following hyperparameters were used during training:
|
|
74 |
| 0.0115 | 2.45 | 4000 | 0.0172 | 0.9821 | 0.9871 | 0.9845 | 0.9962 |
|
75 |
|
76 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
77 |
### Framework versions
|
78 |
|
79 |
- Transformers 4.30.2
|
|
|
74 |
| 0.0115 | 2.45 | 4000 | 0.0172 | 0.9821 | 0.9871 | 0.9845 | 0.9962 |
|
75 |
|
76 |
|
77 |
+
### Limitations
|
78 |
+
|
79 |
+
**Warning**: This model works best when punctuation is separated from the prior words, so
|
80 |
+
```python
|
81 |
+
# ✅
|
82 |
+
model.predict("He plays J. Robert Oppenheimer , an American theoretical physicist .")
|
83 |
+
# ❌
|
84 |
+
model.predict("He plays J. Robert Oppenheimer, an American theoretical physicist.")
|
85 |
+
|
86 |
+
# You can also supply a list of words directly: ✅
|
87 |
+
model.predict(["He", "plays", "J.", "Robert", "Oppenheimer", ",", "an", "American", "theoretical", "physicist", "."])
|
88 |
+
```
|
89 |
+
The same may be beneficial for some languages, such as splitting `"l'ocean Atlantique"` into `"l' ocean Atlantique"`.
|
90 |
+
|
91 |
### Framework versions
|
92 |
|
93 |
- Transformers 4.30.2
|