Model type: Transformer-based masked language model
Training data: No additional pretraining, merges two existing models
Languages: 100+ languages
Architecture:
- Base architectures:
- XLM-RoBERTa base (multilingual)
- BERT base cased (multilingual)
Custom merging technique to combine weights from both base models into one unified model
- Downloads last month
- 7
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.