OmniGenome-186M / README.md
yangheng's picture
update
51f9ad7
|
raw
history blame
891 Bytes
---
license: mit
language:
- rna
- dna
tags:
- Genomic-Language-Modeling
- OmniGenome Foundation Model
---
# Multi-species Foundation Model for Universal RNA and DNA Downstream Tasks
# Notes
We are keep updating the checkpoints, the current checkpoint is trained for 0.85 epoch.
## Training Examples
Refer to GitHub [https://github.com/yangheng95/OmniGenome](https://github.com/yangheng95/OmniGenome)
## Usage
This model is available for replacing genomic foundation models such as CDSBERT, Nucleotide Transformers, DNABERT2, etc.
```
from transformers import AutoModel
model = AutoModel.from_pretrained("yangheng/OmniGenome-52M", trust_remote_code=True)
```
## Subtasks
- Secondary structure prediction
- Genome Sequence Classification
- Genome Sequence Regression
- Single Nucleotide Repair
- Genome Masked Language Modeling
- etc.
Part of the codes are adapted from ESM2.