OmniGenome-52M / README.md
yangheng's picture
update
b0f0a2f
|
raw
history blame
892 Bytes
metadata
license: mit
language:
  - rna
  - dna
tags:
  - Genomic-Language-Modeling
  - OmniGenome Foundation Model

Multi-species Foundation Model for Universal RNA and DNA Downstream Tasks

Notes

We are keep updating the checkpoints, the current checkpoint is trained for 0.85 epoch.

Training Examples

Refer to GitHub https://github.com/yangheng95/OmniGenome

Usage

This model is available for replacing genomic foundation models such as CDSBERT, Nucleotide Transformers, DNABERT2, etc.

from transformers import AutoModel
model = AutoModel.from_pretrained("yangheng/OmniGenome-52M", trust_remote_code=True)

Subtasks

  • Secondary structure prediction
  • Genome Sequence Classification
  • Genome Sequence Regression
  • Single Nucleotide Repair
  • Genome Masked Language Modeling
  • etc.

Part of the codes are adapted from ESM2.