open-genome / README.md
exnx's picture
update readme
84369c0
metadata
license: apache-2.0
tags:
  - stripedhyena
  - long context
  - deep signal processing
  - hybrid
  - biology
  - genomics
task_categories:
  - text-generation
language:
  - en
pretty_name: open-genome
configs:
  - config_name: stage1
    data_files:
      - split: train
        path:
          - stage1/gtdb/gtdb_train_shard_*
          - stage1/imgpr/imgpr_train.parquet
      - split: validation
        path:
          - stage1/gtdb/gtdb_valid_small.parquet
          - stage1/imgpr/imgpr_valid_small.parquet
      - split: test
        path:
          - stage1/gtdb/gtdb_test.parquet
          - stage1/imgpr/imgpr_test.parquet
  - config_name: stage2
    data_files:
      - split: train
        path: stage2/train_stage2.parquet
      - split: validation
        path: stage2/valid_stage2.parquet
      - split: test
        path: stage2/test_stage2.parquet
  - config_name: sample
    data_files:
      - split: validation
        path: stage2/valid_stage2.parquet

Dataset organization

The OpenGenome dataset is organized in 2 stages, where stage 1 has context length 8k and stage 2 has context length 131k. Each stage has their own datasplits.

- stage1
  - train
  - validation
  - test

- stage2
  - train
  - validation
  - test

Instructions to download

You can load a dataset using HF's API, with an example below.

from datasets import load_dataset

stage1_data = load_dataset("LongSafari/open-genome", 'stage1')

# access just the train data
stage_1_train_data = stage1_data['train']

Note: stage 1 training dataset is sharded into separate files due to it's large size.

We also provide a small dataset sample to test out the pipeline if you prefer.

sample_data = load_dataset("LongSafari/open-genome", 'sample')['validation']