METAGENE-1-BnB-4Bit

Model Information

METAGENE-1 is a 7-billion-parameter autoregressive transformer language model, which we refer to as a metagenomic foundation model, that was trained on a novel corpus of diverse metagenomic DNA and RNA sequences comprising over 1.5 trillion base pairs. This dataset is sourced from a large collection of human wastewater samples, processed and sequenced using deep metagenomic (next-generation) sequencing methods. Unlike genomic models that focus on individual genomes or curated sets of specific species, the aim of METAGENE-1 is to capture the full distribution of genomic information present across the human microbiome. After pretraining, this model is designed to aid in tasks in the areas of biosurveillance, pandemic monitoring, and pathogen detection.

This repository contains metagene-ai/METAGENE-1-BnB-4Bit quantized using bitsandbytes from BF16 down to NF4 with a block size of 64, and storage type torch.bfloat16.

Downloads last month
0
Safetensors
Model size
3.35B params
Tensor type
F32
·
FP16
·
U8
·
Inference API
Unable to determine this model's library. Check the docs .

Model tree for metagene-ai/METAGENE-1-BnB-4Bit

Quantized
(4)
this model

Collection including metagene-ai/METAGENE-1-BnB-4Bit