Unknown

Dataset Information

0

Understanding the natural language of DNA using encoder-decoder foundation models with byte-level precision.


ABSTRACT:

Summary

This article presents the Ensemble Nucleotide Byte-level Encoder-Decoder (ENBED) foundation model, analyzing DNA sequences at byte-level precision with an encoder-decoder Transformer architecture. ENBED uses a subquadratic implementation of attention to develop an efficient model capable of sequence-to-sequence transformations, generalizing previous genomic models with encoder-only or decoder-only architectures. We use Masked Language Modeling to pretrain the foundation model using reference genome sequences and apply it in the following downstream tasks: (i) identification of enhancers, promotors, and splice sites, (ii) recognition of sequences containing base call mismatches and insertion/deletion errors, an advantage over tokenization schemes involving multiple base pairs, which lose the ability to analyze with byte-level precision, (iii) identification of biological function annotations of genomic sequences, and (iv) generating mutations of the Influenza virus using the encoder-decoder architecture and validating them against real-world observations. In each of these tasks, we demonstrate significant improvement as compared to the existing state-of-the-art results.

Availability and implementation

The source code used to develop and fine-tune the foundation model has been released on Github (https://github.itap.purdue.edu/Clan-labs/ENBED).

SUBMITTER: Malusare A 

PROVIDER: S-EPMC11341122 | biostudies-literature | 2024

REPOSITORIES: biostudies-literature

altmetric image

Publications

Understanding the natural language of DNA using encoder-decoder foundation models with byte-level precision.

Malusare Aditya A   Kothandaraman Harish H   Tamboli Dipesh D   Lanman Nadia A NA   Aggarwal Vaneet V  

Bioinformatics advances 20240812 1


<h4>Summary</h4>This article presents the Ensemble Nucleotide Byte-level Encoder-Decoder (ENBED) foundation model, analyzing DNA sequences at byte-level precision with an encoder-decoder Transformer architecture. ENBED uses a subquadratic implementation of attention to develop an efficient model capable of sequence-to-sequence transformations, generalizing previous genomic models with encoder-only or decoder-only architectures. We use Masked Language Modeling to pretrain the foundation model usi  ...[more]

Similar Datasets

| S-EPMC9044365 | biostudies-literature
| S-EPMC10100163 | biostudies-literature
| S-EPMC4451011 | biostudies-literature
| S-EPMC8292438 | biostudies-literature
| S-EPMC8629217 | biostudies-literature
| S-EPMC9541604 | biostudies-literature
| S-EPMC8756182 | biostudies-literature
| S-EPMC9360199 | biostudies-literature
| S-EPMC11539752 | biostudies-literature
| S-EPMC9411784 | biostudies-literature