Papers
Topics
Authors
Recent
Search
2000 character limit reached

HyenaDNA: Efficient Genomic Foundation Model

Updated 27 February 2026
  • HyenaDNA is a genomic foundation model that uses implicit convolution and gating to achieve long-range sequence modeling at single-nucleotide resolution.
  • It employs fast Fourier transform-based convolutions to attain O(NlogN) runtime and favorable scaling for processing entire genomes and RNA sequences.
  • Adaptable via plug-in adapters, HyenaDNA enhances downstream tasks like rare disease gene discovery and regulatory element classification with reduced computational cost.

HyenaDNA is a genomic foundation model designed for long-range sequence modeling at single-nucleotide resolution, leveraging implicit-convolution architectures to enable order-of-magnitude scaling beyond attention-based transformers. It provides a universal, efficient, and highly parameter-efficient backbone for DNA and, via adapters, RNA modeling. HyenaDNA and related approaches underpin advances in downstream prediction, rare disease gene discovery, and general retrieval-augmented inference across genomic modalities.

1. Architectural Principles of HyenaDNA

HyenaDNA replaces the quadratic-cost self-attention mechanism in standard transformers with the Hyena "implicit convolution + gating" operator. Each Hyena block computes

H(x)=Dx2ThDx1H(x) = D_{x_2}\,T_h\,D_{x_1}

where Dx1,Dx2RL×LD_{x_1}, D_{x_2} \in \mathbb{R}^{L \times L} are diagonal matrices from learned projections of the input xRLx \in \mathbb{R}^L, and ThT_h is a Toeplitz matrix parameterizing a global 1D convolution. The convolution filter hkh_k is generated via a small MLP γθ(k)\gamma_\theta(k) (as a neural field) rather than being directly learned, which decouples parameter count from window size. Each block includes pointwise nonlinearity (GELU), layer normalization, and a feed-forward network, with residual connections throughout.

A defining feature is the ability to process single-nucleotide tokens, eschewing fixed kk-mer tokenization to preserve maximal nucleotide resolution—crucial for tasks involving SNPs or rare mutations. Architectural efficiency stems from the use of fast Fourier transform-based convolution (O(NlogN)O(N\log N) time/space per layer), making HyenaDNA orders of magnitude more efficient in both parameter count and runtime for long genomic inputs (Nguyen et al., 2023, Du et al., 6 Aug 2025).

2. Computational Complexity and Scaling

Traditional transformer models exhibit O(N2)O(N^2) time and memory complexity, limiting input lengths (typically $512$–Dx1,Dx2RL×LD_{x_1}, D_{x_2} \in \mathbb{R}^{L \times L}0 bases) to a small fraction of the human genome. In contrast, each HyenaDNA layer, via implicit convolution and diagonal gating, achieves Dx1,Dx2RL×LD_{x_1}, D_{x_2} \in \mathbb{R}^{L \times L}1 runtime and Dx1,Dx2RL×LD_{x_1}, D_{x_2} \in \mathbb{R}^{L \times L}2 memory, where Dx1,Dx2RL×LD_{x_1}, D_{x_2} \in \mathbb{R}^{L \times L}3 is sequence length and Dx1,Dx2RL×LD_{x_1}, D_{x_2} \in \mathbb{R}^{L \times L}4 embedding dimension. Empirical benchmarks demonstrate up to 160Dx1,Dx2RL×LD_{x_1}, D_{x_2} \in \mathbb{R}^{L \times L}5 speedup relative to FlashAttention-Transformers at Dx1,Dx2RL×LD_{x_1}, D_{x_2} \in \mathbb{R}^{L \times L}6 tokens, with practical context windows up to Dx1,Dx2RL×LD_{x_1}, D_{x_2} \in \mathbb{R}^{L \times L}7 nucleotides and efficient scaling to tens of layers (Dx1,Dx2RL×LD_{x_1}, D_{x_2} \in \mathbb{R}^{L \times L}8–Dx1,Dx2RL×LD_{x_1}, D_{x_2} \in \mathbb{R}^{L \times L}9), allowing true whole-gene or ultra-long-range context (Nguyen et al., 2023, Datta et al., 6 Aug 2025).

A comparison of key operational metrics is summarized below:

Model Max Context Params (M) Time Complexity
HyenaDNA xRLx \in \mathbb{R}^L0 4–6.5 xRLx \in \mathbb{R}^L1
DNABERT/Enformer xRLx \in \mathbb{R}^L2 20–80 xRLx \in \mathbb{R}^L3
NucleotideTransf. xRLx \in \mathbb{R}^L4 500–2\,500 xRLx \in \mathbb{R}^L5

3. Training Regime and Embedding Strategies

HyenaDNA is pre-trained as a masked LLM on the full human reference genome (GRCh38/hg38), using contexts of up to xRLx \in \mathbb{R}^L6 bases. The objective is standard cross-entropy over masked positions:

xRLx \in \mathbb{R}^L7

where xRLx \in \mathbb{R}^L8 flags masked positions. The model learns nucleotide representations at high resolution.

For downstream inference, HyenaDNA serves as a frozen feature extractor (retrieval-augmented pipeline). For input xRLx \in \mathbb{R}^L9 of length ThT_h0, the output ThT_h1 from the last layer is mean-aggregated and L2-normalized to yield sequence embedding ThT_h2, which is suitable for k-NN retrieval or as input to lightweight classifiers (Datta et al., 6 Aug 2025).

Enhancer classification with z-Curve features, which capture 3D sequence geometry, demonstrates that HyenaDNA embeddings combined with z-Curve systematically improve accuracy (from ThT_h3 base to ThT_h4 with z-Curve) and achieve inference >8ThT_h5 faster with 8ThT_h6 lower COThT_h7 emissions than fine-tuned models (Datta et al., 6 Aug 2025).

4. Adapter-Based Modalities: CodonMoE and RNA Analysis

HyenaDNA forms the basis for plug-in adapters enabling DNA-trained models to function on RNA-centric tasks. The CodonMoE adapter applies a codon-level mixture-of-experts (MoE) on HyenaDNA outputs:

  • Every three nucleotides are averaged to form codon embeddings ThT_h8.
  • A gating network ThT_h9 assigns weights hkh_k0 for hkh_k1 expert MLPs hkh_k2.
  • Output:

hkh_k3

  • Codon features are tiled back to nucleotide resolution, residual connections and normalization applied, and a lightweight head yields the property prediction.

The CodonMoE architecture is a universal approximator for codon-to-RNA mappings, with formal guarantees. Standard configurations (hkh_k4–hkh_k5 experts, hkh_k6–hkh_k7M params) or the "pro" version (hkh_k8M params) add 1D convolutions over codon neighborhoods, all maintaining sub-quadratic complexity, hkh_k9 (Du et al., 6 Aug 2025).

Benchmarks in mRNA expression and stability demonstrate state-of-the-art rank correlations (Spearman's γθ(k)\gamma_\theta(k)0 up to γθ(k)\gamma_\theta(k)1), with HyenaDNA+CodonMoE matching or exceeding specialized RNA models (CodonBERT, SpliceBERT) at %%%%52Dx1,Dx2RL×LD_{x_1}, D_{x_2} \in \mathbb{R}^{L \times L}053%%%% of their parameter count, and delivering 5–10γθ(k)\gamma_\theta(k)4 faster inference (Du et al., 6 Aug 2025).

5. Empirical Results and Downstream Applications

HyenaDNA sets new top-1 accuracy on regulatory element classification, enhancer detection, chromatin profile prediction, and species assignment, often outperforming baseline CNNs, DNABERT, GPT-style transformers, and Nucleotide Transformer models. On GenomicBenchmarks, it achieves γθ(k)\gamma_\theta(k)5–γθ(k)\gamma_\theta(k)6 point improvements over prior state-of-the-art for multiple tasks (Nguyen et al., 2023).

Embedding-extraction pipelines using HyenaDNA maintain strong predictive performance across data splits with shifted distributions, indicating superior generalization to unseen genomic contexts compared to models reliant on full fine-tuning. Carbon calculations confirm γθ(k)\gamma_\theta(k)7–γθ(k)\gamma_\theta(k)8 lower emissions for retrieval-augmented approaches.

In rare disease genomics (Saadat et al., 2024):

  • HyenaDNA is used to generate sample- and variant-aware gene embeddings by processing full gene sequences personalized by individual pathogenic variants.
  • Embeddings at variant positions are averaged to yield a dynamic gene embedding γθ(k)\gamma_\theta(k)9, which is sensitive to deleterious changes.
  • Embeddings across genes are used as features in a protein-protein interaction (PPI) graph neural network and further refined by a genetic algorithm to extract functionally-coherent diagnosis subnetworks.
  • The workflow re-identifies known disease genes (e.g., IFIH1 with kk0, kk1) and pathways (e.g., interferon signaling), validating the capacity for interpretable target discovery.

6. Broader Implications and Future Directions

The HyenaDNA framework demonstrates that single nucleotide–resolution models with sub-quadratic scaling enable both parameter-efficient and computation-efficient solutions for broad genomic inference tasks. CodonMoE and similar adapters "RNA-ize" pretrained DNA models without full RNA pretraining, illustrating a unifying template for multi-modality in genomics. The plug-and-play MoE principle is extensible: analogous adapters could exploit amino-acid context for protein tasks or be designed for locus-level tasks in chromatin modeling (Du et al., 6 Aug 2025).

Practically, HyenaDNA’s architecture and embedding strategies favor scalable, low-footprint, and robust solutions, making them suitable for resource-constrained high-throughput genomics and for systematic interrogation of genome function in rare or common diseases. The variant-sensitivity of embeddings highlights the model’s utility for personalized genomics, offering precise and explainable representations optimized for downstream machine learning integration.

A plausible implication is that continued refinement of such architectures and adapter-based multi-modal pipelines may further reduce the computational and carbon cost of universal genomic modeling while enabling rapid, interpretable hypothesis generation across increasingly complex biological tasks.

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to HyenaDNA.