Build faithful mouse models of human variants faster. See how H2M maps DNA and protein effects to improve disease modeling for drug discovery.

Model Human Variants in Mice Faster With H2M
A surprising amount of drug discovery still depends on one stubborn bottleneck: we can sequence millions of human genetic variants, but we can’t quickly turn the “right” ones into faithful animal models. Teams end up spending weeks doing manual lookups across ortholog tables, transcript annotations, codon differences, and surrounding sequence constraints—only to learn that the “same” variant behaves differently in mouse.
A new computational pipeline called H2M (human-to-mouse) tackles that bottleneck head-on. It builds a high-throughput “dictionary” of human variants and their best mouse equivalents—down to DNA change options that produce the same protein-level effect when the exact nucleotide swap won’t.
For pharma and biotech teams working in the AI in Pharmaceuticals & Drug Discovery space, this matters for a simple reason: better disease models create better decisions earlier—from target validation and mechanism studies to biomarker work and even immuno-oncology neoantigen testing.
The real problem: “equivalent variant” is not a simple lookup
An “equivalent” variant sounds straightforward: find the mouse ortholog, edit the same position, done. In practice, most companies get this wrong.
Here’s what derails cross-species variant modeling:
- Orthology is messy. Not every human gene maps cleanly 1:1 to mouse. Some map 1:many, many:many, or not at all.
- Transcripts don’t match. The “canonical” transcript in humans may not align to the mouse transcript you care about in your model.
- Local sequence context changes everything. Even when an amino acid is conserved, the surrounding nucleotides (and codons) often aren’t.
- Same DNA change ≠same protein change. Because of codon redundancy and frame context, the identical nucleotide substitution in mouse can yield a different amino acid outcome.
Drug discovery teams feel this pain in very concrete ways:
- A target looks validated in a GEMM… but the engineered mutation wasn’t functionally comparable.
- A resistance mutation is modeled, but the protein effect isn’t matched, leading to misleading PK/PD hypotheses.
- Immunogenicity screens don’t translate because the neoepitope in mouse isn’t the neoepitope in humans.
H2M’s core contribution is making variant equivalence explicit, standardized, and scalable.
What H2M does (and why it’s useful for drug discovery)
H2M is a computational pipeline that takes human variant data and predicts how to model the “same” variant in mouse—using both DNA-level and protein-level reasoning.
The authors built:
- A Python package (compatible with common variant formats)
- An online portal for browsing and downloading results
- A large human→mouse mapping database
The scale is the headline: 3,171,709 human-to-mouse mutation mappings (database version 1, built from publicly cataloged clinical variants). They report that over 80% of human variants can be modeled in mouse, with higher coverage for coding variants than deep noncoding ones.
This connects cleanly to how AI is being used in pharma R&D: automate the expensive “glue work” between data and experiments. Variant modeling is exactly that glue.
H2M’s four steps (the “assembly line”)
H2M runs four major steps:
- Query orthologous genes (human↔mouse)
- Align transcripts or peptides (depending on coding vs noncoding)
- Simulate mutations
- Check and model functional effects at nucleotide and peptide levels
Under the hood, it aligns sequences and then decides how to represent the variant across species in a way that’s honest about what’s conserved and what isn’t.
NCE vs PCE: the key idea you’ll want your team to adopt
H2M formalizes two types of “effect”:
- NCE (nucleotide change effect): the DNA-level edit
- PCE (peptide change effect): the protein-level change (amino acid substitution/indel effect)
That distinction is more than academic. In drug discovery, PCE is often what you truly care about (target binding site changes, conformational effects, catalytic residue changes), while NCE drives feasibility (can CRISPR base/prime editing do it efficiently at that locus?).
H2M supports three modeling strategies:
- NCE-only: same DNA alteration (often for noncoding or frameshifts)
- NCE-for-PCE: same DNA change and same protein effect
- Extended NCE-for-PCE: different DNA change in mouse to achieve the same protein effect (using codon redundancy)
If you’ve ever had a team argue “we modeled the human mutation” when they really modeled a nearby proxy, you can see why this structure helps.
Why this matters right now: AI, genome editing, and model throughput
December 2025 is a weirdly good time to care about variant-to-model translation.
Two trends are colliding:
- Variant volume keeps exploding (clinical sequencing + tumor profiling + population studies)
- Precision genome editing keeps getting more practical (base editing and prime editing are moving from boutique methods into routine platforms)
But most organizations still have a throughput mismatch: variant interpretation is fast; variant modeling is slow.
H2M shrinks that gap by producing standardized outputs you can plug into downstream workflows:
- Coordinates and edits in standardized formats
- HGVS-style annotations for protein changes n- Compatibility signals for base/prime editing design
In the paper’s demonstration subset, the team paired H2M with a guide design workflow to produce large editing libraries, including:
- 24,680 base-editing gRNAs for thousands of mutations
- 48,255 prime-editing gRNAs for thousands more
From a drug discovery perspective, that’s a signal that variant modeling is shifting from one-off artisan builds to library-scale experimentation—which is exactly where AI-driven discovery gets powerful.
Practical use cases in pharma: where H2M fits in the pipeline
The fastest way to evaluate H2M is to map it onto concrete R&D decisions.
1) Target validation that actually matches the human allele
Answer first: H2M helps you avoid “false validation” caused by non-equivalent mouse edits.
If your therapeutic hypothesis depends on a specific human missense change (say, a kinase domain alteration), H2M lets you:
- Confirm whether the same residue exists in mouse
- Quantify local conservation with a tunable “flank size” metric
- Choose between modeling the identical DNA change or the identical amino acid change
That last part is critical. Many teams over-index on “same nucleotide change” because it feels faithful. If that nucleotide change doesn’t yield the same amino acid substitution in mouse, you’ve built a different biological question.
2) Resistance modeling for oncology and anti-infectives
Answer first: For resistance, PCE fidelity beats everything.
Resistance mutations often work by changing a binding pocket or a protein interaction surface. A mouse model that matches the protein change but uses a different codon is usually the better choice.
H2M’s “extended NCE-for-PCE” strategy is designed for this exact issue: produce the same amino acid substitution even when the DNA-level edit must differ between species.
3) Immuno-oncology neoantigen screening that doesn’t mislead you
Answer first: Mouse models can validate human neoantigens only when the peptide outcome is conserved.
The authors explored mapping mutation-derived neoantigen peptides between species and found that many predicted peptide pairs can be presented in both humans and mice—but the correlation isn’t automatically strong. Translation depends on peptide identity, MHC context, and local sequence.
For drug developers building cancer vaccines, TCR therapies, or bispecifics, this reinforces a practical stance:
- Use cross-species mapping to filter for high-likelihood conserved neoepitopes
- Don’t assume murine immunogenicity means human immunogenicity
- Treat peptide-level verification as a first-class requirement, not a nice-to-have
4) Variant prioritization: conservation can be a feature, not a bias
Answer first: Conservation signals can help you pick variants that will behave similarly in mouse and human.
H2M introduces “flank size” as a way to quantify how much surrounding sequence matches across species at the mutation site. Larger conserved flanks tend to indicate more evolutionary constraint.
This is a helpful heuristic in early discovery:
- If you need a mouse model to de-risk a target fast, prioritize variants with stronger conserved context.
- If you’re investigating a human-specific regulatory mechanism, accept that mouse may be the wrong organism—or plan alternative systems.
In other words: modelability is part of variant prioritization, not an afterthought.
“People also ask” questions (answered plainly)
Can most human variants be modeled in mouse?
Yes—H2M predicts more than 80% of clinically observed human variants in their aggregated dataset can be modeled in mouse, with higher coverage for coding variants.
Why do noncoding variants translate worse across species?
Because noncoding regions (especially deep intronic and regulatory sequences) have lower sequence conservation and stronger species-specific context effects.
If the amino acid change is the same, is the model automatically equivalent?
No. Even with identical PCE, differences in expression, splicing, protein interaction networks, and compensatory biology can change phenotype. But matching PCE is still the correct starting point when mechanism depends on protein structure/function.
How to adopt H2M without creating workflow chaos
If you’re running computational biology or disease modeling in pharma, I’d implement H2M as a standard pre-modeling gate—not a bespoke tool a single scientist runs.
A workable rollout plan looks like this:
- Start with a high-value variant set (top clinical hotspots, resistance mutations, or biomarker-linked alleles)
- Define what “equivalence” means per program (NCE-only vs PCE-first)
- Set flank-size thresholds for different use cases (fast in vivo validation vs deep mechanistic modeling)
- Integrate with guide design (base/prime editing feasibility should be computed, not guessed)
- Track model fidelity in a registry (variant → modeling strategy → mouse edit → phenotype outcomes)
This is where AI in drug discovery becomes operational: repeatable pipelines, measurable assumptions, and fewer one-off decisions.
Where this goes next: model catalogs as R&D infrastructure
The bigger story isn’t just H2M. It’s the shift toward model catalogs as infrastructure—databases that connect clinical variants to edit designs, predicted functional impacts, and experimental readouts.
As variant effect predictors (including AI-based scoring) mature, the winning organizations will be the ones that connect:
- variant selection (clinical + computational signals)
- model feasibility (editing strategy + conservation)
- functional validation (phenotype + omics)
- therapeutic decision-making (go/no-go, biomarker strategy, patient stratification)
If you’re serious about AI in pharmaceuticals, you don’t just need better models. You need better ways to decide which models to build.
A forward-looking question worth asking internally: How many of our current mouse models would still be considered “equivalent” if we enforced explicit NCE/PCE definitions and conservation thresholds?