New engineered base editors cut bystander edits while staying efficient—opening safer gene therapy and precision medicine paths for pharma and biotech teams.

Precise Base Editing: Less Bystander, More Therapeutic
Base editing has always come with an uncomfortable trade-off: you can edit efficiently, or you can edit precisely—but getting both at once is hard. That tension is a big reason many pharma teams still treat base editing as “promising” rather than “ready,” especially when the goal is a therapeutic where every unintended nucleotide change becomes a safety discussion.
A new Nature Biotechnology paper published today (Dec 18, 2025) reports a pragmatic way out of that box: engineered adenine base editors that reduce bystander editing while maintaining (and sometimes increasing) on-target activity. The authors didn’t rely on one trick. They combined guide RNA engineering, directed evolution, and machine learning-guided protein design—a workflow that should feel familiar to anyone building AI-enabled discovery platforms.
For teams working in gene therapy, functional genomics, or target validation, this matters because precision editing isn’t just a “nice-to-have”—it changes what indications are feasible.
Why bystander editing is a clinical problem, not a lab nuisance
Bystander editing happens because many base editors act within a small “editing window” (often around five nucleotides). If the target window contains multiple editable bases, the enzyme can convert more than you intended. In coding regions, that can introduce missense changes or splicing alterations—and in a therapeutic setting, that’s a regulatory and safety headache.
A sobering number from prior analyses: if you’re using adenine base editors (ABEs) and there’s at least one bystander adenine in the window, there’s an estimated 38% probability that bystander edits produce a nonsilent mutation. That’s not a rounding error; it’s a program risk.
Here’s the operational reality I see across pharma:
- Many targetable pathogenic variants are “correctable” on paper, but unacceptable once you model bystanders.
- Screening teams spend weeks cycling through guides hoping to land on a window/context that behaves.
- Translation teams then face the “it worked in HEK cells” trap—only to discover context dependence in primary cells or iPS-derived models.
The paper’s core contribution is a design strategy that treats bystanders as a first-class optimization target, rather than a post-hoc filtering step.
The new playbook: guide engineering + directed evolution + protein language models
The authors combine three ideas that are each useful alone, but more powerful together:
- 3′-extended guide RNAs (“anchor-guide RNAs,” agRNAs) to narrow the functional editing window.
- Phage-assisted noncontinuous evolution (PANCE) with a custom selection circuit that rewards perfect edits and penalizes bystanders.
- Protein language models (PLMs) to propose “evolutionarily plausible” mutations that improve fitness/precision without brute-force wet-lab screening.
The throughline is a mindset shift: editing outcomes are an interaction between editor protein, guide architecture, and local sequence context. Optimizing only one component (protein-only or guide-only) leaves performance on the table.
1) agRNAs: guide architecture as a physical constraint
The team built and screened a library of roughly 60,000 3′-extended sgRNAs (agRNAs). The premise is mechanical: if the DNA/R-loop substrate has less freedom to move in the deaminase active site, the editor acts in a narrower window.
In a screening locus in the human DNMT1 gene, top agRNA candidates reduced bystander editing substantially. One guide (agRNA 56114) reduced edits at specific bystander positions by 44% and 34% while preserving desirable editing.
Why this is relevant to drug discovery workflows:
- It creates a new optimization knob—guide scaffold/extension design—not just spacer selection.
- It suggests your “guide design model” shouldn’t only score on-target binding and Cas9 off-targets. It should score editing distributions.
If your team already uses AI to propose gRNAs, this is a clear opportunity: incorporate engineered guide backbones (not only 20-nt spacers) into the search space.
2) PANCE with dual selection: evolve precision without killing activity
Most precision-focused editor engineering runs into the same wall: narrowing the window often reduces activity. The authors built a PANCE selection circuit designed specifically to avoid that trap.
Their selection does two things at once:
- Rewards perfect editing that restores a functional phage protein (pIII), increasing phage replication.
- Penalizes bystander edits that introduce amino acid changes harming pIII function, decreasing replication.
This matters because it mirrors how pharma discovery actually works: you rarely optimize one metric. You optimize efficacy under constraints (safety, PK, manufacturability).
From this evolution, they identified key TadA-8e deaminase variants. The standout is V28C, which showed:
- ~2–3× higher precision across many tested sites (especially in adenine-rich windows)
- ~20% higher on-target efficiency than ABE8e in tested pathogenic contexts
At the DNMT1 locus, pairing V28C + agRNA 56114 increased “perfect edit” outcomes dramatically (reported as 27.4% perfect editing in that assay context).
Just as important for translational teams: evolved variants also reduced Cas9-independent off-target DNA editing and RNA A-to-I off-target events. In RNA editing noise:
- V28C showed ~7.9-fold less A-to-I than ABE8e-SpRY
- L34W showed ~20.4-fold less A-to-I than ABE8e-SpRY
Those are the kinds of deltas that can change the shape of a safety package.
3) PLMs: faster jumps to plausible, functional mutations
Directed evolution is powerful, but slow and resource-intensive. The paper uses an ensemble of protein language models to propose single amino-acid substitutions likely to be compatible with protein function.
One mutation, M151E, narrowed the editing window and improved on-target editing in certain contexts. Interestingly, many ML-predicted changes were “reversions” toward ancestral TadA residues—hinting that some ABE8e mutations might be redundant depending on context.
This is a useful lesson for AI in drug discovery teams: ML doesn’t need to “solve the whole problem.” It can do what it’s best at—constraining the search space to mutations likely to fold, express, and function.
And when they cross-referenced PANCE results, position 151 also showed enrichment toward a negatively charged residue (aspartate) during evolution—meaning experimental selection and ML predictions converged on similar chemistry.
What V28C changes for therapeutic programs and target validation
The most compelling value of V28C isn’t a single benchmark. It’s how consistently it improves the edit distribution across many targets.
The authors tested thousands of pathogenic variants (around 12,000 sites in library experiments) and validated across 12 endogenous human genomic loci. Across those loci, V28C:
- Improved precision at every tested site
- Produced a constrained 4-adenine editing window
- Increased average on-target efficiency by 27.1% (reported across their endogenous panel)
That package—more precision and higher editing at the most-edited adenine—directly addresses the classic “precision vs potency” dilemma.
Example 1: PCSK9—practical potency in a validated therapeutic target
PCSK9 is already a marquee target in lipid lowering, including prior in vivo base editing work.
In this study, at a splice donor site strategy in PCSK9, the V28C variant achieved about 75.7% editing efficiency, outperforming the baseline ABE8e-nSpCas9-WT by 19.7% in that setup.
Even though PCSK9 editing doesn’t necessarily suffer from bystanders in the same way as coding corrections, higher efficiency at similar specificity is exactly what delivery-limited programs need.
Example 2: SNCA E46K—where bystanders can kill the idea
The SNCA E46K mutation is linked to early-onset Parkinson’s disease. This is the type of target where “pretty good” editing isn’t good enough: any extra amino-acid changes risk confounding phenotype or worsening aggregation.
The authors show that V28C improves perfect correction markedly:
- V28C achieved 11.6% perfect edits (as a fraction of edited reads) vs 0.65% for ABE8e
- That’s a 21.1-fold increase in precision by their reported metric
- Pairing with agRNA 56114 increased perfect edits further to 17.5% and 37.6-fold precision improvement over ABE8e in that comparison
This is a strong example of where precision engineering makes a previously “too messy” edit directionally feasible.
How pharma teams can apply this immediately (even without building new editors)
Most organizations won’t re-run PANCE next week. You don’t need to. The actionable part is the workflow philosophy.
Update your base editing design criteria
If your internal rubric still prioritizes “highest on-target percent” first, you’ll keep picking guides that fail later. Replace that with a distribution-first approach:
- Primary metric: fraction of reads that are the desired allele (no bystanders)
- Secondary metric: total on-target conversion
- Explicit filters: bystander edits that change amino acids, create cryptic splice sites, or alter regulatory motifs
A simple practical move is to track a “perfect edit rate” alongside editing efficiency for every guide/editor pair.
Treat guide architecture as a search space (and let AI help)
agRNAs make the guide scaffold part of the optimization. That’s a natural fit for AI-driven discovery because:
- The combinatorial space is large (this paper screened ~60k)
- Outcomes are measurable (NGS readouts across contexts)
- The goal is multi-objective (perfect edit rate vs efficiency)
If you’re building models for gene editing outcomes, broaden inputs beyond spacer sequence and PAM. Include:
- guide extension motifs/hairpins
- predicted secondary structure
- local adenine density and motif context (YA vs RA preferences)
Use “ML to propose, selection to verify” as a repeatable pattern
The most transferable concept is the division of labor:
- ML proposes plausible mutations (reducing wasted wet-lab cycles)
- selection/evolution enforces the true objective (precision under biological constraints)
This is the same pattern many AI in pharmaceuticals teams use for protein engineering and antibody optimization. The novelty here is applying it cleanly to genome editor proteins, where the phenotype is an editing distribution, not a binding affinity.
A good editor isn’t the one that edits the most. It’s the one that edits the right allele most often.
What to watch next in 2026: delivery, generalizability, and regulatory narratives
This paper is a tools paper, but the implications are product-facing.
Three things will determine how fast precision base editing translates into therapeutic pipelines:
- Generalizable agRNA rules. The paper shows strong effects but also suggests many targets may need custom agRNAs. Expect more data and, ideally, structural insight into guide-target complexes.
- Editor performance in primary cells and in vivo. V28C looks strong in HEK293T and iPS contexts; delivery constraints and immunogenicity concerns remain the real gatekeepers.
- Regulatory storytelling around bystanders. If perfect edit rates rise and RNA off-target signatures fall by an order of magnitude, the risk narrative changes from “unpredictable collateral” to “quantifiable residual.” That matters.
If you’re leading an AI in drug discovery initiative, this is also a reminder: AI’s impact isn’t limited to small molecules. It’s now shaping the engineering of the therapeutic modality itself.
Most pharma teams say they want safer, more predictable gene editing. The teams that win will be the ones who operationalize “precision” as a measurable, optimizable outcome—then build AI and wet-lab loops that improve it continuously.
Next step: make precision an engineering KPI
If base editing is part of your platform—target validation, disease modeling, or therapeutic development—set a near-term goal: quantify bystander risk systematically across your top programs, then identify where narrowed-window editors and engineered guides could rescue a target.
If you want help designing a precision-first evaluation framework (perfect edit rate, bystander consequence scoring, and ML-ready datasets), that’s exactly the kind of build that pays off across an entire genome editing portfolio.
What therapeutic target in your pipeline is currently “on hold” because bystander edits make it too risky—and what would you do if you could cut those bystanders 2–3× without losing potency?