It is a sponsored article dropped at you by MBZUAI.
For those who’ve ever tried to guess how a cell will change form after a drug or a gene edit, it’s half science, half artwork, and largely costly trial-and-error. Imaging hundreds of situations is gradual; exploring thousands and thousands is unimaginable.
A new paper in Nature Communications proposes a special route: simulate these mobile “after” photographs instantly from molecular readouts, so you may preview the morphology earlier than you choose up a pipette. The workforce calls their mannequin MorphDiff, and it’s a diffusion mannequin guided by the transcriptome, the sample of genes turned up or down after a perturbation.
At a excessive stage, the thought flips a well-recognized workflow. Excessive-throughput imaging is a confirmed approach to uncover a compound’s mechanism or spot bioactivity however profiling each candidate drug or CRISPR goal isn’t possible. MorphDiff learns from instances the place each gene expression and cell morphology are identified, then makes use of solely the L1000 gene expression profile as a situation to generate lifelike post-perturbation photographs, both from scratch or by remodeling a management picture into its perturbed counterpart. The declare is that aggressive constancy on held-out (unseen) perturbations throughout massive drug and genetic datasets plus features on mechanism-of-action (MOA) retrieval can rival actual photographs.
This analysis led by MBZUAI researchers begins from a organic remark: gene expression in the end drives proteins and pathways that form what a cell seems like underneath the microscope. The mapping isn’t one-to-one, however there’s sufficient shared sign for studying. Conditioning on the transcriptome affords a sensible bonus too: there’s merely much more publicly accessible L1000 information than paired morphology, making it simpler to cowl a large swath of perturbation house. In different phrases, when a brand new compound arrives, you’re more likely to discover its gene signature which MorphDiff can then leverage.
Underneath the hood, MorphDiff blends two items. First, a Morphology Variational Autoencoder (MVAE) compresses five-channel microscope photographs right into a compact latent house and learns to reconstruct them with excessive perceptual constancy. Second, a Latent Diffusion Mannequin learns to denoise samples in that latent house, steering every denoising step with the L1000 vector through consideration.
Wang et al., Nature Communications (2025), CC BY 4.0
Diffusion is an efficient match right here: it’s intrinsically sturdy to noise, and the latent house variant is environment friendly sufficient to coach whereas preserving picture element. The workforce implements each gene-to-image (G2I) era (begin from noise, situation on the transcriptome) and image-to-image (I2I) transformation (push a management picture towards its perturbed state utilizing the identical transcriptomic situation). The latter requires no retraining due to an SDEdit-style process, which is helpful once you need to clarify modifications relative to a management.
It’s one factor to generate photogenic photos; it’s one other to generate biologically trustworthy ones. The paper leans into each: on the generative facet, MorphDiff is benchmarked in opposition to GAN and diffusion baselines utilizing commonplace metrics like FID, Inception Rating, protection, density, and a CLIP-based CMMD. Throughout JUMP (genetic) and CDRP/LINCS (drug) check splits, MorphDiff’s two modes sometimes land first and second, with significance assessments run throughout a number of random seeds or impartial management plates. The result’s constant: higher constancy and variety, particularly on OOD perturbations the place sensible worth lives.
The larger image is that generative AI has lastly reached a constancy stage the place in-silico microscopy can stand in for first-pass experiments.
Extra attention-grabbing for biologists, the authors step past picture aesthetics to morphology options. They extract lots of of CellProfiler options (textures, intensities, granularity, cross-channel correlations) and ask whether or not the generated distributions match the bottom fact.
In side-by-side comparisons, MorphDiff’s characteristic clouds line up with actual information extra intently than baselines like IMPA. Statistical assessments present that over 70 p.c of generated characteristic distributions are indistinguishable from actual ones, and feature-wise scatter plots present the mannequin accurately captures variations from management on essentially the most perturbed options. Crucially, the mannequin additionally preserves correlation construction between gene expression and morphology options, with increased settlement to floor fact than prior strategies, proof that it’s modeling greater than floor fashion.
Wang et al., Nature Communications (2025), CC BY 4.0
The drug outcomes scale up that story to hundreds of therapies. Utilizing DeepProfiler embeddings as a compact morphology fingerprint, the workforce demonstrates that MorphDiff’s generated profiles are discriminative: classifiers skilled on actual embeddings additionally separate generated ones by perturbation, and pairwise distances between drug results are preserved.
Wang et al., Nature Communications (2025), CC BY 4.0
That issues for the downstream process everybody cares about: MOA retrieval. Given a question profile, can you discover reference medicine with the identical mechanism? MorphDiff’s generated morphologies not solely beat prior image-generation baselines but additionally outperform retrieval utilizing gene expression alone, they usually method the accuracy you get utilizing actual photographs. In top-k retrieval experiments, the common enchancment over the strongest baseline is 16.9 p.c and eight.0 p.c over transcriptome-only, with robustness proven throughout a number of okay values and metrics like imply common precision and folds-of-enrichment. That’s a robust sign that simulated morphology accommodates complementary info to chemical construction and transcriptomics which is sufficient to assist discover look-alike mechanisms even when the molecules themselves look nothing alike.
MorphDiff’s generated morphologies not solely beat prior image-generation baselines but additionally outperform retrieval utilizing gene expression alone, they usually method the accuracy you get utilizing actual photographs.
The paper additionally lists some present limitations that trace at potential future enhancements. Inference with diffusion stays comparatively gradual; the authors recommend plugging in newer samplers to hurry era. Time and focus (two components that biologists care about) aren’t explicitly encoded as a consequence of information constraints; the structure might take them as further situations when matched datasets change into accessible. And since MorphDiff relies on perturbed gene expression as enter, it may well’t conjure morphology for perturbations that lack transcriptome measurements; a pure extension is to chain with fashions that predict gene expression for unseen medicine (the paper cites GEARS for instance). Lastly, generalization inevitably weakens as you stray removed from the coaching distribution; bigger, better-matched multimodal datasets will assist, as will conditioning on extra modalities equivalent to buildings, textual content descriptions, or chromatin accessibility.
What does this imply in observe? Think about a screening workforce with a big L1000 library however a smaller imaging funds. MorphDiff turns into a phenotypic copilot: generate predicted morphologies for brand spanking new compounds, cluster them by similarity to identified mechanisms, and prioritize which to picture for affirmation. As a result of the mannequin additionally surfaces interpretable characteristic shifts, researchers can peek underneath the hood. Did ER texture and mitochondrial depth transfer the way in which we’d anticipate for an EGFR inhibitor? Did two structurally unrelated molecules land in the identical phenotypic neighborhood? These are the sorts of hypotheses that speed up mechanism looking and repurposing.
The larger image is that generative AI has lastly reached a constancy stage the place in-silico microscopy can stand in for first-pass experiments. We’ve already seen text-to-image fashions explode in shopper domains; right here, a transcriptome-to-morphology mannequin exhibits that the identical diffusion equipment can do scientifically helpful work equivalent to capturing refined, multi-channel phenotypes and preserving the relationships that make these photographs greater than eye sweet. It received’t substitute the microscope. But when it reduces the variety of plates it’s important to run to search out what issues, that’s money and time you may spend validating the hits that rely.
