Latent Diffusion Model Provides Efficient and High-quality Paraphrase
en-GBde-DEes-ESfr-FR

Latent Diffusion Model Provides Efficient and High-quality Paraphrase

05/02/2026 Frontiers Journals

Paraphrase generation requires diverse generation of high-quality utterance by the given semantics, which is a challenge for traditional end-2-end text generation.
Inspired by the diffusion modeling for diverse image generations, a research team from Nanjing University led by Wei Zou managed to reconcile the quality and diversity for paraphrase generation via latent diffusion modeling (Latent Diffusion Paraphraser, LDP), and published their new research on 15 January 2026 in Frontiers of Computer Science co-published by Higher Education Press and Springer Nature.
The team proposed a latent diffusion modeling at encoded text space, which offers a controllable semantic intervention. The off-the-shelf pre-trained text encoder and decoder bridges the diffusion semantic space with the valid text, preventing overheads of diffusion process for straightforward text space. Compared to the straightforward textual diffusion generations, LDP shares the similar efficiency of traditional end-2-end text generation
model.
The diffusion process allows additional semantic controls to ensure the paraphrase quality, where semantic controls are fine-tuned by sampled input segments from training inputs without additional annotations, which is a great convenience.
Experiments verified the LDP in English paraphrase generation on Quora Question Pair, Twitter-URL, PAWS-wiki paraphrase datasets, with a state-of-the-art generation even comparable to the open-source large language model generations, but at much cheaper costs. LDP with input semantic controls also surpass the diffusion baselines in generation quality and efficiency. Further analysis shows that LDP with its controller implementation is also helpful for similar text generations with diversity and quality requirements, such as question generation and domain adaptation.
DOI:10.1007/s11704-025-40633-9
Attached files
  • Figure 1: LDP consists of a diffusion modeling for encoded text space of an off-the-shelf pre-trained encoder and decoder, the diffusion process can be intervened by additional controller .
05/02/2026 Frontiers Journals
Regions: Asia, China
Keywords: Applied science, Computing

Disclaimer: AlphaGalileo is not responsible for the accuracy of content posted to AlphaGalileo by contributing institutions or for the use of any information through the AlphaGalileo system.

Testimonials

For well over a decade, in my capacity as a researcher, broadcaster, and producer, I have relied heavily on Alphagalileo.
All of my work trips have been planned around stories that I've found on this site.
The under embargo section allows us to plan ahead and the news releases enable us to find key experts.
Going through the tailored daily updates is the best way to start the day. It's such a critical service for me and many of my colleagues.
Koula Bouloukos, Senior manager, Editorial & Production Underknown
We have used AlphaGalileo since its foundation but frankly we need it more than ever now to ensure our research news is heard across Europe, Asia and North America. As one of the UK’s leading research universities we want to continue to work with other outstanding researchers in Europe. AlphaGalileo helps us to continue to bring our research story to them and the rest of the world.
Peter Dunn, Director of Press and Media Relations at the University of Warwick
AlphaGalileo has helped us more than double our reach at SciDev.Net. The service has enabled our journalists around the world to reach the mainstream media with articles about the impact of science on people in low- and middle-income countries, leading to big increases in the number of SciDev.Net articles that have been republished.
Ben Deighton, SciDevNet

We Work Closely With...


  • e
  • The Research Council of Norway
  • SciDevNet
  • Swiss National Science Foundation
  • iesResearch
Copyright 2026 by AlphaGalileo Terms Of Use Privacy Statement