From phishing to bias: Study maps the hidden threats behind large language models
en-GBde-DEes-ESfr-FR

From phishing to bias: Study maps the hidden threats behind large language models

25.12.2025 TranSpread

Large language models (LLMs) such as generative pre-trained transformer (GPT), bidirectional encoder representations from transformers (BERT), and T5 have transformed sectors ranging from education and healthcare to digital governance. Their ability to generate fluent, human-like text enables automation and accelerates information workflows. However, this same capability increases exposure to cyber-attacks, model manipulation, misinformation, and biased outputs that can mislead users or amplify social inequalities. Academic researchers warn that without systematic regulation and defense mechanisms, LLM misuse may threaten data security, public trust, and social stability. Based on these challenges, further research is required to improve model governance, strengthen defenses, and mitigate ethical risks.

A research team from Shanghai Jiao Tong University and East China Normal University published (DOI: 10.1007/s42524-025-4082-6) a comprehensive review in Frontiers of Engineering Management, (2025) examining ethical security risks in large language models. The study screened over 10,000 documents and distilled 73 key works to summarize threats such as phishing attacks, malicious code generation, data leakage, hallucination, social bias, and jailbreaking. The review further evaluates defense tools including adversarial training, input preprocessing, watermarking, and model alignment strategies.

The review categorizes LLM-related security threats into two major domains: misuse-based risks and malicious attacks targeting models. Misuse includes phishing emails crafted with near-native fluency, automated malware scripting, identity spoofing, and large-scale false information production. Malicious attacks appear at both data/model level—such as model inversion, poisoning, extraction—and user interaction level including prompt injection and jailbreak techniques. These attacks may access private training data, bypass safety filters, or induce harmful content output.

On defense strategy, the study summarizes three mainstream technical routes: parameter processing, which removes redundant parameters to reduce attack exposure; input preprocessing, which paraphrases prompts or detects adversarial triggers without retraining; and adversarial training, including red-teaming frameworks that simulate attacks for robustness improvement. The review also introduces detection technologies like semantic watermarking and CheckGPT, which can identify model-generated text with up to 98–99% accuracy. Despite progress, defenses often lag behind evolving attacks, indicating urgent need for scalable, low-cost, multilingual-adaptive solutions.

The authors emphasize that technical safeguards must coexist with ethical governance. They argue that hallucination, bias, privacy leakage, and misinformation are social-level risks, not merely engineering problems. To ensure trust in LLM-based systems, future models should integrate transparency, verifiable content traceability, and cross-disciplinary oversight. Ethical review frameworks, dataset audit mechanisms, and public awareness education will become essential in preventing misuse and protecting vulnerable groups.

The study suggests that secure and ethical development of LLMs will shape how societies adopt AI. Robust defense systems may protect financial systems from phishing, reduce medical misinformation, and maintain scientific integrity. Meanwhile, watermark-based traceability and red-teaming may become industry standards for model deployment. The researchers encourage future work toward AI responsible governance, unified regulation frameworks, safer training datasets, and model transparency reporting. If well-managed, LLMs can evolve into reliable tools supporting education, digital healthcare, and innovation ecosystems while minimizing risks linked to cybercrime and social misinformation.

###

References

DOI

10.1007/s42524-025-4082-6

Original Source URL

https://doi.org/10.1007/s42524-025-4082-6

Funding Information

This study was supported by the Beijing Key Laboratory of Behavior and Mental Health, Peking University, China.

About Frontiers of Engineering Management

Frontiers of Engineering Management(FEM) is an international academic journal supervised by the Chinese Academy of Engineering, focusing on cutting-edge management issues across all fields of engineering. The journal publishes research articles, reviews, and perspectives that advance theoretical and practical understanding in areas such as manufacturing, construction, energy, transportation, environmental systems, and logistics. FEM emphasizes methodologies in systems engineering, information management, technology and innovation management, as well as the management of large-scale engineering projects. Serving both scholars and industry leaders, the journal aims to promote knowledge exchange and support innovation in global engineering management.

Paper title: The ethical security of large language models: A systematic review
Angehängte Dokumente
  • Systematic selection process of relevant literature.
25.12.2025 TranSpread
Regions: North America, United States, Asia, China
Keywords: Science, Mathematics

Disclaimer: AlphaGalileo is not responsible for the accuracy of content posted to AlphaGalileo by contributing institutions or for the use of any information through the AlphaGalileo system.

Referenzen

We have used AlphaGalileo since its foundation but frankly we need it more than ever now to ensure our research news is heard across Europe, Asia and North America. As one of the UK’s leading research universities we want to continue to work with other outstanding researchers in Europe. AlphaGalileo helps us to continue to bring our research story to them and the rest of the world.
Peter Dunn, Director of Press and Media Relations at the University of Warwick
AlphaGalileo has helped us more than double our reach at SciDev.Net. The service has enabled our journalists around the world to reach the mainstream media with articles about the impact of science on people in low- and middle-income countries, leading to big increases in the number of SciDev.Net articles that have been republished.
Ben Deighton, SciDevNet
AlphaGalileo is a great source of global research news. I use it regularly.
Robert Lee Hotz, LA Times

Wir arbeiten eng zusammen mit...


  • e
  • The Research Council of Norway
  • SciDevNet
  • Swiss National Science Foundation
  • iesResearch
Copyright 2025 by DNN Corp Terms Of Use Privacy Statement