《计算机应用》唯一官方网站 ›› 2024, Vol. 44 ›› Issue (6): 1655-1662.DOI: 10.11772/j.issn.1001-9081.2023060885

• CCF第38届中国计算机应用大会 (CCF NCCA 2023) •    下一篇

大语言模型的技术应用前景与风险挑战

徐月梅1(), 胡玲1, 赵佳艺1, 杜宛泽2, 王文清2   

  1. 1.北京外国语大学 信息科学技术学院,北京 100089
    2.北京大学 软件与微电子学院,北京 102600
  • 收稿日期:2023-07-06 修回日期:2023-08-09 接受日期:2023-08-15 发布日期:2023-09-14 出版日期:2024-06-10
  • 通讯作者: 徐月梅
  • 作者简介:胡玲(2000—),女,江西南昌人,硕士研究生,主要研究方向:自然语言处理
    赵佳艺(2002—),女,陕西渭南人,主要研究方向:自然语言处理
    杜宛泽(2000—),女,北京人,硕士研究生,主要研究方向:自然语言处理
    王文清(2000—),女,河南商丘人,硕士研究生,主要研究方向:自然语言处理。
  • 基金资助:
    中央高校基本科研业务费专项(2022JJ006)

Technology application prospects and risk challenges of large language models

Yuemei XU1(), Ling HU1, Jiayi ZHAO1, Wanze DU2, Wenqing WANG2   

  1. 1.School of Information Science and Technology,Beijing Foreign Studies University,Beijing 100089,China
    2.School of Software and Microelectronics,Peking University,Beijing 102600,China
  • Received:2023-07-06 Revised:2023-08-09 Accepted:2023-08-15 Online:2023-09-14 Published:2024-06-10
  • Contact: Yuemei XU
  • About author:HU Ling, born in 2000, M. S. candidate. Her research interests include natural language processing.
    ZHAO Jiayi, born in 2002. Her research interests include natural language processing.
    DU Wanze, born in 2000, M. S. candidate. Her research interests include natural language processing.
    WANG Wenqing, born in 2000, M. S. candidate. Her research interests include natural language processing.
  • Supported by:
    Fundamental Research Funds for Central Universities(2022JJ006)

摘要:

针对大语言模型(LLM)技术的快速发展,剖析它的技术应用前景和风险挑战,对通用人工智能(AGI)的发展和治理有重要参考价值。首先,以Multi-BERT(Multilingual Bidirectional Encoder Representations from Transformers)、GPT(Generative Pre-trained Transformer)和ChatGPT(Chat Generative Pre-Trained Transformer)等语言模型为代表,综述LLM的发展脉络、核心技术和评估体系;其次,分析LLM现存的技术局限和安全风险;最后,提出LLM在技术上改进、政策上跟进的建议。分析指出作为发展阶段的LLM,现有模型存在非真实性及偏见性输出、实时自主学习能力欠缺,算力需求庞大,对数据质量和数量依赖性强,语言风格单一;存在数据隐私、信息安全和伦理等方面的安全风险。未来发展可从技术上继续改进,从“大规模”转向“轻量化”、从“单模态”走向“多模态”、从“通用”迈入“垂类”;从政策上实时跟进,实施有针对性的监管措施,规范应用和发展。

关键词: 大语言模型, 风险挑战, 技术监管, 应用前景, 通用人工智能

Abstract:

In view of the rapid development of Large Language Model (LLM) technology, a comprehensive analysis was conducted on its technical application prospects and risk challenges which has great reference value for the development and governance of Artificial General Intelligence (AGI). Firstly, with representative language models such as Multi-BERT (Multilingual Bidirectional Encoder Representations from Transformer), GPT (Generative Pre-trained Transformer) and ChatGPT (Chat Generative Pre-trained Transformer) as examples, the development process, key technologies and evaluation systems of LLM were reviewed. Then, a detailed analysis of LLM on technical limitations and security risks was conducted. Finally, suggestions were put forward for technical improvement and policy follow-up of the LLM. The analysis indicates that at a developing status, the current LLMs still produce non-truthful and biased output, lack real-time autonomous learning ability, require huge computing power, highly rely on data quality and quantity, and tend towards monotonous language style. They have security risks related to data privacy, information security, ethics, and other aspects. Their future developments can continue to improve technically, from “large-scale” to “lightweight”, from “single-modal” to “multi-modal”, from “general-purpose” to “vertical”; for real-time follow-up in policy, their applications and developments should be regulated by targeted regulatory measures.

Key words: Large Language Model (LLM), risk challenge, technology supervision, application prospect, Artificial General Intelligence (AGI)

中图分类号: