Author Login Chief Editor Login Reviewer Login Editor Login Remote Office

Computer Engineering ›› 2026, Vol. 52 ›› Issue (3): 41-61. doi: 10.19678/j.issn.1000-3428.0069925

• Frontier Perspectives and Reviews • Previous Articles     Next Articles

Research Progress of Interpretable Artificial Intelligence

LIAO Yong1,*(), HAN Xiaojin1, LIU Jinlin1, WANG Hao2   

  1. 1. School of Microelectronics and Communication Engineering, Chongqing University, Chongqing 400044, China
    2. Chongqing Shouxun Technology Co., Ltd., Chongqing 401147, China
  • Received:2024-05-27 Revised:2024-08-26 Online:2026-03-15 Published:2024-10-28
  • Contact: LIAO Yong

可解释人工智能研究进展

廖勇1,*(), 韩小金1, 刘金林1, 汪浩2   

  1. 1. 重庆大学微电子与通信工程学院, 重庆 400044
    2. 重庆首讯科技股份有限公司, 重庆 401147
  • 通讯作者: 廖勇
  • 作者简介:

    廖勇(CCF杰出会员), 男, 副研究员、博士、博士生导师, 主研方向为移动通信、人工智能技术及其应用

    韩小金, 硕士研究生

    刘金林, 硕士研究生

    汪浩, 高级工程师、硕士

  • 基金资助:
    重庆市自然科学基金(CSTB2023NSCQ-MSX0025)

Abstract:

Artificial intelligence has made remarkable progress across many fields, encouraging countries to attach great importance to its research and development. However, the rapid development of artificial intelligence has also brought about a series of problems and threats, and overreliance on and blind trust in such models can lead to serious risks. Therefore, interpretable artificial intelligence has become a key element in building trusted and transparent intelligent systems, and its research and development requires immediate attention. This survey comprehensively summarizes the research progress on explainable artificial intelligence at home and abroad comprehensively from multiple dimensions and levels. Based on current research results in the industry, this survey subdivides the key technologies of explainable artificial intelligence into four categories: interpretation model, interpretation method, safety testing, and experimental verification, with the aim of clarifying the technical focus and development direction of each field. Furthermore, the survey explores specific application examples of explainable artificial intelligence across key industry sectors, including but not limited to education, healthcare, finance, autonomous driving, and justice, demonstrating the significant role it plays in enhancing decision-making transparency. Finally, this survey provides an in-depth analysis of the major technical challenges of interpretable artificial intelligence and presents future development trends, in addition to a special investigation and in-depth analysis of the interpretability of large models, which has attracted considerable attention recently.

Key words: interpretability, trustworthy, artificial intelligence, demonstration application, large model

摘要:

人工智能在诸多领域的应用取得了突破性的进展, 引起了全球各国对其研发的高度重视。然而, 人工智能的快速发展也带来了一系列的问题, 过度依赖和盲目信任人工智能模型可能导致严重的风险。因此, 可解释人工智能成为构建可信、透明的智能系统的关键要素, 其研发变得尤为迫切。为此, 本文综述可解释人工智能的国内外研究进展, 从多维度、多层次进行全面梳理与归纳。首先, 基于当前行业内的研究成果, 将可解释人工智能的关键技术细分为解释模型、解释方法、安全测试及实验验证4类, 旨在明确各领域的技术焦点与发展方向。然后, 探讨可解释人工智能在多个关键行业领域的具体应用实例, 包括但不限于教育、医疗、金融、自动驾驶及司法等, 展示其在提升决策透明度等方面的重要作用。最后, 深入剖析可解释人工智能当前面临的主要技术挑战, 并展望其未来的发展趋势, 尤其针对当前备受瞩目的大模型可解释性问题, 进行了专项调研与探讨分析。

关键词: 可解释性, 可信, 人工智能, 示范应用, 大模型