作者投稿和查稿 主编审稿 专家审稿 编委审稿 远程编辑

计算机工程 ›› 2025, Vol. 51 ›› Issue (12): 56-67. doi: 10.19678/j.issn.1000-3428.0070658

• 热点与综述 • 上一篇    下一篇

一种基于不确定性校正的可信连续学习方法

沈勤丰1,2, 黄璐瑶1,*()   

  1. 1. 东南大学网络空间安全学院, 江苏 南京 211189
    2. 江阴市华姿中等专业学校, 江苏 无锡 214401
  • 收稿日期:2024-11-27 修回日期:2025-01-24 出版日期:2025-12-15 发布日期:2025-05-14
  • 通讯作者: 黄璐瑶

A Trusted Continuous Learning Method Based on Uncertainty Correction

SHEN Qinfeng1,2, HUANG Luyao1,*()   

  1. 1. School of Cyber Science and Engineering, Southeast University, Nanjing 211189, Jiangsu, China
    2. Jiangyin Huazi Secondary Specialized School, Wuxi 214401, Jiangsu, China
  • Received:2024-11-27 Revised:2025-01-24 Online:2025-12-15 Published:2025-05-14
  • Contact: HUANG Luyao

摘要:

连续学习是一种能在连续的数据流中训练模型的深度学习范式, 适合日益开放复杂的智能应用场景。连续学习最大的挑战是"灾难性遗忘", 即模型在学习新知识后会遗忘过去学习的知识。现有的连续学习忽视了不确定性对模型训练的影响, 且大多数方法都聚焦于连续学习后续阶段的改进, 对模型初始阶段研究较少。提出一种基于不确定性校正的可信连续学习方法, 通过在初始阶段约束模型输出的不确定性来弥补模型参数漂移带来的分类误差, 从而缓解"灾难性遗忘"。该方法可以与其他连续学习方法相结合并形成改进模型, 具有较强的通用性。对3种经典的连续学习方法进行改进, 实验结果表明所提方法均能有效提高原始模型的性能, 在2个数据集上平均分类准确率(ACA)提升1.2~19.1百分点。除此之外, 引入期望校准误差(ECE)评判连续学习模型的可靠性, 对比原始模型, 基于所提方法的改进模型具有更低的ECE, 这证明基于该方法的改进模型更加可信。

关键词: 连续学习, 不确定性, 校正, 信息熵, 知识蒸馏

Abstract:

As a deep learning paradigm that can be trained in continuous data streams, continuous learning is suitable for increasingly open and complex intelligent application scenarios. The main challenge in incremental learning is catastrophic forgetting, which refers to a precipitous drop in performance on previously learned tasks after learning a new task. State-of-the-art studies on continuous learning ignore the impact of uncertainty on model training. In addition, existing studies mainly focus on mitigating forgetting in the phases after the initial phase, whereas the role of the initial phase is largely neglected. Motivated by this, this study proposes a trusted continuous learning method based on uncertainty correction that constrains the uncertainty of a model at the initial stage. Thus, this constraint can alleviate errors caused by model parameter drift, and catastrophic forgetting can be relieved. The proposed method can be combined with other continuous learning methods; therefore, it is fairly universal; for example, this study improves three traditional continuous learning models using the proposed method, and the experimental results show that the improved models outperform the original ones, with the Average Classification Accuracy (ACA) improving by 1.2 to 19.1 percentage points on two datasets. The Expected Calibration Error (ECE) is used to evaluate the reliability of the models. The experimental results show that the improved models have a lower ECE, which proves that the proposed method improves the reliability of the original models.

Key words: continuous learning, uncertainty, correction, information entropy, knowledge distillation