作者投稿和查稿 主编审稿 专家审稿 编委审稿 远程编辑

计算机工程 ›› 2024, Vol. 50 ›› Issue (7): 271-281. doi: 10.19678/j.issn.1000-3428.0068104

• 图形图像处理 • 上一篇    下一篇

结合注意力和低光增强的夜间语义分割

肖慈, 徐杨*(), 张永丹, 冯明文, 黄易仟   

  1. 贵州大学大数据与信息工程学院, 贵州 贵阳 550025
  • 收稿日期:2023-07-19 出版日期:2024-07-15 发布日期:2023-12-05
  • 通讯作者: 徐杨
  • 基金资助:
    贵州省科技计划项目(黔科合支撑一般326)

Nighttime Semantic Segmentation with Attention and Low-Light Enhancement

Ci XIAO, Yang XU*(), Yongdan ZHANG, Mingwen FENG, Yiqian HUANG   

  1. College of Big Data and Information Engineering, Guizhou Universivy, Guiyang 550025, Guizhou, China
  • Received:2023-07-19 Online:2024-07-15 Published:2023-12-05
  • Contact: Yang XU

摘要:

随着深度学习技术的发展和计算能力的提升, 对白天拍摄的自然场景图像进行语义分割能够取得良好的效果。然而, 在夜间图像语义分割任务中, 由于存在曝光不平衡、缺乏标记数据等问题, 由白天数据训练的模型往往无法取得良好的表现。为此, 提出一种新的无监督夜间图像语义分割网络(AI-USeg)。首先, 使用一个轻量级的自校准照明网络(SCI)对夜间图像进行增强, 以减少光照变化对后续语义分割网络的影响; 其次, 引入领域自适应(DA)方法, 将模型从包含大量有标签数据的Cityscapes自适应到Dark Zurich-D, 解决缺乏标记数据的问题; 随后, AI-USeg在基于全卷积网络(FCN)实现的判别器中引入SENet, 通过在输出空间进行对抗学习来适应夜间低光照环境下的图像特征, 以提升夜间图像语义分割任务的效果。实验使用Cityscapes和Dark Zurich-train中的2 416个昼夜图像对进行无监督训练, 结果表明, AI-USeg在Dark Zurich-test和Nighttime Driving-test上的平均交并比(mIoU)分别达到了47.9%和51.5%, 相较于MGCDA方法分别提高了5.4和2.1个百分点。AI-USeg对夜间图像的特征适应性更强, 具有更高的鲁棒性, 为夜间场景下的图像分割任务提供了一种有效的解决方案。

关键词: 深度学习, 语义分割, 自动驾驶, 低光图像增强, 注意力机制

Abstract:

With the development of deep learning technology and the improvements in computing power, semantic segmentation of natural scene images captured during the day shows promising results. However, in nighttime image semantic segmentation tasks, models trained on daytime data often fail to deliver satisfactory performance due to challenges such as imbalanced exposure and a lack of labeled data. To address these challenges, a new unsupervised nighttime image semantic segmentation network called AI-USeg is proposed. First, a lightweight Self-Calibrating Illumination (SCI) network is used to enhance nighttime images, thereby mitigating the impact of lighting variations on subsequent semantic segmentation networks. Next, a Domain Adaptation (DA) method is introduced to transition the model from Cityscapes containing a large amount of labeled data to Dark Zurich-D, addressing the lack of labeled data. Subsequently, AI-USeg introduces a Squeeze-and-Excitation Network (SENet) into the discriminator, built upon a Fully Convolutional Network (FCN). This adaptation facilitates the adjustment of image features in low-light nighttime settings through adversarial learning in the output space, ultimately improving the performance of semantic segmentation tasks for nighttime images. The experiment used two sets of 2 416 day and night image pairs sourced from Cityscapes and Dark Zurich-train for unsupervised training. The results show that AI-USeg achieved Mean Intersection over Union (mIoU) values of 47.9% and 51.5% on the Dark Zurich-test and Nighttime Driving-test datasets, respectively. These values were 5.4 and 2.1 percentage points higher than those obtained using the MGCDA method. In conclusion, AI-USeg displayed stronger adaptability to nighttime image features and higher robustness than previous segmentation models, providing an effective solution for image segmentation tasks in nighttime scenes.

Key words: deep learning, semantic segmentation, autonomous driving, low-light image enhancement, attention mechanism