IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing (Jan 2024)
Improved Few-Shot SAR Image Generation by Enhancing Diversity
Abstract
Due to their remarkable capabilities of generation, deep-learning-based (DL) generative models have been widely applied in the field of synthetic aperture radar (SAR) image synthesis. This kind of data-driven DL methods usually requires abundant training samples to guarantee the performance. However, the number of SAR images for training is often insufficient because of expensive acquisitions. This typical few-shot image generation (FSIG) task still remains not fully investigated. In this article, we propose an optical-to-SAR (O2S) image translation model with a pairwise distance (PD) loss to enhance the diversity of generation. First, we replace the semantic maps used as the input of network in previous studies with more easily available optical images and apply the popular pix2pix model in image-to-image translation tasks as the foundation network. Second, inspired by the FSIG works in the traditional computer vision field, we propose a similarity preservation term in the loss function, which encourages the generated images to inherit the similarity relationship of the corresponding simulated SAR images. Third, the data augmentation experiments on the MSTAR dataset indicates the effectiveness of our model. With only five samples for each target category and six categories in total, the basic O2S network boosts the classification accuracy by 4.81% and 2.27% for data of depression angle of 15° and 17°, respectively. The PD loss is capable of bringing additional 2.23% and 1.78% improvement. The investigation on similarity curves also suggests that the generated images enhanced by the PD loss have closer similarity behaviors to the real SAR images.
Keywords