Collecting aerial data from physical world is usually time-consuming. Image simulation is a significant data source for various ground target detection systems. Unfortunately, the reality gap between simulated and real data makes the model trained on simulated image not workable on real image. A translation method is proposed for tackling the simulation-toreality problem in this paper. First, image simulation system is employed for data preparation. Then, the simulated data is converted into a more similar one to the real image. The segmentation map is the bridge between simulated and real data. At last, the target detection model is used as the utility evaluation method for the simulated data. The simulated and synthesized data is used to train a vehicle detection model. Experiments show that results trained by synthesized data are really close to the real results. The proposed translation method can assist real image for target detection task, which is an effective data augmentation method for aerial data.
When the real infrared image is insufficient, the simulation infrared image is an important data supplement to the real infrared image. However, the authenticity of simulated infrared image often does not meet the requirements of real images. So improving the authenticity of simulated infrared image plays an important role in related fields. In order to achieve this goal, a method based on deep learning is proposed in this paper. Unlike traditional methods of using manual modification by experience, the proposed method can convert non-realistic simulation infrared image input into a realistic one with similar scene structure. First, we generate a large number of simulation infrared images through the simulation system. Then, we propose an optimization method to improve the authenticity of simulated infrared images. Finally, we designed a comparison experiment between the original simulation infrared image and the optimized simulation infrared image, and finally verify the effectiveness.
Image segmentation has always been a key research issue in the field of computer vision. Image segmentation networks that use deep learning methods require a large number of finely labeled samples, which is difficult to obtain. In this paper, we combine the focal loss function with the fully convolutional networks to improve network performance. And we collected and built a dataset contents 1500 samples with complex background. We trained the improved network with the dataset to achieve 81.55% in mean average precision and 76.13% in mean intersection over union.