Cross-modality person re-identification (Re-ID) between RGB and infrared domains is a hot and challenging problem, which aims to retrieve pedestrian images cross-modality and cross-camera views. Since there is a huge gap between two modalities, the difficulty of solving the problem is how to bridge the cross-modality gap with images. However, most approaches solve this issue mainly by increasing interclass discrepancy between features, and few research studies focus on decreasing intraclass cross-modality discrepancy, which is crucial for cross-modality Re-ID. Moreover, we find that despite the huge gap, the attribute representations of the pedestrian are generally unchanged. We provide a different view of the cross-modality person Re-ID problem, which uses additional attribute labels as auxiliary information to increase intraclass cross-modality similarity. First, we manually annotate attribute labels for a large-scale cross-modality Re-ID dataset. Second, we propose an end-to-end network to learn modality-invariant and identity-specific local features with the joint supervision of attribute classification loss and identity classification loss. The experimental results on a large-scale cross-modality Re-ID benchmarks show that our model achieves competitive Re-ID performance compared with the state-of-the-art methods. To demonstrate the versatility of the model, we report the results of our model on the Market-1501 dataset.
Example-based face sketch synthesis technology generally requires face photo-sketch images with face alignment and size normalize. To break through the limitation, we propose a global face sketch synthesis method: In training, all training photo-sketch patch pairs are collected together and a photo feature dictionary is learned from the photo patches. For each atom of the dictionary, its K closest photo-sketch patch pairs are clustered, namely “Anchored Neighborhood”. In testing, for each test photo patch, we search its nearest photo patch in the Anchored Neighborhood determined by its closest atom, then the corresponding sketch patch is the output. By the same way, we train and test in the high-frequency domain and synthesis the high-frequency results. Finally, the fusion of the initial and the high-frequency results is the final sketch. The experiments on three public face sketch datasets and various real-world photos demonstrate the effectiveness and robustness of the proposed method