Paper Title
Cross-Potent Visuval Data Matching by Nuterilized Sacular Model Analysis

Abstract
The fundamental problems in many real-world in vision tasks is cross-potent visual data matching, e.g., matching persons across ID photos and surveillance videos. The basic approaches to this problem naturally involve two steps: i) Experiencing samples from different domains into a natural space, and ii) computing (dis-)sacular in this space based on a certain distance. In this paper, we present a Dogma pair-wise sacular measure that advances existing models by i) Enlarging traditional linear projections into affine conversions i.e transformations and ii) fusing affine Maha-lanobis distance and Cosine sacular by a data-driven combination. We unify our sacular measure with feature representation convolution via deep convolution neural networks. Specifically, we in-corporate the sacular measure matrix into the deep architecture, opening an end-to-end way of model optimization. We extensively evaluate our Neutralisation sacular model in several challenging cross-potent matching tasks: person re-checking under different views and face verification over different modalities (i.e., faces from moving images and videos, senior citizens and younger faces, and sketch and photo portraits). The exploration results demonstrate accurate performance of our model over other state-of-the-art methods. Index Terms - Sacular model, Cross-potent matching, Person Identifiable, Deep convolution.