GENERALIZABLE EMBEDDINGS WITH CROSS-BATCH METRIC LEARNING

2023-10-08
GÜRBÜZ, YETİ ZİYA
Alatan, Abdullah Aydın
Global average pooling (GAP) is a popular component in deep metric learning (DML) for aggregating features. Its effectiveness is often attributed to treating each feature vector as a distinct semantic entity and GAP as a combination of them. Albeit substantiated, such an explanation's algorithmic implications to learn generalizable entities to represent unseen classes, a crucial DML goal, remain unclear. To address this, we formulate GAP as a convex combination of learnable prototypes. We then show that the prototype learning can be expressed as a recursive process fitting a linear predictor to a batch of samples. Building on that perspective, we consider two batches of disjoint classes at each iteration and regularize the learning by expressing the samples of a batch with the prototypes that are fitted to the other batch. We validate our approach on 4 popular DML benchmarks.
IEEE International Conference on Image Processing
Citation Formats
Y. Z. GÜRBÜZ and A. A. Alatan, “GENERALIZABLE EMBEDDINGS WITH CROSS-BATCH METRIC LEARNING,” presented at the IEEE International Conference on Image Processing, Kuala-Lumpur, Malezya, 2023, Accessed: 00, 2024. [Online]. Available: https://ieeexplore.ieee.org/xpl/conhome/10221937/proceeding.