GENERALIZABLE EMBEDDINGS WITH CROSS-BATCH METRIC LEARNING


GÜRBÜZ Y. Z., ALATAN A. A.

IEEE International Conference on Image Processing, Kuala-Lumpur, Malezya, 08 Ekim 2023 identifier identifier

  • Yayın Türü: Bildiri / Tam Metin Bildiri
  • Doi Numarası: 10.1109/icip49359.2023.10222102
  • Basıldığı Şehir: Kuala-Lumpur
  • Basıldığı Ülke: Malezya
  • Orta Doğu Teknik Üniversitesi Adresli: Evet

Özet

Global average pooling (GAP) is a popular component in deep metric learning (DML) for aggregating features. Its effectiveness is often attributed to treating each feature vector as a distinct semantic entity and GAP as a combination of them. Albeit substantiated, such an explanation's algorithmic implications to learn generalizable entities to represent unseen classes, a crucial DML goal, remain unclear. To address this, we formulate GAP as a convex combination of learnable prototypes. We then show that the prototype learning can be expressed as a recursive process fitting a linear predictor to a batch of samples. Building on that perspective, we consider two batches of disjoint classes at each iteration and regularize the learning by expressing the samples of a batch with the prototypes that are fitted to the other batch. We validate our approach on 4 popular DML benchmarks.