Skip to content

Introduction

For a variety of recognition applications, the most accurate results are now obtained by adapting large foundation models pre-trained on massive curated or raw data [1]

CLIP as a foundation model could power inumerous ethnicity classification solutions, and as such it is imperial that we explored inherent bias in its zero-shot performance.

References

  1. Visual Prompt Tuning. Jia et al. 2022