China • Affiliation: Aizawa Laboratory, D3 Graduate School of Information Science and Technology The University of Tokyo • Research interest: Open-set recognition Semi-supervised learning Domain adaptation • HP: yu1ut.com 2
result MSE loss back prop 0 1 0 0 back prop Groundtruth Label Train: Mean Teacher [Tarvainen+, NeurIPS 17] Cross-entropy loss labeled 0.1 0.2 0.6 0.1 0.2 0.1 0.7 0.0
• Training Unlabeled Sample prediction pseudo label one-hot CNN CNN CNN prediction prediction prediction back prop Cross-entropy loss same network same network
≠ The performance in real-world • vs self-supervised learning • When the dataset is large enough, fine-tuning self-supervised models is enough? • vs zero-shot prediction (e.g. CLIP) • Is CLIP enough for most datasets? 50
to train a model when only limited labeled data is available. • Recent methods can train a high performance model even when only 40 labeled samples are labeled. • Self-supervised learning and pretrained vision-and-language models are challenging semi-supervised learning. 51