Limited Additional Knowledge from Naive Self-Training with Pseudo Labels
Ascertain whether naive self-teaching for monocular depth estimation—implemented by directly combining labeled images and pseudo-labeled unlabeled images for joint training—yields only limited additional visual knowledge when sufficient labeled data and strong pre-trained encoders are available, thereby failing to improve over training solely on labeled images.
References
In our preliminary attempts, directly combining labeled and pseudo labeled images failed to improve the baseline of solely using labeled images. We conjecture that, the additional knowledge acquired in such a naive self-teaching manner is rather limited.
— Depth Anything: Unleashing the Power of Large-Scale Unlabeled Data
(2401.10891 - Yang et al., 19 Jan 2024) in Section 1 (Introduction)