Question

All or nearly all of the papers using dropout are using it for supervised learning. It seems that it could just as easily be used to regularize deep autoencoders, RBMs and DBNs. So why isn't dropout used in unsupervised learning?

Was it helpful?

Solution

Dropout is used in unsupervised learning. For example:

Shuangfei Zhai, Zhongfei Zhang: Dropout Training of Matrix Factorization and Autoencoder for Link Prediction in Sparse Graphs (arxiv, 14 Dec 2015)

OTHER TIPS

Labeled data is relatively scarce, and that's why supervised learning often benefits from strong regularization, like DropOut.

On the other hand, unlabeled data is usually plentiful, and that's why DropOut is typically not needed, and may be detrimental (as it reduces the model capacity).

Even gigantic models like GPT-3 (175e9 parameters) are still underfitting after being updated on 300e9 tokens.

Licensed under: CC-BY-SA with attribution
Not affiliated with StackOverflow
scroll top