Question

All or nearly all of the papers using dropout are using it for supervised learning. It seems that it could just as easily be used to regularize deep autoencoders, RBMs and DBNs. So why isn't dropout used in unsupervised learning?

Était-ce utile?

La solution

Dropout is used in unsupervised learning. For example:

Shuangfei Zhai, Zhongfei Zhang: Dropout Training of Matrix Factorization and Autoencoder for Link Prediction in Sparse Graphs (arxiv, 14 Dec 2015)

Autres conseils

Labeled data is relatively scarce, and that's why supervised learning often benefits from strong regularization, like DropOut.

On the other hand, unlabeled data is usually plentiful, and that's why DropOut is typically not needed, and may be detrimental (as it reduces the model capacity).

Even gigantic models like GPT-3 (175e9 parameters) are still underfitting after being updated on 300e9 tokens.

Licencié sous: CC-BY-SA avec attribution
Non affilié à StackOverflow
scroll top