简体   繁体   English

在深度学习中是否可以在训练集的子集上进行训练以找到最佳超参数?

[英]Is it possible in deep learning to train on a subset of training set in order to find the best hyper-parameters?

In classic machine learning, it is not uncommon to do a search for hyper-parameters by training different configurations on a small subset of training set.在经典机器学习中,通过在一小部分训练集上训练不同配置来搜索超参数的情况并不少见。 Usually, for each set of hyper-parameters, a k-fold cross validation is done over a small subset of training set.通常,对于每组超参数,都会对一小部分训练集进行 k 折交叉验证。 However, in deep learning, models are usually very hungry of data.然而,在深度学习中,模型通常非常需要数据。

So, my question is that do you think is it still possible to use the same strategy in deep learning?所以,我的问题是,您认为在深度学习中是否仍然可以使用相同的策略? What is your experience?你的经验是什么?

Yes, but as you noticed, the deep learning models usually work best with large samples.是的,但正如您所注意到的,深度学习模型通常最适合大样本。 So your subset would need to be large as well.所以你的子集也需要很大。 With insufficient data, the model would underperform and wouldn't help for hyperparameter tuning.如果数据不足,模型将表现不佳,并且无助于超参数调整。

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM