TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Transfer Learning and Fine-Tuning Deep Convolutional Neural Networks

88 点作者 rasmi超过 8 年前

3 条评论

eveningcoffee超过 8 年前
What I have not seen is the explanation about transferring the data normalization parameters. Say you apply the contrast normalization to the images you use to train the first network.<p>Is there anything better you can do than applying the same parameters to the second training set?
评论 #13244436 未加载
iraphael超过 8 年前
This is the interesting part:<p>&gt; New dataset is smaller in size and similar in content compared to original dataset: If the data is small, it is not a good idea to fine-tune the DCNN due to overfitting concerns. Since the data is similar to the original data, we expect higher-level features in the DCNN to be relevant to this dataset as well. Hence, the best idea might be to train a linear classifier on the CNN-features.<p>&gt; New dataset is relatively large in size and similar in content compared to the original dataset: Since we have more data, we can have more confidence that we would not over fit if we were to try to fine-tune through the full network.<p>&gt; New dataset is smaller in size but very different in content compared to the original dataset: Since the data is small, it is likely best to only train a linear classifier. Since the dataset is very different, it might not be best to train the classifier from the top of the network, which contains more dataset-specific features. Instead, it might work better to train a classifier from activations somewhere earlier in the network.<p>&gt; New dataset is relatively large in size and very different in content compared to the original dataset: Since the dataset is very large, we may expect that we can afford to train a DCNN from scratch. However, in practice it is very often still beneficial to initialize with weights from a pre-trained model. In this case, we would have enough data and confidence to fine-tune through the entire network.
rasmi超过 8 年前
This is one of the most concise and accessible explanations of fine-tuning CNNs that I&#x27;ve come across. I hope someone finds it as helpful as I did.