TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

A Primer in BERTology: What We Know About How Bert Works

81 点作者 whym超过 4 年前

4 条评论

mobilio超过 4 年前
&quot; BERT is a method of pre-training language representations, meaning that we train a general-purpose &quot;language understanding&quot; model on a large text corpus (like Wikipedia), and then use that model for downstream NLP tasks that we care about (like question answering). BERT outperforms previous methods because it is the first unsupervised, deeply bidirectional system for pre-training NLP.&quot;<p><a href="https:&#x2F;&#x2F;github.com&#x2F;google-research&#x2F;bert" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;google-research&#x2F;bert</a>
niea_11超过 4 年前
Can anyone please explain (in layman terms if it&#x27;s possible) how did the researchers come up with the method in the first place if the process how the method finds the answers is not understood?
评论 #25045683 未加载
评论 #25045603 未加载
评论 #25045540 未加载
评论 #25045617 未加载
评论 #25045802 未加载
评论 #25050223 未加载
taneq超过 4 年前
For anyone who, like me, isn&#x27;t a BERTologist, BERT is a neural network architecture.
评论 #25044045 未加载
hallqv超过 4 年前
Any new information in the paper since the first version came out in Mars? Otherwise a 6 month old meta-study seems kind of dated given rate of progress in NLP atm.