TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Distill: a modern machine learning journal

930 pointsby jasikparkabout 8 years ago

30 comments

j2kunabout 8 years ago
I sure hope this catches on, but we should all be aware of the hurdles:<p>- Little incentive for researchers to do this beyond their own good will.<p>- Most ML researchers are bad writers, and it&#x27;s unlikely that the editing team will do the work needed (which is often a larger reorganization of a paper and ideas) to improve clarity.<p>- Producing great writing and clear, interactive figures, and managing an ongoing github repo require nontrivial amounts of extra time, and researchers already have strained time budgets.<p>- It requires you to learn git, front-end web design, random javascript libraries (I for one think d3 is a nuisance), exacerbating the time suck on tangents to research.<p>Maybe you could convince researchers to contribute with prizes that aligned with their university&#x27;s goals. Just spitballing here, but maybe for each &quot;top paper&quot; award, get a team together to further clarify the ideas for a public audience, collaborate with the university and their department and some pop-science writers, and get some serious publicity beyond academic circles. If that doesn&#x27;t convince a university administration that the work is worth the lower publication count, what will?<p>In the worst case it&#x27;ll be the miserable graduate students&#x27; jobs to implement all these publication efforts, and they won&#x27;t be able to spend time learning how to do research.
评论 #13916814 未加载
评论 #13919477 未加载
评论 #13917315 未加载
评论 #13916825 未加载
评论 #13917335 未加载
评论 #13918613 未加载
评论 #13917116 未加载
评论 #13916923 未加载
评论 #13916970 未加载
评论 #13918140 未加载
评论 #13920917 未加载
colah3about 8 years ago
Various announcements:<p>Google Research: <a href="https:&#x2F;&#x2F;research.googleblog.com&#x2F;2017&#x2F;03&#x2F;distill-supporting-clarity-in-machine.html" rel="nofollow">https:&#x2F;&#x2F;research.googleblog.com&#x2F;2017&#x2F;03&#x2F;distill-supporting-c...</a><p>DeepMind: <a href="https:&#x2F;&#x2F;deepmind.com&#x2F;blog&#x2F;distill-communicating-science-machine-learning&#x2F;" rel="nofollow">https:&#x2F;&#x2F;deepmind.com&#x2F;blog&#x2F;distill-communicating-science-mach...</a><p>OpenAI: <a href="https:&#x2F;&#x2F;openai.com&#x2F;blog&#x2F;Distill&#x2F;" rel="nofollow">https:&#x2F;&#x2F;openai.com&#x2F;blog&#x2F;Distill&#x2F;</a><p>YC Research: <a href="http:&#x2F;&#x2F;blog.ycombinator.com&#x2F;distill-an-interactive-visual-journal-for-machine-learning-research&#x2F;" rel="nofollow">http:&#x2F;&#x2F;blog.ycombinator.com&#x2F;distill-an-interactive-visual-jo...</a><p>Chris Olah: <a href="http:&#x2F;&#x2F;colah.github.io&#x2F;posts&#x2F;2017-03-Distill&#x2F;" rel="nofollow">http:&#x2F;&#x2F;colah.github.io&#x2F;posts&#x2F;2017-03-Distill&#x2F;</a>
评论 #13916111 未加载
评论 #13918359 未加载
评论 #13916307 未加载
评论 #13917345 未加载
评论 #13916713 未加载
choxiabout 8 years ago
I&#x27;ve been trying to read more primary source information, sort of as my own way of combatting &quot;fake news&quot; but before that term was coined. There&#x27;s a learning curve to it, but I&#x27;ve found that reading S1 filings and Quarterly Earnings Reports can be more enlightening than reading a news article on any given company. Likewise, reading research papers on biology and deep learning is <i>significantly</i> more valuable than reading articles or educational content on those topics.<p>As you&#x27;d imagine though, it&#x27;s really hard. Reading a two page research paper is a very different experience from reading a NYTimes or WSJ article. The information density is enormous, the vocabulary is very domain specific, and it can take days or weeks of re-reading and looking up terms to finally understand a paper.<p>I&#x27;m really excited about Distill, there&#x27;s a lot of value in making research papers more accessible and interesting. I&#x27;ve noticed that the ML&#x2F;AI field has been very pioneering about research publication process, some papers are now published with source code on GitHub and the authors answering questions on r&#x2F;machinelearning. This seems like a really great next step, I hope other fields of science will break away from traditional journals and do the same.
TuringNYCabout 8 years ago
I don&#x27;t want to undermine visualizations, they are awesome, but one of the big problems I see with ML research is the lack of re-produceability. I know that Google, Facebook and some others already share associated source repos, but it should almost be mandatory when working with public benchmark datasets. Source + Docker Images would be even better.<p>I worked in clinical research in a past life and studies would be highly discounted if they couldn&#x27;t be reproduced. A highly detailed methods section was key. Many ML papers I see tend to have incredibly formalized LaTeX+Greek obsessed methods section, but far short of anything to allow reproduction. Some ML papers, <i>i swear</i> must have run their parameter searches a 1000 times to overfit and magically achieve 99% AUC.<p>Worse, I actually have tons of spare GPU farm capacity i&#x27;d love to devote to re-producing research, tweaking, trying it on adjacent datasets, etc. But the effort to re-produce is too high for most papers.<p>It is also disappointing to see various input datasets strewn about individuals&#x27; personal homepages, and sometimes end up broken. Sometimes the &quot;original&quot; dataset is in a pickled form after having already gone through multiple upstream transformations. I hope Distill can <i>instill</i> some good best practices to the community.
评论 #13916903 未加载
minimaxirabout 8 years ago
The announcements and About page indicate an emphasis on visuals and presentation, which I apprI&#x27;ve. But when I think of &quot;modern machine learning,&quot; I think of open-source and reproducibility (e.g. Jupyter notebooks).<p>Will the papers published on Distill maintain transparency of the statistical process?<p>I see in the submission notes that articles are required to be a public GitHub repo, which is a positive indicator. Although the actual code itself does not seem to be a requirement.
评论 #13916943 未加载
Xeoncrossabout 8 years ago
As a developer with a weaker background in mathematics, I face a language barrier with many modern algorithms. After lots of research I can understand and explain them in code, but I have no idea what your artistic-looking MathXML means.<p>Visualizations or algorithms described using code are much, much easier for me to understand and serve as a great starting point for unpacking the math explanations.
评论 #13916912 未加载
blinryabout 8 years ago
Shameless self-plug: If you like interactive explanations, check out <a href="http:&#x2F;&#x2F;explorableexplanations.com&#x2F;" rel="nofollow">http:&#x2F;&#x2F;explorableexplanations.com&#x2F;</a> and the explorables subreddit: <a href="https:&#x2F;&#x2F;www.reddit.com&#x2F;r&#x2F;explorables&#x2F;" rel="nofollow">https:&#x2F;&#x2F;www.reddit.com&#x2F;r&#x2F;explorables&#x2F;</a>
cingabout 8 years ago
Is there any concern about a web-native journal being less &quot;future-proof&quot;? I&#x27;ve come across quite a few interactive learning demonstrations in Flash&#x2F;Java that no longer work.
评论 #13916754 未加载
评论 #13918139 未加载
dangabout 8 years ago
YC Research&#x27;s (and longtime HNer!) michael_nielsen wrote an announcement here: <a href="http:&#x2F;&#x2F;blog.ycombinator.com&#x2F;distill-an-interactive-visual-journal-for-machine-learning-research&#x2F;" rel="nofollow">http:&#x2F;&#x2F;blog.ycombinator.com&#x2F;distill-an-interactive-visual-jo...</a>. Hopefully he&#x27;ll participate in the discussion too.
rememberlennyabout 8 years ago
I wish there was a way to subscribe to a weekly email related to this.
评论 #13916346 未加载
sytelusabout 8 years ago
This is great but it would have been even better if Distill was designed to play well with the current system. Vast majority of researchers are focused on publishing at various conferences with strict deadlines. Even if they had all the skillsets and time to produce these beautiful illustrations, I highly doubt this will change.<p>Also, it is very likely that veterans in the field might think of this format as too verbose and too sugar coated, more appropriate for less math-savvy users and therefore not mainstream. Furthermore, I really feel TeX is irreplaceable unless you got all of its feature covered. All of the historic effort to replace TeX - even with bells and whistles of WYSIWYG editors - in research has failed and its important to learn from those failures. You will be surprised how many researchers insist on printing out the paper for reading even when they have access to tablets and PC.<p>Instead of being another peer reviewed journal, Distill could act as the following:<p>- platform to publish supplemental material and code<p>- platform to manage communication&#x2F;issues post publication<p>- platform for readers to invite other readers for peer review and generate &quot;front page&quot; based on some sort of reviewer trust relationship.<p>- platform to host Python and MatLab code with web frontends without researchers having to learn new developer skills<p>- support pdf submissions but without all the eliteness of arxiv and using algorithms to create the &quot;front page&quot; based on some sort of peer reviewer rankings.<p>Above features are indeed sorely missing and Distill has good opportunity to become an &quot;add-on&quot; to current academic publishing systems as opposed to another peer reviewed journal.
transcranialabout 8 years ago
This is really exciting! Chris et al: have you guys seen Keras.js (<a href="https:&#x2F;&#x2F;github.com&#x2F;transcranial&#x2F;keras-js" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;transcranial&#x2F;keras-js</a>)? It could probably be useful for certain interactive visualizations or papers.
fnlabout 8 years ago
How does this provide IF ratings? Probably irrelevant for industry, but publishing in academia is all about IF, no matter how bad and corrupt one might think it is.<p>And what about long-term stability&#x2F;presence. Most top journals and their publishing houses (NPG, Elsevier, Springer) are likely to hang around for another decade (or two...), while I don&#x27;t feel so sure about that for a product like GitHub. Maybe Distill is&#x2F;will be officially backed (financially) by the industry names supporting it?<p>That being said, I&#x27;d love seeing this succeed, but there seems much to be done to get this really &quot;off the ground&quot; beyond being a (much?!) nicer GitXiv.
评论 #13919340 未加载
radarsat1about 8 years ago
While this is very nice, I&#x27;m a bit confused about the target. What kind of material is intended to be published here in the future?<p>Because the blog post and title seems to be describing it as a &quot;journal&quot; intended to replace PDF publications, but the actual content appears to be more in the tutorial&#x2F;survey category, e.g. &quot;how to use t-SNE,&quot; etc. Is this intended to be a place to publish <i>new</i> research in the future, or is it meant more for enhanced &quot;medium&quot;-style blog posts?<p>Both are fine, I just find the dissonance between the announcement and the actual content a bit confusing.
chairmanwowabout 8 years ago
I feel like science publication in general could benefit from disruption of the publishing model. I&#x27;m not sure that the toolkit that Distill has provided is quite enough to totally change the paradigm, and it currently restricted to only one field.<p>I like the idea of having research being approachable for the non-scientist, and the more important question of whether there is a more efficient form (in terms of communicating new science between scientists) for research papers to take.<p>Is there any relevant work along this vector of thought that I should check out? Because I would really love to do some work on this.
评论 #13921298 未加载
ycHammerabout 8 years ago
Would saving jupyter notebooks as .html work? PS: I have published in all of top-4 tier ML conferences but s<i></i>k at html&#x2F;css&#x2F;js. What is my pathway to distill now? I, like every other researcher worth her&#x2F;his name in salt is always running behind clock when it comes to deadlines and lit to review. So, yeah? Coaxing myself into investing time for css&#x2F;html&#x2F;js in lieu of picking up more math tools seems criminal to me. Am I alone in this ?
mysoreabout 8 years ago
Wow this comes with great timing!<p>I am a UI-developer who has been wanting to learn ML forever. I started working on<p>1. fast.ai 2. think bayes 3. UW data science @ scale w&#x2F; coursera 4. udacity car nano degree<p>I&#x27;m going to write some articles about what I learn and hopefully move into the ML field as a data engineer in 6 months. I figure I got into my current job with a visual portfolio of nicely designed css&#x2F;js demos, maybe the same thing will work for AI.
Old_Thrashbargabout 8 years ago
I don&#x27;t see it written explicitly; can anyone confirm that this journal is fully open-access?
评论 #13916419 未加载
评论 #13916506 未加载
JorgeGTabout 8 years ago
You should definitely assign a DOI to each article.
评论 #13917591 未加载
评论 #13917275 未加载
EternalDataabout 8 years ago
Looks very good (especially the team behind it!), but I wonder if there&#x27;s a discrete step down to where you make machine learning materials accessible to the general public beyond data visualizations and clear writing. This will certainly be a more interactive experience, but it seems to cater to those who are &quot;in-the-know&quot; and require a bit more interactivity&#x2F;clarity. It&#x27;d be nice to discuss the format changes or the &quot;TLDR&quot; bot of machine learning that makes machine learning research truly accessible to the general public.
fwxabout 8 years ago
This is amazing! My burning question - as has been pointed out in the thread, the effort to produce a great article on Distill - generating interactive figures, doing front end web dev etc. would require a lot of time and resources on the part of the researchers. Is it possible to include within Distill an option to connect researchers to willing-and-able developers in those domains (for example, me) to help them get it done?
aabajianabout 8 years ago
I already have a nomination. The guy who wrote this blog post:<p><a href="http:&#x2F;&#x2F;adilmoujahid.com&#x2F;posts&#x2F;2016&#x2F;06&#x2F;introduction-deep-learning-python-caffe&#x2F;" rel="nofollow">http:&#x2F;&#x2F;adilmoujahid.com&#x2F;posts&#x2F;2016&#x2F;06&#x2F;introduction-deep-lear...</a><p>It&#x27;s the only way I could get a working model of Caffe while understanding the data preparation steps. I&#x27;ve already retrofitted it to classify tumors.
taliesinbabout 8 years ago
Great stuff! I&#x27;m a fan of what&#x27;s gone up on distill so far. Question for colah and co if they&#x27;re still around: When does the first issue of the journal come out (edit: looks like individual articles just get published when they get published, n&#x2F;m). Also, that &quot;before&#x2F;after&quot; visualization of the gradient descent convergence is intriguing -- where&#x27;s it from?
评论 #13919015 未加载
blunteabout 8 years ago
I don&#x27;t know jack about machine learning, but these illustrations are gorgeous - simple, elegant, and aesthetically very pleasing.
wodenokotoabout 8 years ago
Looking at the how-to section[1] for creating distil articles, I fail to find how to write math and some notes on how best to reference sections of the document.<p>Other than that, this looks, much, much easier to write than LaTex.<p>[1] <a href="http:&#x2F;&#x2F;distill.pub&#x2F;guide&#x2F;" rel="nofollow">http:&#x2F;&#x2F;distill.pub&#x2F;guide&#x2F;</a>
djabattabout 8 years ago
It would be cool to see greater diversity of thinking on the about page. perhaps the pub is designed for insiders.<p>Having more research transparency is great for community of likes minds to learn from. A suggested addition is an section and team to lead a discussion ML ethics.
good_vibesabout 8 years ago
I will definitely submit my first paper to Distill. It draws upon a few different fields but the foundation is definitely machine learning.<p>What a time to be alive!
mastaziabout 8 years ago
r&#x2F;MachineLearning discussion:<p><a href="https:&#x2F;&#x2F;www.reddit.com&#x2F;r&#x2F;MachineLearning&#x2F;comments&#x2F;60hy0t&#x2F;the_journal_distill_launches_today_in_a_nutshell&#x2F;" rel="nofollow">https:&#x2F;&#x2F;www.reddit.com&#x2F;r&#x2F;MachineLearning&#x2F;comments&#x2F;60hy0t&#x2F;the...</a>
ycHammerabout 8 years ago
Anyone here has any idea if Jupyter notebook -&gt; save as .html would do the trick?
skynodeabout 8 years ago
Hopefully this won&#x27;t be another ResearchGate dressed in open source clothing.