TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Introduction to Reinforcement Learning (2015)

243 点作者 arithma超过 4 年前

13 条评论

jointpdf超过 4 年前
This is (roughly) the same course but the lectures are from 2018. The audio&#x2F;video quality is quite a a bit better:<p><a href="https:&#x2F;&#x2F;youtube.com&#x2F;playlist?list=PLqYmG7hTraZBKeNJ-JE_eyJHZ7XgBoAyb" rel="nofollow">https:&#x2F;&#x2F;youtube.com&#x2F;playlist?list=PLqYmG7hTraZBKeNJ-JE_eyJHZ...</a>
hfkldjsjfkdj超过 4 年前
When it comes to education online I prefer these, where they publish actual university course lectures and assignments than what you can find on Coursera and alike. There is still a big gap in the quality and depth.
jsemrau超过 4 年前
That&#x27;s one of these moments in life where you see tech and you know it will change the world, but don&#x27;t see the problem yet.
评论 #25551251 未加载
评论 #25550254 未加载
ilaksh超过 4 年前
I used to be a bit more excited about RL. I mean, it&#x27;s still definitely something I have to learn, but one aspect of it _seems_ lacking to me and is messing with my motivation to learn it. I&#x27;m sure someone will happily explain all the ways I am ignorant.<p>It seems like there is a lot of emphasis on &quot;direct RL&quot; or whatever where they don&#x27;t even really think about the model much, but it&#x27;s I guess often inside of the policy or something?<p>But it seems to me as someone who has just started learning about robotics, that I absolutely need to first verify that I have an accurate model of the environment which I can inspect. It seems like a lot of RL approaches might not even be able to supply that.<p>I mean what I am stuck on as far as creating a robot (or virtual robot) is having a vision system that does all of the hard things I want. I feel like if I can detect edges and surfaces and shapes in 3D, parts of objects and objects, with orientation etc., and in a way I can display and manipulate it, that level of understanding will give me a firm base to build the rest of learning and planning on.<p>I know all of that is very hard. It seems like they must have tried that for awhile and then kind of gave up to head down the current direction of RL? Or just decided it wasn&#x27;t important. I still think it&#x27;s important.
评论 #25552468 未加载
评论 #25552859 未加载
评论 #25552814 未加载
in3d超过 4 年前
I watched this course and David Silver is a great lecturer, better than anybody else I’ve seen actually. I hope he does more publicly viewable courses in the future.
hideo7746超过 4 年前
On one hand you&#x27;re right, methods like Q-learning are model-free and do not necessarily encode much about state dynamics. The Q-function is a feature (function) of state and while ita may not say much about the model, it does encode the most important aspect of the model in terms of solving the task. Namely, it predicts the accumulated reward conditional on next actions actions. That makes it a somewhat narrow representation of state on its own. But, if you consider an environment that has many reward signals, and you learn Q functions for each, this ensemble of Q functions can consitute a rich representation of state. Depending on what the reward functions are, the associated Q functions may be sufficient to construct a full model. so I guess my point is that the learned quantities in RL encode key aspects of state, and when you expand beyond the single task&#x2F;single reward RL setting the lines between value and model can become blurred.
dgb23超过 4 年前
Asking from a layman&#x27;s perspective:<p>I&#x27;ve read a bit about genetic algorithms or evolutionary computation at some point. Apparently it achieves good results as it can find discrete solutions for complex, well defined problems.<p>Reinforcement learning is something I know even less about. But from what I gathered it is also most successful in well defined problems and systems (such as games).<p>So my question is: How do they relate? Is there overlap and what are the most significant conceptual differences?
评论 #25552183 未加载
评论 #25551989 未加载
评论 #25561449 未加载
luplex超过 4 年前
I&#x27;m taking an adaptation of this class. My professor is simply reusing Silver&#x27;s slides, so I&#x27;m watching the original lecture instead. Highly recommend!
visarga超过 4 年前
Watched the course and it&#x27;s great, probably the best intro to RL. Multiple watches needed as the subject is very deep.
foobaw超过 4 年前
David Silver is awesome! Loved him in the Alpha Go documentary as well
评论 #25550294 未加载
spicyramen超过 4 年前
Just watched his interview with Lex, inspiring engineer
captn3m0超过 4 年前
I did the course during my RC batch this year. Highly recommended if you&#x27;re looking to learn RL.
评论 #25551024 未加载
platz超过 4 年前
Hrm so are markov decision processes RL?
评论 #25551919 未加载