In this course, you will learn about several algorithms that can learn near optimal policies based on trial and error interaction with the environment---learning from the agent’s own experience. Learning from actual experience is striking because it requires no prior knowledge of the environment’s dynamics, yet can still attain optimal behavior. We will cover intuitively simple but powerful Monte Carlo methods, and temporal difference learning methods including Q-learning. We will wrap up this course investigating how we can get the best of both worlds: algorithms that can combine model-based planning (similar to dynamic programming) and temporal difference updates to radically accelerate learning.Lee mas.
Este recurso es ofrecido por un socio afiliado. Si paga por la capacitación, podemos ganar una comisión para respaldar este sitio.
Las técnicas y herramientas cubiertas en Sample-based Learning Methods son muy similares a los requisitos que se encuentran en los anuncios de trabajo de Científico de datos.