fileCoursera-Practical-Reinf-1LRJt

Coursera Practical Reinforcement Learning
  • MP4001.Welcome\\/001. Why should you care.mp432.42MB
  • MP4001.Welcome\\/002. Reinforcement learning vs all.mp410.80MB
  • MP4002.Reinforcement Learning\\/003. Multi-armed bandit.mp417.88MB
  • MP4002.Reinforcement Learning\\/004. Decision process & applications.mp423.01MB
  • MP4003.Black box optimization\\/005. rkov Decision Process.mp418.00MB
  • MP4003.Black box optimization\\/006. Crossentropy method.mp436.01MB
  • MP4003.Black box optimization\\/007. Approxite crossentropy method.mp419.27MB
  • MP4003.Black box optimization\\/008. More on approxite crossentropy method.mp422.89MB
  • MP4004.All the cool stuff that isn\t in the ba<x>se track\\/009. Evolution strategies core idea.mp420.86MB
  • MP4004.All the cool stuff that isn\t in the ba<x>se track\\/010. Evolution strategies th problems.mp417.73MB
  • MP4004.All the cool stuff that isn\t in the ba<x>se track\\/011. Evolution strategies log-derivative trick.mp427.84MB
  • MP4004.All the cool stuff that isn\t in the ba<x>se track\\/012. Evolution strategies duct tape.mp421.17MB
  • MP4004.All the cool stuff that isn\t in the ba<x>se track\\/013. Blackbox optimization drawbacks.mp415.21MB
  • MP4005.Striving for reward\\/014. Reward design.mp449.70MB
  • MP4006.Belln equations\\/015. State and Action Value Functions.mp437.31MB
  • MP4006.Belln equations\\/016. Measuring Policy Optimality.mp418.08MB
  • MP4007.Generalized Policy Iteration\\/017. Policy evaluation & improvement.mp431.92MB
  • MP4007.Generalized Policy Iteration\\/018. Policy and value iteration.mp424.16MB
  • MP4008.Model-free learning\\/019. Model-ba<x>sed vs model-free.mp428.78MB
  • MP4008.Model-free learning\\/020. Monte-Carlo & Temporal Difference; Q-learning.mp430.11MB
  • MP4008.Model-free learning\\/021. Exploration vs Exploitation.mp428.23MB
  • MP4008.Model-free learning\\/022. Footnote Monte-Carlo vs Temporal Difference.mp410.30MB
  • MP4009.On-policy vs off-policy\\/023. Accounting for exploration. Expected Value SARSA..mp437.73MB
  • MP4010.Experience Replay\\/024. On-policy vs off-policy; Experience replay.mp426.72MB
  • MP4011.Limitations of Tabular Methods\\/025. Supervised & Reinforcement Learning.mp450.61MB
  • MP4011.Limitations of Tabular Methods\\/026. Loss functions in value ba<x>sed RL.mp433.76MB
  • MP4011.Limitations of Tabular Methods\\/027. Difficulties with Approxite Methods.mp447.03MB
  • MP4012.Case Study Deep Q-Network\\/028. DQN bird\s eye view.mp427.76MB
  • MP4012.Case Study Deep Q-Network\\/029. DQN the internals.mp429.63MB
  • MP4013.Honor\\/030. DQN statistical issues.mp419.22MB
  • MP4013.Honor\\/031. Double Q-learning.mp420.46MB
  • MP4013.Honor\\/032. More DQN tricks.mp433.94MB
  • MP4013.Honor\\/033. Partial observability.mp457.23MB
  • MP4014.Policy-ba<x>sed RL vs Value-ba<x>sed RL\\/034. Intuition.mp434.87MB
  • MP4014.Policy-ba<x>sed RL vs Value-ba<x>sed RL\\/035. All Kinds of Policies.mp416.05MB
  • MP4014.Policy-ba<x>sed RL vs Value-ba<x>sed RL\\/036. Policy gradient forli.mp431.56MB
  • MP4014.Policy-ba<x>sed RL vs Value-ba<x>sed RL\\/037. The log-derivative trick.mp413.29MB
  • MP4015.REINFORCE\\/038. REINFORCE.mp431.42MB
  • MP4016.Actor-critic\\/039. Advantage actor-critic.mp424.63MB
  • MP4016.Actor-critic\\/040. Duct tape zone.mp417.53MB
  • MP4016.Actor-critic\\/041. Policy-ba<x>sed vs Value-ba<x>sed.mp416.79MB
  • MP4016.Actor-critic\\/042. Case study A3C.mp426.09MB
  • MP4016.Actor-critic\\/043. A3C case study (2 2).mp414.96MB
  • MP4016.Actor-critic\\/044. Combining supervised & reinforcement learning.mp424.02MB
  • MP4017.Measuting exploration\\/045. Recap bandits.mp424.66MB
  • MP4017.Measuting exploration\\/046. Regret measuring the quality of exploration.mp421.27MB
  • MP4017.Measuting exploration\\/047. The message just repeats. \Regret Regret Regret.\.mp418.43MB
  • MP4018.Uncertainty-ba<x>sed exploration\\/048. Intuitive explanation.mp422.26MB
  • MP4018.Uncertainty-ba<x>sed exploration\\/049. Thompson Sampling.mp417.09MB
  • MP4018.Uncertainty-ba<x>sed exploration\\/050. Optimi in face of uncertainty.mp416.54MB
  • MP4018.Uncertainty-ba<x>sed exploration\\/051. UCB-1.mp422.19MB
  • MP4018.Uncertainty-ba<x>sed exploration\\/052. Bayesian UCB.mp440.80MB
  • MP4019.Planning with Monte Carlo Tree Search\\/053. Introduction to planning.mp451.63MB
  • MP4019.Planning with Monte Carlo Tree Search\\/054. Monte Carlo Tree Search.mp430.92MB
Latest Search: 1.WSS-207   2.IDBD-336   3.IDBD-229   4.ANHD-003   5.RKI-136   6.KCPB-004   7.TMRD-426   8.MVSD-139   9.MILD-714   10.SDDL-194   11.CESD-020   12.TYWD-029   13.DDB-171   14.SERO-0043   15.DMBA-129   16.NEO-311   17.GUR-014   18.PAR-232   19.ONSD-488   20.SND-43   21.PPSD-033   22.RKI-231   23.DJNO-115   24.DOKS-140   25.KIBD-096   26.MIBD-596   27.FETI-025   28.ONSD-180   29.MKCK-055   30.ELO-254   31.IDBD-363   32.ONSD-169   33.ONSD-686   34.BIB-082   35.IDBD-194   36.HRND-085   37.MJD-15   38.CMC-086   39.MILD-503   40.CMC-002   41.PMH-141   42.DKDN-004   43.DVH-436   44.DDB-157   45.MIRD-125   46.ONSD-676   47.ONSD-633   48.RBD-408   49.SDMS-461   50.CRPD-230   51.MADV-281   52.AMD-200   53.GWAZ-035   54.FAX-326   55.VIKG-127   56.DSE-211   57.RKI-283   58.PGD-628   59.AGEMIX-156   60.STAR-251   61.AGEMIX-083   62.RBD-335   63.LPTM-0019   64.REQ-080   65.DPM-003   66.MDB-399   67.ARMG-016   68.TND-019   69.BWSD-032   70.LMHG-004   71.207   72.336   73.229   74.003   75.136   76.004   77.426   78.139   79.714   80.194   81.020   82.029   83.171   84.0043   85.129   86.311   87.014   88.232   89.488   90.43   91.033   92.231   93.115   94.140   95.096   96.596   97.025   98.180   99.055   100.254   101.363   102.169   103.686   104.082   105.194   106.085   107.15   108.086   109.503   110.002   111.141   112.004   113.436   114.157   115.125   116.676   117.633   118.408   119.461   120.230   121.281   122.200   123.035   124.326   125.127   126.211   127.283   128.628   129.156   130.251   131.083   132.335   133.0019   134.080   135.003   136.399   137.016   138.019   139.032   140.004