Blog

Learn About Our Meetup

4500+ Members

[D] What do you see as the most promising directions for reducing sample inefficiency in reinforcement learning?

I often read from ML researchers, but more from computational cognitive scientists, that humans are able to generalize patterns from only a few data points or use “rich, informative priors” even as children, and how that is very important for us as cognitive beings that sets us apart from the current neural network approaches to RL used today.

I’m also not entirely convinced that the current neural net paradigm with the McCulloch–Pitts-esque neurons is ever going to become sample efficient enough for real-world reinforcement learning tasks. It seems like despite our best efforts to increase sample efficiency in NN techniques, the most impressive results still use hundreds of thousands or more simulations/data points that could be infeasible to implement for any sufficiently complex real-world environments.

That being said, what approaches are you most excited for in reducing sample efficiency in reinforcement learning or in neural network techniques in general?

submitted by /u/p6m_lattice
[link] [comments]

Next Meetup

 

Days
:
Hours
:
Minutes
:
Seconds

 

Plug yourself into AI and don't miss a beat

 


Toronto AI is a social and collaborative hub to unite AI innovators of Toronto and surrounding areas. We explore AI technologies in digital art and music, healthcare, marketing, fintech, vr, robotics and more. Toronto AI was founded by Dave MacDonald and Patrick O'Mara.