A Recurrent Memory Model Implementation of Reinforcement Learning
| dc.contributor.author | Church, Kinsey Antonina | |
| dc.contributor.supervisor | Chartier, Sylvain | |
| dc.date.accessioned | 2024-11-25T16:49:43Z | |
| dc.date.available | 2024-11-25T16:49:43Z | |
| dc.date.issued | 2024-11-25 | |
| dc.description.abstract | Reinforcement Learning (RL) provides a robust framework for understanding how humans and animals learn from and adapt to our environments through trial and error. It is comprised of two processes that work together, exploration and exploitation. This thesis presents an implementation of these two aspects of RL using recurrent neural networks. Several different challenges were approached including one-to-many problems, nonlinearly separable problems, and the generation and representation of random behaviour. The first study is an implementation of exploitation that is capable of cycling through previously learned behaviours and stabilizing on the correct one for each given problem. This is achieved with contextual tags and a unit that represents environmental feedback. It is able to solve nonlinearly separable and one-to-many problems. The second study is an implementation of exploration that is able to randomly select from available options and adapt based on the feedback from those decisions. It tackles the question of how to generate and represent random behaviour, as well as how to represent and apply reward. The implementations and techniques detailed in this thesis could be applied to many other similar models of cognitive behaviour. Further research could experiment by combining these two models and investigating the implementation of different trade-off strategies. | |
| dc.identifier.uri | http://hdl.handle.net/10393/49885 | |
| dc.identifier.uri | https://doi.org/10.20381/ruor-30710 | |
| dc.language.iso | en | |
| dc.publisher | Université d'Ottawa / University of Ottawa | |
| dc.subject | Reinforcement Learning | |
| dc.subject | Randomness | |
| dc.subject | Recurrent associative memory | |
| dc.subject | Exploration-Exploitation | |
| dc.subject | Artificial neural networks | |
| dc.subject | Biasing decisions | |
| dc.subject | Cognitive psychology | |
| dc.title | A Recurrent Memory Model Implementation of Reinforcement Learning | |
| dc.type | Thesis | en |
| thesis.degree.discipline | Sciences sociales / Social Sciences | |
| thesis.degree.level | Masters | |
| thesis.degree.name | MA | |
| uottawa.department | Psychologie / Psychology |
