Veuillez utiliser cette adresse pour citer ce document :
http://dspace1.univ-tlemcen.dz/handle/112/23064
Affichage complet
Élément Dublin Core | Valeur | Langue |
---|---|---|
dc.contributor.author | Koudad, Zoulikha | - |
dc.date.accessioned | 2024-09-23T12:33:59Z | - |
dc.date.available | 2024-09-23T12:33:59Z | - |
dc.date.issued | 2024-07-15 | - |
dc.identifier.uri | http://dspace1.univ-tlemcen.dz/handle/112/23064 | - |
dc.description.abstract | The hierarchical reinforcement learning framework breaks down the reinforcement learning problem into subtasks or extended actions called options in order to facilitate its resolution. Di erent models have been proposed where options were manually prede ned or semi-automatically discovered. However, the automatic discovery of options has become a real challenge for research in hierarchical reinforcement learning. In this thesis we propose two automatic option discovery method for hierarchical reinforcement learning. The rst method that we call FAOD (Fast Automatic Option Discovery). In this contribution, we took inspiration from robot learning methods to categorize the sensorimotor ow during navigation. Thus, FAOD agent moves along the walls to discover the rooms' contour, closed spaces, doors and bottleneck regions to de ne terminate states and initiation sets for options. In the second contribution our learning agent uses his sense of direction to discover the shortest paths and shortcuts after an exploration based on intrinsic motivation, without resorting to the algorithms of the graph theory, these discoveries subsequently serve to discover the termination conditions and the initiation states of the options. For the learning of options policies, the agent uses his experience of exploration as well as learning by temporal di erence strategy. We tested and validated this approach on di erent maze problems and on the tic-tac-toe game. | en_US |
dc.language.iso | en | en_US |
dc.publisher | University of tlemcen | en_US |
dc.relation.ispartofseries | 761 Doct Informatique; | - |
dc.subject | Hierarchical reinforcement learning; Reinforcement learning; Option Discovery; Markov decision process; Actor-criticlearning, Way nding; Intrinsic motivation. | en_US |
dc.title | Methods for Automatic Option Discovery in Hierarchical Reinforcement Learning | en_US |
dc.type | Thesis | en_US |
Collection(s) : | Doctorat Classique SIC |
Fichier(s) constituant ce document :
Fichier | Description | Taille | Format | |
---|---|---|---|---|
Methods_for_Automatic_Option_Discovery_in.pdf | 3,3 MB | Adobe PDF | Voir/Ouvrir |
Tous les documents dans DSpace sont protégés par copyright, avec tous droits réservés.