A reinforcement learning approach to dynamic norm generation

dc.contributor.advisorUlieru, Mihaela
dc.contributor.authorHosseini, Hadi
dc.date.accessioned2023-03-01T16:19:53Z
dc.date.available2023-03-01T16:19:53Z
dc.date.issued2010
dc.date.updated2016-07-15T00:00:00Z
dc.description.abstractThis thesis proposes a two-level learning framework for dynamic norm generation. This framework uses the Bayesian reinforcement learning technique to extract behavioral norms and domain-dependent knowledge in a certain environment and later incorporates them into the learning agents in different settings. Reinforcement learning (RL) and norms are mutually beneficial: norms can be extracted through RL, and RL can be improved by incorporating behavioral norms as prior probability distributions into learning agents. An agent should be confident about its beliefs in order to generalize them and use them in future settings. The confidence level is developed by checking two conditions: how familiar the agent is with the current world and its dynamics (including the norm system), and whether it has converged to an optimal policy. A Bayesian dynamic programming technique is implemented and then compared to other methods such as Q-learning and Dyna. It is shown that Bayesian RL outperforms other techniques in finding the best equilibrium for the exploration-exploitation problem. This thesis demonstrates how an agent can extract behavioral norms and adapt its beliefs based on the domain knowledge it has acquired through the learning process. Scenarios with different percentages of similarity and goals are examined. The experimental results show that the normative agent, having been trained in an initial environment, is able to adjust its beliefs about the dynamics and behavioral norms in a new environment, and thus it converges to the optimal policy more quickly, especially in the early stages of learning.
dc.description.copyright© Hadi Hosseini, 2010
dc.description.note(UNB accession number) Thesis 8607. (OCoLC)821799515
dc.description.noteM.C.S. University of New Brunswick, Faculty of Computer Science, 2010
dc.formattext/xml
dc.format.extentxiv, 123 pages ; illustrations
dc.format.mediumelectronic
dc.identifier.oclc(OCoLC)821799515
dc.identifier.otherThesis 8607
dc.identifier.urihttps://unbscholar.lib.unb.ca/handle/1882/13533
dc.identifier.urlhttp://gateway.proquest.com/openurl?url_ver=Z39.88-2004&rft_val_fmt=info:ofi/fmt:kev:mtx:dissertation&res_dat=xri:pqdiss&rft_da
dc.language.isoen_CA
dc.publisherUniversity of New Brunswick
dc.rightshttp://purl.org/coar/access_right/c_abf2
dc.subject.disciplineComputer Science
dc.subject.lcshReinforcement learning.
dc.subject.lcshDynamic programming.
dc.subject.lcshBayesian statistical decision theory.
dc.subject.lcshAutomatic programming (Computer science)
dc.titleA reinforcement learning approach to dynamic norm generation
dc.typemaster thesis
thesis.degree.disciplineComputer Science
thesis.degree.fullnameMaster of Computer Science
thesis.degree.grantorUniversity of New Brunswick
thesis.degree.levelmasters
thesis.degree.nameM.C.S.

Files

Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
item.pdf
Size:
1.61 MB
Format:
Adobe Portable Document Format