Please use this identifier to cite or link to this item:
http://hdl.handle.net/10263/7506
Title: | Verification of Reinforcement Learning Models: |
Other Titles: | Comparing Construction of Environment Models |
Authors: | A, Patrick Jeeva |
Keywords: | Gaussian Ensemble Model Bayesian Neural Network Monte Carlo Dropout Model Markov Decision Processes |
Issue Date: | Jun-2024 |
Publisher: | Indian Statistical Institute, Kolkata |
Citation: | 65p. |
Series/Report no.: | MTech(CS) Dissertation;22-20 |
Abstract: | In recent years of advancements in reinforcement learning (RL), utilizing neural network based models to make decisions in dynamic and complex environments has emerged as a powerful paradigm. In particular, model based reinforcement learning has been widely used for its ability to increase learning efficiency and performance. By constructing an environment model beforehand, the agent attains a prior knowledge of the dynamics of the model to take informed decisions and converge fast to optimal policies. Real-world environments are often intricate and subject to external disturbances, posing substantial challenges for accurate modeling. Addressing these challenges requires the application of sophisticated neural network-based models that can effectively approximate the underlying environment dynamics. In this work, we develop and evaluate extensive neural network models, specifically focusing on Gaussian Ensemble models, Bayesian neural networks, and Monte Carlo Dropout techniques, to approximate various standard gym environments. These models are trained on different numbers of samples to understand their efficiency and accuracy in capturing environment dynamics. Once trained, the neural network models are used to construct Markov Decision Processes (MDPs) with various discretization strategies. The constructed MDPs are then analyzed and compared to evaluate the performance of each neural network approach. The purpose of this thesis is to present a comprehensive study on the construction of environment models using advanced neural network techniques. We aim to approximate the standard environments in the reinforcement learning setup, utilizing a variety of neural networks and compare the efficiency based on the reconstruction of MDPs. |
Description: | Dissertation under the supervision of Dr. Swarup Mohalik and Dr. Ansuman Banerjee. |
URI: | http://hdl.handle.net/10263/7506 |
Appears in Collections: | Dissertations - M Tech (CS) |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
Patrick_jeeva-Cs2220-Mtech-CS-2024.pdf | Dissertations - M Tech (CS) | 2.69 MB | Adobe PDF | View/Open |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.