Single and multi-agent finite horizon reinforcement learning algorithms for smart grids (Record no. 432759)

MARC details
000 -LEADER
fixed length control field 04750nam a22002417a 4500
008 - FIXED-LENGTH DATA ELEMENTS--GENERAL INFORMATION
fixed length control field 241022b |||||||| |||| 00| 0 eng d
041 ## - LANGUAGE CODE
Language code of text/sound track or separate title Eng.
082 ## - DEWEY DECIMAL CLASSIFICATION NUMBER
Classification number 621.31
Item number VIV
100 ## - MAIN ENTRY--PERSONAL NAME
Personal name Vivek, V P
245 ## - TITLE STATEMENT
Title Single and multi-agent finite horizon reinforcement learning algorithms for smart grids
260 ## - PUBLICATION, DISTRIBUTION, ETC. (IMPRINT)
Place of publication, distribution, etc Bangalore :
Name of publisher, distributor, etc Indian Institute of Science,
Date of publication, distribution, etc 2024.
300 ## - PHYSICAL DESCRIPTION
Extent ix, 100p. :
Other physical details col. ill.
Accompanying material eThesis
Size of unit 3.936 Mb
500 ## - GENERAL NOTE
General note Includes bibliographical references.
502 ## - DISSERTATION NOTE
Dissertation note PhD;2024:Computer Science and Automation.
520 ## - SUMMARY, ETC.
Summary, etc In this thesis, we study sequential decision-making under uncertainty in the context of smart grids using reinforcement learning. The underlying mathematical model for reinforcement learning algorithms are Markov Decision Processes. A smart grid is essentially a concept for efficient electricity management using various technologies. We consider different models of smart grids involving single and multiple decision-making agents. We then develop reinforcement learning algorithms that can be applied to these models for efficient energy management. We also rigorously prove the convergence and stability of these algorithms. We then demonstrate the efficiency of these algorithms on the smart grid models we considered. Additionally, we run these algorithms on different randomly generated Markov Decision Processes to establish their correctness and convergence. We give a brief description of various studies given in this thesis. 1. Finite Horizon Q-learning algorithm for smart grids In this study, we develop a model of smart grid including different components like a main grid, microgrid with battery, renewal energy, and microcontroller. Subsequently, we define the problem of energy management in this model. This is modeled as a finite horizon Markov decision process. To address the complex decision-making process for energy management in the finite horizon Markov Decision Process, we develop a Q-learning algorithm. We apply this algorithm to our model effectively and demonstrate its performance. Additionally, we give rigorous mathematical proof establishing the stability and correctness of the algorithm. Our analysis of stability and convergence is purely based on ordinary differential equations. We also demonstrate the performance of our algorithm on different Markov Decision Processes generated randomly. 2. Finite Horizon Minimax Q-learning algorithm for smart grids. In this work, we developed a comprehensive model for smart grid that takes into account the the competition between two microgrids. Each microgrid have a battery, renewal energy, and microcontroller. Stochastic games are an important framework to capture the competitive environment. It is an extension of the Markov Decision Process by including multiple decision makers. It can also be viewed as an extension of games including a state factor. We model the competition between the two microgrids in our smart grid model as a finite horizon stochastic game. The interaction between the microgrids happens over a finite number of stages. We aim to solve the equilibrium of this competitive interaction. To this interest, the minimax concept is used to capture instantaneous interaction. Subsequently, we develop a finite horizon minimax Q-learning algorithm to capture the long-term equilibrium of the competition between two microgrids. The performance of the algorithm is effectively demonstrated on smart grid setup. Additionally, we demonstrate the correctness and convergence of the algorithm on randomly generated stochastic games. Furthermore, a rigorous mathematical proof of the stability and convergence of the algorithm is given. 3. Finite Horizon SOR Q-learning In this final part of our study, we proposed a generalization of the finite horizon problem using discounting and proposed an improvement of the finite horizon Q-learning algorithm for this problem. The rate of convergence of a reinforcement learning algorithm is an important parameter of its performance. There are techniques used in the literature to improve the rate of convergence of reinforcement learning algorithms. One of them is successive over-relaxation. This was originally used in linear algebra to improve the performance of the Gauss-Siedel iterative scheme used for solving linear system of equations. We apply this technique in the finite horizon Q-learning of discounted problems to get a better algorithm that has better asymptotic performance.
650 ## - SUBJECT ADDED ENTRY--TOPICAL TERM
Topical term or geographic name as entry element Reinforcement Learning
650 ## - SUBJECT ADDED ENTRY--TOPICAL TERM
Topical term or geographic name as entry element Smart Grids
650 ## - SUBJECT ADDED ENTRY--TOPICAL TERM
Topical term or geographic name as entry element Markov Decision Processes
650 ## - SUBJECT ADDED ENTRY--TOPICAL TERM
Topical term or geographic name as entry element Q-learning algorithm
700 ## - ADDED ENTRY--PERSONAL NAME
Personal name Advised by Bhatnagar, Shalabh
856 ## - ELECTRONIC LOCATION AND ACCESS
Uniform Resource Identifier https://etd.iisc.ac.in/handle/2005/6644
942 ## - ADDED ENTRY ELEMENTS (KOHA)
Koha item type Thesis

No items available.

                                                                                                                                                                                                    Facebook    Twitter

                             Copyright © 2024. J.R.D. Tata Memorial Library, Indian Institute of Science, Bengaluru - 560012

                             Contact   Phone: +91 80 2293 2832