Abstract
With the development of deregulated retail power markets, it is possible for end users equipped with smart meters and controllers to optimize their consumption cost portfolios by choosing various pricing plans from different retail electricity companies. This paper proposes a reinforcement learning-based decision system for assisting the selection of electricity pricing plans, which can minimize the electricity payment and consumption dissatisfaction for individual smart grid end user. The decision problem is modeled as a transition probability-free Markov decision process (MDP) with improved state framework. The proposed problem is solved using a Kernel approximator-integrated batch Q-learning algorithm, where some modifications of sampling and data representation are made to improve the computational and prediction performance. The proposed algorithm can extract the hidden features behind the time-varying pricing plans from a continuous high-dimensional state space. Case studies are based on data from real-world historical pricing plans and the optimal decision policy is learned without a priori information about the market environment. Results of several experiments demonstrate that the proposed decision model
can construct a precise predictive policy for individual user, effectively reducing their cost and energy consumption dissatisfaction.
can construct a precise predictive policy for individual user, effectively reducing their cost and energy consumption dissatisfaction.
Original language | English |
---|---|
Journal | IEEE Transactions on Smart Grid |
Volume | 12 |
Issue number | 3 |
Pages (from-to) | 2176 - 2187 |
ISSN | 1949-3053 |
DOIs | |
Publication status | Published - 2021 |
Keywords
- Smart grid end user
- Decision system
- Electricity market
- Value-based Q learning
- Demand response