Operation scheduling in a solar thermal system: A reinforcement learning-based framework
Author | dc.contributor.author | Correa Jullian, Camila | |
Author | dc.contributor.author | López Droguett, Enrique | |
Author | dc.contributor.author | Cardemil Iglesias, José | |
Admission date | dc.date.accessioned | 2020-07-01T23:55:55Z | |
Available date | dc.date.available | 2020-07-01T23:55:55Z | |
Publication date | dc.date.issued | 2020 | |
Cita de ítem | dc.identifier.citation | Applied Energy 268 (2020) 114943 | es_ES |
Identifier | dc.identifier.other | 10.1016/j.apenergy.2020.114943 | |
Identifier | dc.identifier.uri | https://repositorio.uchile.cl/handle/2250/175738 | |
Abstract | dc.description.abstract | Reinforcement learning (RL) provides an alternative method for designing condition-based decision making in engineering systems. In this study, a simple and flexible RL tabular Q-learning framework is employed to identify the optimal operation schedules for a solar hot water system according to action-reward feedback. The system is simulated in TRNSYS software. Three energy sources must supply a building's hot-water demand: low-cost heat from solar thermal collectors and a heat-recovery chiller, coupled to a conventional heat pump. Key performance indicators are used as rewards for balancing the system's performance with regard to energy efficiency, heat-load delivery, and operational costs. A sensitivity analysis is performed for different reward functions and meteorological conditions. Optimal schedules are obtained for selected scenarios in January, April, July, and October, according to the dynamic conditions of the system. The results indicate that when solar radiation is widely available (October through April), the nominal operation schedule frequently yields the highest performance. However, the obtained schedule differs when the solar radiation is reduced, for instance, in July. On average, with prioritization of the efficient use of both low-cost energy sources, the performance in July can be on average 21% higher than under nominal schedule-based operation. | es_ES |
Patrocinador | dc.description.sponsorship | ANID/FONDAP "Solar Energy Research Center" SERC-Chile 15110019 Comision Nacional de Investigacion Cientifica y Tecnologica (CONICYT) CONICYT FONDECYT 1190720 | es_ES |
Lenguage | dc.language.iso | en | es_ES |
Publisher | dc.publisher | Elsevier | es_ES |
Type of license | dc.rights | Attribution-NonCommercial-NoDerivs 3.0 Chile | * |
Link to License | dc.rights.uri | http://creativecommons.org/licenses/by-nc-nd/3.0/cl/ | * |
Source | dc.source | Applied Energy | es_ES |
Keywords | dc.subject | Solar hot water systems | es_ES |
Keywords | dc.subject | Reinforcement learning | es_ES |
Keywords | dc.subject | Intelligent control systems | es_ES |
Keywords | dc.subject | Condition-based decision making | es_ES |
Keywords | dc.subject | Q-learning | es_ES |
Keywords | dc.subject | Machine learning | es_ES |
Título | dc.title | Operation scheduling in a solar thermal system: A reinforcement learning-based framework | es_ES |
Document type | dc.type | Artículo de revista | es_ES |
dcterms.accessRights | dcterms.accessRights | Acceso Abierto | |
Cataloguer | uchile.catalogador | laj | es_ES |
Indexation | uchile.index | Artículo de publicación ISI | |
Indexation | uchile.index | Artículo de publicación SCOPUS |
Files in this item
This item appears in the following Collection(s)
-
Artículos de revistas
Artículos de revistas