{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,28]],"date-time":"2026-04-28T12:56:19Z","timestamp":1777380979665,"version":"3.51.4"},"reference-count":27,"publisher":"SAGE Publications","issue":"1","license":[{"start":{"date-parts":[[2025,2,1]],"date-time":"2025-02-01T00:00:00Z","timestamp":1738368000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/journals.sagepub.com\/page\/policies\/text-and-data-mining-license"}],"content-domain":{"domain":["journals.sagepub.com"],"crossmark-restriction":true},"short-container-title":["Journal of Ambient Intelligence and Smart Environments"],"published-print":{"date-parts":[[2025,2]]},"abstract":"<jats:p>To solve the problems of sleep disorders such as difficulty in falling asleep and insufficient sleep depth caused by uncomfortable indoor temperature, this paper proposes a deep reinforcement learning method based on deep Q-network (DQN) with human sleep electroencephalogram (EEG) as input to improve human sleep. Firstly, the EEG is subjected to a short-time Fourier transform to construct a time-frequency feature data set, which is used as input to DQN along with temperature. Secondly, the agent performs environmental interaction actions in each time step and returns a reward value. Finally, the optimal strategy for indoor temperature control is formulated by the agent. The simulation results show that this method can dynamically adjust the indoor temperature to the optimal temperature for human sleep, and can alleviate sleep disorders, which has certain practical significance.<\/jats:p>","DOI":"10.3233\/ais-230294","type":"journal-article","created":{"date-parts":[[2023,10,27]],"date-time":"2023-10-27T11:14:55Z","timestamp":1698405295000},"page":"63-74","update-policy":"https:\/\/doi.org\/10.1177\/sage-journals-update-policy","source":"Crossref","is-referenced-by-count":1,"title":["Research on human sleep improvement method based on DQN"],"prefix":"10.1177","volume":"17","author":[{"given":"Yunzhi","family":"Tian","sequence":"first","affiliation":[{"name":"School of Electrical and Control Engineering, Shaanxi University of Science &amp; Technology, Xi\u2019an, China"},{"name":"Shaanxi Artificial Intelligence Joint Laboratory, Xi\u2019an, China"}]},{"given":"Qiang","family":"Zhou","sequence":"additional","affiliation":[{"name":"School of Electrical and Control Engineering, Shaanxi University of Science &amp; Technology, Xi\u2019an, China"},{"name":"Shaanxi Artificial Intelligence Joint Laboratory, Xi\u2019an, China"}]},{"given":"Wan","family":"Li","sequence":"additional","affiliation":[{"name":"School of Electronic Information and Artificial Intelligence, Shaanxi University of Science &amp; Technology, Xi\u2019an, China"},{"name":"Shaanxi Artificial Intelligence Joint Laboratory, Xi\u2019an, China"}]}],"member":"179","published-online":{"date-parts":[[2025,3,19]]},"reference":[{"key":"e_1_3_2_2_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.psychres.2019.112579"},{"key":"e_1_3_2_3_2","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0004609"},{"key":"e_1_3_2_4_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.sleh.2022.02.001"},{"key":"e_1_3_2_5_2","unstructured":"Fan J. Wang Z. Xie Y. Yang Z. A theoretical analysis of deep Q-learning in: Proceedings of the 2nd Conference on Learning for Dynamics and Control (L4DC) PMLR 2020 pp.\u00a0486\u2013489."},{"key":"e_1_3_2_6_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.buildenv.2021.108666"},{"key":"e_1_3_2_7_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.energy.2022.125679"},{"key":"e_1_3_2_8_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.applthermaleng.2022.118552"},{"key":"e_1_3_2_9_2","doi-asserted-by":"publisher","DOI":"10.3233\/AIS-140288"},{"key":"e_1_3_2_10_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.enbuild.2022.112284"},{"key":"e_1_3_2_11_2","doi-asserted-by":"publisher","unstructured":"Jinqiang X. Zi-ping F. Application of two-stage fuzzy PID control in variable refrigerant volume air conditioning systems in: 2009 Chinese Control and Decision Conference (CCDC) IEEE 2009 pp.\u00a05619\u20135622. doi:10.1109\/CCDC.2009.5195199.","DOI":"10.1109\/CCDC.2009.5195199"},{"key":"e_1_3_2_12_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.jobe.2023.105830"},{"key":"e_1_3_2_13_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.enbuild.2022.111903"},{"key":"e_1_3_2_14_2","unstructured":"Lin L.J. Reinforcement Learning for Robots Using Neural Networks Carnegie Mellon University 1992."},{"key":"e_1_3_2_15_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.bspc.2023.104693"},{"key":"e_1_3_2_16_2","unstructured":"Mnih V. Kavukcuoglu K. Silver D. Graves A. Antonoglou I. Wierstra D. Riedmiller M. Playing atari with deep reinforcement learning 2013. arXiv:1312.5602."},{"key":"e_1_3_2_17_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.bspc.2022.103819"},{"key":"e_1_3_2_18_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.bspc.2022.103751"},{"key":"e_1_3_2_19_2","doi-asserted-by":"publisher","DOI":"10.1038\/nature16961"},{"key":"e_1_3_2_20_2","doi-asserted-by":"publisher","DOI":"10.7507\/1001-5515.202211021"},{"key":"e_1_3_2_21_2","doi-asserted-by":"publisher","DOI":"10.1109\/TSMC.2016.2623766"},{"key":"e_1_3_2_22_2","doi-asserted-by":"publisher","DOI":"10.1109\/TIE.2020.3001849"},{"key":"e_1_3_2_23_2","doi-asserted-by":"publisher","DOI":"10.1109\/TCYB.2019.2906694"},{"key":"e_1_3_2_24_2","doi-asserted-by":"publisher","unstructured":"Wei T. Wang Y. Zhu Q. Deep reinforcement learning for building HVAC control in: Proceedings of the 54th Annual Design Automation Conference 2017 (DAC) ACM 2017 pp.\u00a01\u20136. doi:10.1145\/3061639.3062224.","DOI":"10.1145\/3061639.3062224"},{"key":"e_1_3_2_25_2","doi-asserted-by":"publisher","DOI":"10.1109\/TFUZZ.2014.2312979"},{"key":"e_1_3_2_26_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.compag.2022.107546"},{"key":"e_1_3_2_27_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.sleepx.2022.100053"},{"key":"e_1_3_2_28_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.sleep.2022.08.014"}],"container-title":["Journal of Ambient Intelligence and Smart Environments"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/journals.sagepub.com\/doi\/pdf\/10.3233\/AIS-230294","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/journals.sagepub.com\/doi\/full-xml\/10.3233\/AIS-230294","content-type":"application\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/journals.sagepub.com\/doi\/pdf\/10.3233\/AIS-230294","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,4,28]],"date-time":"2026-04-28T09:19:33Z","timestamp":1777367973000},"score":1,"resource":{"primary":{"URL":"https:\/\/journals.sagepub.com\/doi\/10.3233\/AIS-230294"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,2]]},"references-count":27,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2025,2]]}},"alternative-id":["10.3233\/AIS-230294"],"URL":"https:\/\/doi.org\/10.3233\/ais-230294","relation":{},"ISSN":["1876-1364","1876-1372"],"issn-type":[{"value":"1876-1364","type":"print"},{"value":"1876-1372","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,2]]}}}