{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,24]],"date-time":"2025-11-24T07:15:14Z","timestamp":1763968514498,"version":"3.40.3"},"publisher-location":"Cham","reference-count":24,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031204999"},{"type":"electronic","value":"9783031205002"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-20500-2_50","type":"book-chapter","created":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T05:12:32Z","timestamp":1672549952000},"page":"613-625","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Benchmarking Deep Reinforcement Learning Based Energy Management Systems for\u00a0Hybrid Electric Vehicles"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-0408-9319","authenticated-orcid":false,"given":"Wu","family":"Yuankai","sequence":"first","affiliation":[]},{"given":"Lian","family":"Renzong","sequence":"additional","affiliation":[]},{"given":"Wang","family":"Yong","sequence":"additional","affiliation":[]},{"given":"Lin","family":"Yi","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,1,1]]},"reference":[{"issue":"6","key":"50_CR1","doi-asserted-by":"publisher","first-page":"26","DOI":"10.1109\/MSP.2017.2743240","volume":"34","author":"K Arulkumaran","year":"2017","unstructured":"Arulkumaran, K., Deisenroth, M.P., Brundage, M., Bharath, A.A.: Deep reinforcement learning: a brief survey. IEEE Sig. Process. Mag. 34(6), 26\u201338 (2017)","journal-title":"IEEE Sig. Process. Mag."},{"issue":"7","key":"50_CR2","doi-asserted-by":"publisher","first-page":"6453","DOI":"10.1109\/TVT.2019.2914457","volume":"68","author":"A Biswas","year":"2019","unstructured":"Biswas, A., Emadi, A.: Energy management systems for electrified powertrains: state-of-the-art review and future trends. IEEE Trans. Veh. Technol. 68(7), 6453\u20136467 (2019)","journal-title":"IEEE Trans. Veh. Technol."},{"unstructured":"Haarnoja, T., Zhou, A., Abbeel, P., Levine, S.: Soft actor-critic: off-policy maximum entropy deep reinforcement learning with a stochastic actor. In: International Conference on Machine Learning, pp. 1861\u20131870 (2018)","key":"50_CR3"},{"key":"50_CR4","doi-asserted-by":"publisher","DOI":"10.1016\/j.apenergy.2019.113708","volume":"254","author":"X Han","year":"2019","unstructured":"Han, X., He, H., Wu, J., Peng, J., Li, Y.: Energy management based on reinforcement learning with double deep q-learning for a hybrid electric tracked vehicle. Appl. Energy 254, 113708 (2019)","journal-title":"Appl. Energy"},{"doi-asserted-by":"crossref","unstructured":"Henderson, P., Islam, R., Bachman, P., Pineau, J., Precup, D., Meger, D.: Deep reinforcement learning that matters. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 32 (2018)","key":"50_CR5","DOI":"10.1609\/aaai.v32i1.11694"},{"doi-asserted-by":"crossref","unstructured":"Hilleli, B., El-Yaniv, R.: Toward deep reinforcement learning without a simulator: an autonomous steering example. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 32 (2018)","key":"50_CR6","DOI":"10.1609\/aaai.v32i1.11490"},{"issue":"2","key":"50_CR7","doi-asserted-by":"publisher","first-page":"187","DOI":"10.3390\/app8020187","volume":"8","author":"Y Hu","year":"2018","unstructured":"Hu, Y., Li, W., Xu, K., Zahid, T., Qin, F., Li, C.: Energy management strategy for a hybrid electric vehicle based on deep reinforcement learning. Appl. Sci. 8(2), 187 (2018)","journal-title":"Appl. Sci."},{"key":"50_CR8","doi-asserted-by":"publisher","DOI":"10.1016\/j.apenergy.2019.113762","volume":"255","author":"Y Li","year":"2019","unstructured":"Li, Y., He, H., Khajepour, A., Wang, H., Peng, J.: Energy management for a power-split hybrid electric bus via deep reinforcement learning with terrain information. Appl. Energy 255, 113762 (2019)","journal-title":"Appl. Energy"},{"issue":"8","key":"50_CR9","doi-asserted-by":"publisher","first-page":"7416","DOI":"10.1109\/TVT.2019.2926472","volume":"68","author":"Y Li","year":"2019","unstructured":"Li, Y., He, H., Peng, J., Wang, H.: Deep reinforcement learning-based energy management for a series hybrid electric vehicle enabled by history cumulative trip information. IEEE Trans. Veh. Technol. 68(8), 7416\u20137430 (2019)","journal-title":"IEEE Trans. Veh. Technol."},{"doi-asserted-by":"crossref","unstructured":"Lian, R., Peng, J., Wu, Y., Tan, H., Zhang, H.: Rule-interposing deep reinforcement learning based energy management strategy for power-split hybrid electric vehicle. Energy, 117297 (2020)","key":"50_CR10","DOI":"10.1016\/j.energy.2020.117297"},{"unstructured":"Lillicrap, T.P., et al.: Continuous control with deep reinforcement learning. In: ICLR (Poster) (2016)","key":"50_CR11"},{"issue":"4","key":"50_CR12","doi-asserted-by":"publisher","first-page":"1497","DOI":"10.1109\/TMECH.2017.2707338","volume":"22","author":"T Liu","year":"2017","unstructured":"Liu, T., Hu, X., Li, S.E., Cao, D.: Reinforcement learning optimized look-ahead energy management of a parallel hybrid electric vehicle. IEEE\/ASME Trans. Mechatron. 22(4), 1497\u20131507 (2017)","journal-title":"IEEE\/ASME Trans. Mechatron."},{"key":"50_CR13","doi-asserted-by":"publisher","first-page":"544","DOI":"10.1016\/j.energy.2018.07.022","volume":"160","author":"T Liu","year":"2018","unstructured":"Liu, T., Wang, B., Yang, C.: Online Markov chain-based energy management for a hybrid tracked vehicle with speedy q-learning. Energy 160, 544\u2013555 (2018)","journal-title":"Energy"},{"issue":"2\u20133","key":"50_CR14","doi-asserted-by":"publisher","first-page":"458","DOI":"10.1016\/j.neunet.2007.12.043","volume":"21","author":"DV Prokhorov","year":"2008","unstructured":"Prokhorov, D.V.: Toyota Prius HEV neurocontrol and diagnostics. Neural Netw. 21(2\u20133), 458\u2013465 (2008)","journal-title":"Neural Netw."},{"key":"50_CR15","doi-asserted-by":"publisher","first-page":"67","DOI":"10.1016\/j.trc.2018.12.018","volume":"99","author":"X Qi","year":"2019","unstructured":"Qi, X., Luo, Y., Wu, G., Boriboonsomsin, K., Barth, M.: Deep reinforcement learning enabled self-learning control for energy efficient driving. Transp. Res. Part C: Emerg. Technol. 99, 67\u201381 (2019)","journal-title":"Transp. Res. Part C: Emerg. Technol."},{"unstructured":"Schulman, J., Levine, S., Abbeel, P., Jordan, M., Moritz, P.: Trust region policy optimization. In: International Conference on Machine Learning, pp. 1889\u20131897. PMLR (2015)","key":"50_CR16"},{"unstructured":"Schulman, J., Wolski, F., Dhariwal, P., Radford, A., Klimov, O.: Proximal policy optimization algorithms. arXiv preprint arXiv:1707.06347 (2017)","key":"50_CR17"},{"key":"50_CR18","doi-asserted-by":"publisher","first-page":"548","DOI":"10.1016\/j.enconman.2019.05.038","volume":"195","author":"H Tan","year":"2019","unstructured":"Tan, H., Zhang, H., Peng, J., Jiang, Z., Wu, Y.: Energy management of hybrid electric bus based on deep reinforcement learning in continuous state and action space. Energy Convers. Manag. 195, 548\u2013560 (2019)","journal-title":"Energy Convers. Manag."},{"doi-asserted-by":"crossref","unstructured":"Van Hasselt, H., Guez, A., Silver, D.: Deep reinforcement learning with double q-learning. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 30 (2016)","key":"50_CR19","DOI":"10.1609\/aaai.v30i1.10295"},{"doi-asserted-by":"crossref","unstructured":"Won, D.O., M\u00fcller, K.R., Lee, S.W.: An adaptive deep reinforcement learning framework enables curling robots with human-like performance in real-world conditions. Sci. Robot. 5(46), eabb9764 (2020)","key":"50_CR20","DOI":"10.1126\/scirobotics.abb9764"},{"unstructured":"Wong, J.Y.: Theory of Ground Vehicles. Wiley, Hoboken (2008)","key":"50_CR21"},{"key":"50_CR22","doi-asserted-by":"publisher","first-page":"799","DOI":"10.1016\/j.apenergy.2018.03.104","volume":"222","author":"J Wu","year":"2018","unstructured":"Wu, J., He, H., Peng, J., Li, Y., Li, Z.: Continuous reinforcement learning of energy management with deep q network for a power split hybrid electric bus. Appl. Energy 222, 799\u2013811 (2018)","journal-title":"Appl. Energy"},{"key":"50_CR23","doi-asserted-by":"publisher","first-page":"454","DOI":"10.1016\/j.apenergy.2019.04.021","volume":"247","author":"Y Wu","year":"2019","unstructured":"Wu, Y., Tan, H., Peng, J., Zhang, H., He, H.: Deep reinforcement learning of energy management with continuous control strategy and traffic information for a series-parallel plug-in hybrid electric bus. Appl. Energy 247, 454\u2013466 (2019)","journal-title":"Appl. Energy"},{"key":"50_CR24","doi-asserted-by":"publisher","first-page":"372","DOI":"10.1016\/j.apenergy.2016.03.082","volume":"171","author":"Y Zou","year":"2016","unstructured":"Zou, Y., Liu, T., Liu, D., Sun, F.: Reinforcement learning-based real-time energy management for a hybrid tracked vehicle. Appl. Energy 171, 372\u2013382 (2016)","journal-title":"Appl. Energy"}],"container-title":["Lecture Notes in Computer Science","Artificial Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-20500-2_50","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T05:45:29Z","timestamp":1672551929000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-20500-2_50"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031204999","9783031205002"],"references-count":24,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-20500-2_50","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"1 January 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"CICAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"CAAI International Conference on Artificial Intelligence","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Beijing","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 August 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28 August 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"cicai2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/cicai.caai.cn\/#\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"472","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"164","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"35% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.1","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.7","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}