{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T07:34:46Z","timestamp":1740123286951,"version":"3.37.3"},"reference-count":18,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2024,10,30]],"date-time":"2024-10-30T00:00:00Z","timestamp":1730246400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,10,30]],"date-time":"2024-10-30T00:00:00Z","timestamp":1730246400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Ann Oper Res"],"published-print":{"date-parts":[[2025,1]]},"DOI":"10.1007\/s10479-024-06282-0","type":"journal-article","created":{"date-parts":[[2024,10,30]],"date-time":"2024-10-30T16:17:46Z","timestamp":1730305066000},"page":"345-382","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Complexity Bounds for Deterministic Partially Observed Markov Decision Processes"],"prefix":"10.1007","volume":"344","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-9076-1932","authenticated-orcid":false,"given":"Cyrille","family":"Vessaire","sequence":"first","affiliation":[]},{"given":"Pierre","family":"Carpentier","sequence":"additional","affiliation":[]},{"given":"Jean-Philippe","family":"Chancelier","sequence":"additional","affiliation":[]},{"given":"Michel","family":"De Lara","sequence":"additional","affiliation":[]},{"given":"Alejandro","family":"Rodr\u00edguez-Mart\u00ednez","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,10,30]]},"reference":[{"issue":"1","key":"6282_CR1","doi-asserted-by":"publisher","first-page":"174","DOI":"10.1016\/0022-247X(65)90154-X","volume":"10","author":"KJ \u00c5str\u00f6m","year":"1965","unstructured":"\u00c5str\u00f6m, K. J. (1965). Optimal control of Markov processes with incomplete state information. Journal of Mathematical Analysis and Applications, 10(1), 174\u2013205. https:\/\/doi.org\/10.1016\/0022-247X(65)90154-X","journal-title":"Journal of Mathematical Analysis and Applications"},{"key":"6282_CR2","volume-title":"Dynamic programming","author":"R Bellman","year":"1957","unstructured":"Bellman, R. (1957). Dynamic programming. Princeto: Princeton Univ. Pr."},{"key":"6282_CR3","volume-title":"Dynamic programming and optimal control","author":"DP Bertsekas","year":"2000","unstructured":"Bertsekas, D. P. (2000). Dynamic programming and optimal control (2nd ed., Vol. 1 and 2). Belmont: Athena Scientific.","edition":"2"},{"key":"6282_CR4","series-title":"Mathematics in science and engineering","volume-title":"Stochastic optimal control: The discrete time case","author":"DP Bertsekas","year":"1978","unstructured":"Bertsekas, D. P., & Shreve, S. E. (1978). Stochastic optimal control: The discrete time case Mathematics in science and engineering (Vol. 139). New York: Academic Press."},{"key":"6282_CR5","unstructured":"Bonet, B. (2009). Deterministic POMDPs revisited. In Proceedings of the Twenty-Fifth Conference on Uncertainty in Artificial Intelligence, UAI \u201909 (pp. 59\u201366). Arlington, Virginia, USA. AUAI Press."},{"issue":"1","key":"6282_CR6","doi-asserted-by":"publisher","first-page":"222","DOI":"10.1287\/moor.22.1.222","volume":"22","author":"AN Burnetas","year":"1997","unstructured":"Burnetas, A. N., & Katehakis, M. N. (1997). Optimal adaptive policies for markov decision processes. Mathematics of Operations Research, 22(1), 222\u2013255. ISSN 0364765X, 15265471.","journal-title":"Mathematics of Operations Research"},{"key":"6282_CR7","unstructured":"Geffner, H., & Bonet, B. (1998). Solving Large POMDPs using Real Time Dynamic Programming. In Proceedings of Fall AAAI Symposium on POMDPS, Orlando, FL."},{"key":"6282_CR8","doi-asserted-by":"publisher","unstructured":"Kurniawati, H., Hsu, D., & Sun Lee, W. (2008). SARSOP: Efficient point-based POMDP planning by approximating optimally reachable belief spaces. In Robotics: Science and Systems IV. Robotics: Science and Systems Foundation. https:\/\/doi.org\/10.15607\/RSS.2008.IV.009.","DOI":"10.15607\/RSS.2008.IV.009"},{"key":"6282_CR9","unstructured":"Littman, M. L. (1996). Algorithms for Sequential Decision Making. PhD thesis, Brown University."},{"key":"6282_CR10","doi-asserted-by":"publisher","first-page":"213","DOI":"10.1016\/j.artint.2015.04.001","volume":"247","author":"J Pajarinen","year":"2017","unstructured":"Pajarinen, J., & Kyrki, V. (2017). Robotic manipulation of multiple objects as a POMDP. Artificial Intelligence, 247, 213\u2013228. https:\/\/doi.org\/10.1016\/j.artint.2015.04.001","journal-title":"Artificial Intelligence"},{"key":"6282_CR11","series-title":"Wiley Series in Probability and Statistics","doi-asserted-by":"publisher","DOI":"10.1002\/9780470316887.","volume-title":"Markov decision processes: Discrete stochastic dynamic programming","author":"ML Puterman","year":"1994","unstructured":"Puterman, M. L. (1994). Markov decision processes: Discrete stochastic dynamic programming Wiley Series in Probability and Statistics (1st ed.). London: Wiley. https:\/\/doi.org\/10.1002\/9780470316887.","edition":"1"},{"issue":"5","key":"6282_CR12","doi-asserted-by":"publisher","first-page":"1071","DOI":"10.1287\/opre.21.5.1071","volume":"21","author":"RD Smallwood","year":"1973","unstructured":"Smallwood, R. D., & Sondik, E. J. (1973). The optimal control of partially observable Markov processes over a finite horizon. Operations Research, 21(5), 1071\u20131088. https:\/\/doi.org\/10.1287\/opre.21.5.1071","journal-title":"Operations Research"},{"issue":"10","key":"6282_CR13","doi-asserted-by":"publisher","first-page":"1124","DOI":"10.1080\/24725854.2021.1895454","volume":"53","author":"LN Steimle","year":"2021","unstructured":"Steimle, L. N., Kaufman, D. L., & Denton, B. T. (2021). Multi-model Markov decision processes. IISE Transactions, 53(10), 1124\u20131139. https:\/\/doi.org\/10.1080\/24725854.2021.1895454","journal-title":"IISE Transactions"},{"key":"6282_CR14","volume-title":"Reinforcement learning: An introduction","author":"RS Sutton","year":"2018","unstructured":"Sutton, R. S., & Barto, A. G. (2018). Reinforcement learning: An introduction (2nd ed.). London: The MIT Press.","edition":"2"},{"key":"6282_CR15","unstructured":"Vessaire, C. (2022). Design and operation management of oil-fields taking into account partially observed uncertainties. PhD thesis, \u00c9cole des Ponts ParisTech."},{"key":"6282_CR16","unstructured":"Wesley, C., Katehakis, M. N., & Pirutinsky, D. (2019). Accelerating the computation of ucb and related indices for reinforcement learning. arXiv:1909.13158."},{"key":"6282_CR17","doi-asserted-by":"publisher","unstructured":"Wesley, C., Katehakis, M. N., & Pirutinsky, D. (2020). Reinforcement learning: a comparison of UCB versus alternative adaptive policies (pp. 127\u2013138). First Congress of Greek Mathematicians, De Gruyter, Berlin, Boston. ISBN 9783110663075. https:\/\/doi.org\/10.1515\/9783110663075-006.","DOI":"10.1515\/9783110663075-006"},{"issue":"11","key":"6282_CR18","doi-asserted-by":"publisher","first-page":"1073","DOI":"10.2307\/2583870","volume":"44","author":"DJ White","year":"1993","unstructured":"White, D. J. (1993). A survey of applications of Markov decision processes. The Journal of the Operational Research Society, 44(11), 1073. https:\/\/doi.org\/10.2307\/2583870","journal-title":"The Journal of the Operational Research Society"}],"container-title":["Annals of Operations Research"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10479-024-06282-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10479-024-06282-0\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10479-024-06282-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,1,3]],"date-time":"2025-01-03T02:12:29Z","timestamp":1735870349000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10479-024-06282-0"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,30]]},"references-count":18,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2025,1]]}},"alternative-id":["6282"],"URL":"https:\/\/doi.org\/10.1007\/s10479-024-06282-0","relation":{},"ISSN":["0254-5330","1572-9338"],"issn-type":[{"type":"print","value":"0254-5330"},{"type":"electronic","value":"1572-9338"}],"subject":[],"published":{"date-parts":[[2024,10,30]]},"assertion":[{"value":"30 March 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 September 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"30 October 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that there is no Conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"This article does not contain any studies with human participants or animals performed by any of the authors.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}}]}}