{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,28]],"date-time":"2026-02-28T18:08:51Z","timestamp":1772302131967,"version":"3.50.1"},"reference-count":43,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Pattern Anal. Mach. Intell."],"published-print":{"date-parts":[[2022]]},"DOI":"10.1109\/tpami.2022.3174603","type":"journal-article","created":{"date-parts":[[2022,5,12]],"date-time":"2022-05-12T19:36:42Z","timestamp":1652384202000},"page":"1-1","source":"Crossref","is-referenced-by-count":17,"title":["DeepTag: A General Framework for Fiducial Marker Design and Detection"],"prefix":"10.1109","author":[{"given":"Zhuming","family":"Zhang","sequence":"first","affiliation":[{"name":"X-Lab, Guangdong Virtual Reality Technology Co., Ltd., SHENZHEN, GUANGDONG, China, 518000"}]},{"given":"Yongtao","family":"Hu","sequence":"additional","affiliation":[{"name":"X-Lab, Guangdong Virtual Reality Technology Co., Ltd., SHENZHEN, GUANGDONG, China, 518000"}]},{"given":"Guoxing","family":"Yu","sequence":"additional","affiliation":[{"name":"X-lab, Guangdong Virtual Reality Co., Ltd., Shenzhen, Guangdong, China"}]},{"given":"Jingwen","family":"Dai","sequence":"additional","affiliation":[{"name":"X-Lab, Guangdong Virtual Reality Technology Co., Ltd., Shenzhen, Guangdong, China"}]}],"member":"263","reference":[{"key":"ref1","first-page":"139","article-title":"ARToolKitPlus for pose tracking on mobile devices","volume-title":"Proc. 12th Comput. Vis. Winter Workshop","author":"Wagner"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2011.5979561"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2016.7759617"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2014.01.005"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2015.09.023"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/TVCG.2020.2988466"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2011.5995544"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2016.2519024"},{"key":"ref9","article-title":"E2ETag: An end-to-end trainable method for generating and detecting fiducial markers","volume-title":"Proc. 31st Brit. Mach. Vis. Conf.","author":"Peace"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1145\/3476576.3476619"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2019.2929257"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.116"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.494"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.492"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1117\/12.56761"},{"key":"ref16","first-page":"147","article-title":"A multi-ring color fiducial system and an intensity-invariant detection method for scalable fiducial-tracking augmented reality","volume-title":"Proc. Int. Workshop Augmented Reality: Placing Artif. Objects Real Scenes","author":"Cho"},{"key":"ref17","first-page":"80","article-title":"The development of new coded targets for automated point identification and non-contact 3D surface measurements","volume":"5","author":"Knyaz","year":"1998","journal-title":"Int. Arch. Photogrammetry Remote Sens."},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/ISMAR.2002.1115065"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/CRV.2007.34"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/CRV.2011.13"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/APCHI.1998.704151"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1145\/354666.354667"},{"key":"ref23","first-page":"265","article-title":"Using camera-equipped mobile phones for interacting with real-world objects","volume":"176","author":"Rohs","year":"2004","journal-title":"Adv. Pervasive Comput."},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/IWAR.1999.803809"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2005.74"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2009.146"},{"key":"ref27","first-page":"59","article-title":"A lightweight ID-based extension for marker tracking systems","volume-title":"Proc. Eurogr. Symp. Virtual Environments Short Paper Proc.","author":"Flohr"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1016\/j.imavis.2018.05.004"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.164"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.169"},{"key":"ref31","article-title":"Faster R-CNN: Towards real-time object detection with region proposal networks","author":"Ren","year":"2015","journal-title":"arXiv:1506.01497"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.l007\/978-3-319-46448-0_2"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1145\/2733373.2806337"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1145\/2964284.2967300"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46478-7_38"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/TMI.2019.2895318"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00863"},{"key":"ref38","article-title":"MobileNets: Efficient convolutional neural networks for mobile vision applications","author":"Howard","year":"2017","journal-title":"arXiv:1704.04861"},{"key":"ref39","article-title":"Very deep convolutional networks for large-scale image recognition","author":"Simonyan","year":"2014","journal-title":"arXiv:1409.1556"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206537"},{"key":"ref41","article-title":"OpenImages: A public dataset for large-scale multi-label and multi-class image classification","author":"Krasin","year":"2017"},{"key":"ref42","article-title":"Adam: A method for stochastic optimization","author":"Kingma","year":"2017","journal-title":"arXiv:1412.6980"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.5555\/3454287.3455008"}],"container-title":["IEEE Transactions on Pattern Analysis and Machine Intelligence"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/34\/4359286\/09773975.pdf?arnumber=9773975","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,22]],"date-time":"2024-01-22T22:38:43Z","timestamp":1705963123000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9773975\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"references-count":43,"URL":"https:\/\/doi.org\/10.1109\/tpami.2022.3174603","relation":{},"ISSN":["0162-8828","2160-9292","1939-3539"],"issn-type":[{"value":"0162-8828","type":"print"},{"value":"2160-9292","type":"electronic"},{"value":"1939-3539","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]}}}