{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,5]],"date-time":"2026-03-05T15:35:57Z","timestamp":1772724957051,"version":"3.50.1"},"reference-count":51,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"name":"Shanghai Science and Technology Program Project","award":["21JC1400600"],"award-info":[{"award-number":["21JC1400600"]}]},{"name":"Singapore Ministry of Education (MOE) Academic Research Fund"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Multimedia"],"published-print":{"date-parts":[[2025]]},"DOI":"10.1109\/tmm.2024.3371212","type":"journal-article","created":{"date-parts":[[2024,2,28]],"date-time":"2024-02-28T18:56:14Z","timestamp":1709146574000},"page":"2724-2733","source":"Crossref","is-referenced-by-count":10,"title":["From Canteen Food to Daily Meals: Generalizing Food Recognition to More Practical Scenarios"],"prefix":"10.1109","volume":"27","author":[{"ORCID":"https:\/\/orcid.org\/0009-0007-6648-5206","authenticated-orcid":false,"given":"Guoshan","family":"Liu","sequence":"first","affiliation":[{"name":"Fudan Vision and Learning Laboratory (FVL), Fudan University, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8679-2242","authenticated-orcid":false,"given":"Yang","family":"Jiao","sequence":"additional","affiliation":[{"name":"Fudan Vision and Learning Laboratory (FVL), Fudan University, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3148-264X","authenticated-orcid":false,"given":"Jingjing","family":"Chen","sequence":"additional","affiliation":[{"name":"Fudan Vision and Learning Laboratory (FVL), Fudan University, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9213-2611","authenticated-orcid":false,"given":"Bin","family":"Zhu","sequence":"additional","affiliation":[{"name":"Singapore Management University, Singapore"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1907-8567","authenticated-orcid":false,"given":"Yu-Gang","family":"Jiang","sequence":"additional","affiliation":[{"name":"Fudan Vision and Learning Laboratory (FVL), Fudan University, Shanghai, China"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2023.3237871"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00879"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/ICIP46576.2022.9898001"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1145\/2964284.2964315"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-10599-4_29"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2020.3045639"},{"key":"ref7","article-title":"Deep domain confusion: Maximizing for domain invariance","volume-title":"CoRR","author":"Tzeng","year":"2014"},{"key":"ref8","first-page":"97","article-title":"Learning transferable features with deep adaptation networks","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Long","year":"2015"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-49409-8_35"},{"key":"ref10","first-page":"343","article-title":"Domain separation networks","volume-title":"Proc. Annu. Conf. Neural Inf. Process. Syst.","author":"Bousmalis","year":"2016"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2021.12.089"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00127"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2020.2988928"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM.2019.00088"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00503"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2023.109379"},{"key":"ref17","article-title":"Dual adversarial domain adaptation","author":"Du","year":"2020","journal-title":"CoRR"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1145\/3240508.3240512"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2013.274"},{"key":"ref20","article-title":"Foodx-251: A dataset for fine-grained food classification","author":"Kaur","year":"2019","journal-title":"CoRR"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.327"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1145\/2407746.2407775"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/WACV.2009.5403087"},{"key":"ref24","first-page":"2017","article-title":"Multiple-food recognition considering co-occurrence employing manifold ranking","volume-title":"Proc. IEEE Int. Conf. Pattern Recognit.","author":"Matsuda","year":"2012"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/IDAP.2019.8875946"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/TAI.2021.3108126"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1145\/3394171.3413809"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.5555\/2946645.2946704"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v30i1.10306"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-13560-1_76"},{"key":"ref31","first-page":"2208","article-title":"Deep transfer learning with joint adaptation networks","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Long","year":"2017"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1145\/3422622"},{"key":"ref33","first-page":"1180","article-title":"Unsupervised domain adaptation by backpropagation","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Ganin","year":"2015"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00258"},{"key":"ref35","first-page":"1989","article-title":"CyCADA: Cycle-consistent adversarial domain adaptation","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Hoffman","year":"2018"},{"key":"ref36","article-title":"Deep transfer network: Unsupervised domain adaptation","author":"Zhang","year":"2015","journal-title":"CoRR"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v32i1.11767"},{"key":"ref38","first-page":"896","article-title":"Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks","volume-title":"Proc. Workshop Challenges Representation Learn.","author":"Lee","year":"2013"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00912"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1137\/1.9781611976236.53"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-020-01395-y"},{"issue":"1","key":"ref42","first-page":"723","article-title":"A kernel two-sample test","volume":"13","author":"Gretton","year":"2012","journal-title":"J. Mach. Learn. Res."},{"key":"ref43","article-title":"Rethinking softmax with cross-entropy: Neural network classifier as mutual information estimator","author":"Qin","year":"2019","journal-title":"CoRR"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.572"},{"key":"ref45","article-title":"VisDA: The visual domain adaptation challenge","author":"Peng","year":"2017","journal-title":"CoRR"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19830-4_6"},{"key":"ref47","article-title":"CDTrans: Cross-domain transformer for unsupervised domain adaptation","volume-title":"Proc. 10th Int. Conf. Learn. Representations","author":"Xu","year":"2022"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-015-0816-y"},{"key":"ref49","article-title":"ADAM: A method for stochastic optimization","volume-title":"Proc. 3rd Int. Conf. Learn. Representations","author":"Kingma","year":"2015"},{"key":"ref50","article-title":"An image is worth 16x16 words: Transformers for image recognition at scale","volume-title":"Proc. 9th Int. Conf. Learn. Representations","author":"Dosovitskiy","year":"2021"},{"key":"ref51","first-page":"10347","article-title":"Training data-efficient image transformers & distillation through attention","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Touvron","year":"2021"}],"container-title":["IEEE Transactions on Multimedia"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/6046\/10844992\/10453509.pdf?arnumber=10453509","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,5,27]],"date-time":"2025-05-27T04:31:20Z","timestamp":1748320280000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10453509\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"references-count":51,"URL":"https:\/\/doi.org\/10.1109\/tmm.2024.3371212","relation":{},"ISSN":["1520-9210","1941-0077"],"issn-type":[{"value":"1520-9210","type":"print"},{"value":"1941-0077","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]}}}