{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,18]],"date-time":"2025-12-18T14:14:50Z","timestamp":1766067290508,"version":"3.40.3"},"publisher-location":"Cham","reference-count":36,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030687922"},{"type":"electronic","value":"9783030687939"}],"license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021]]},"DOI":"10.1007\/978-3-030-68793-9_33","type":"book-chapter","created":{"date-parts":[[2021,2,20]],"date-time":"2021-02-20T16:28:24Z","timestamp":1613838504000},"page":"451-463","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":8,"title":["VA2Mass: Towards the Fluid Filling Mass Estimation via Integration of Vision and Audio Learning"],"prefix":"10.1007","author":[{"given":"Qi","family":"Liu","sequence":"first","affiliation":[]},{"given":"Fan","family":"Feng","sequence":"additional","affiliation":[]},{"given":"Chuanlin","family":"Lan","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4808-2490","authenticated-orcid":false,"given":"Rosa H. M.","family":"Chan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,2,21]]},"reference":[{"key":"33_CR1","unstructured":"Xompero, R.A., Sanchez-Matilla,R.M., Cavallaro, A.: CORSMAL Containers Manipulation (1.0) [Data set]. https:\/\/doi.org\/10.17636\/101CORSMAL1"},{"key":"33_CR2","unstructured":"Abdel-Hamid, O., Mohamed, A.R., Jiang, H., Deng, L., Penn, G., Yu, D.: Convolutional neural networks for speech recognition. IEEE\/ACM Trans. Audio Speech Lang. Process. 22(10), 1533\u20131545 (2014)"},{"issue":"2","key":"33_CR3","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1109\/MRA.2020.2987186","volume":"27","author":"H Bae","year":"2020","unstructured":"Bae, H., et al.: Iros 2019 lifelong robotic vision: object recognition challenge [competitions]. IEEE Rob. Autom. Mag 27(2), 11\u201316 (2020)","journal-title":"IEEE Rob. Autom. Mag"},{"key":"33_CR4","doi-asserted-by":"crossref","unstructured":"Bhattacharyya, R., Floerkemeier, C., Sarma, S.: Rfid tag antenna based sensing: does your beverage glass need a refill? In: 2010 IEEE International Conference on RFID (IEEE RFID 2010), pp. 126\u2013133. IEEE (2010)","DOI":"10.1109\/RFID.2010.5467235"},{"key":"33_CR5","unstructured":"Bochkovskiy, A., Wang, C.Y., Liao, H.Y.M.: Yolov4: optimal speed and accuracy of object detection. arXiv preprint arXiv:2004.10934 (2020)"},{"key":"33_CR6","doi-asserted-by":"crossref","unstructured":"Brandi, S., Kroemer, O., Peters, J.: Generalizing pouring actions between objects using warped parameters. In: 2014 IEEE-RAS International Conference on Humanoid Robots, pp. 616\u2013621. IEEE (2014)","DOI":"10.1109\/HUMANOIDS.2014.7041426"},{"key":"33_CR7","unstructured":"Clarke, S., Rhodes, T., Atkeson, C.G., Kroemer, O.: Learning audio feedback for estimating amount and flow of granular material. Proc. Mach. Learn. Res. 87 (2018)"},{"key":"33_CR8","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: Imagenet: a large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248\u2013255. IEEE (2009)","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"33_CR9","series-title":"Advances in Intelligent Systems and Computing","doi-asserted-by":"publisher","first-page":"210","DOI":"10.1007\/978-3-030-01370-7_17","volume-title":"Intelligent Autonomous Systems 15","author":"C Do","year":"2019","unstructured":"Do, C., Burgard, W.: Accurate pouring with an autonomous robot using an RGB-D camera. In: Strand, M., Dillmann, R., Menegatti, E., Ghidoni, S. (eds.) IAS 2018. AISC, vol. 867, pp. 210\u2013221. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-01370-7_17"},{"key":"33_CR10","doi-asserted-by":"crossref","unstructured":"Do, C., Schubert, T., Burgard, W.: A probabilistic approach to liquid level detection in cups using an RGB-D camera. In: 2016 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 2075\u20132080. IEEE (2016)","DOI":"10.1109\/IROS.2016.7759326"},{"key":"33_CR11","unstructured":"Griffith, S., Sukhoy, V., Wegter, T., Stoytchev, A.: Object categorization in the sink: Learning behavior-grounded object categories with water. In: Proceedings of the 2012 ICRA Workshop on Semantic Perception, Mapping and Exploration. Citeseer (2012)"},{"key":"33_CR12","doi-asserted-by":"crossref","unstructured":"Gu, S., Holly, E., Lillicrap, T., Levine, S.: Deep reinforcement learning for robotic manipulation with asynchronous off-policy updates. In: 2017 IEEE International Conference on Robotics and Automation (ICRA), pp. 3389\u20133396. IEEE (2017)","DOI":"10.1109\/ICRA.2017.7989385"},{"key":"33_CR13","doi-asserted-by":"crossref","unstructured":"Huang, Y., Sun, Y.: Learning to pour. In: 2017 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 7005\u20137010. IEEE (2017)","DOI":"10.1109\/IROS.2017.8206626"},{"key":"33_CR14","series-title":"Lecture Notes in Electrical Engineering","doi-asserted-by":"publisher","first-page":"7","DOI":"10.1007\/978-4-431-55690-9_2","volume-title":"Haptic Interaction","author":"S Ikeno","year":"2015","unstructured":"Ikeno, S., Watanabe, R., Okazaki, R., Hachisu, T., Sato, M., Kajimoto, H.: Change in the amount poured as a result of vibration when pouring a liquid. In: Kajimoto, H., Ando, H., Kyung, K.-U. (eds.) Haptic Interaction. LNEE, vol. 277, pp. 7\u201311. Springer, Tokyo (2015). https:\/\/doi.org\/10.1007\/978-4-431-55690-9_2"},{"key":"33_CR15","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"issue":"7553","key":"33_CR16","doi-asserted-by":"publisher","first-page":"436","DOI":"10.1038\/nature14539","volume":"521","author":"Y LeCun","year":"2015","unstructured":"LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521(7553), 436\u2013444 (2015)","journal-title":"Nature"},{"issue":"4\u20135","key":"33_CR17","doi-asserted-by":"publisher","first-page":"705","DOI":"10.1177\/0278364914549607","volume":"34","author":"I Lenz","year":"2015","unstructured":"Lenz, I., Lee, H., Saxena, A.: Deep learning for detecting robotic grasps. Int. J. Rob. Res. 34(4\u20135), 705\u2013724 (2015)","journal-title":"Int. J. Rob. Res."},{"key":"33_CR18","doi-asserted-by":"crossref","unstructured":"Liang, H., et al.: Making sense of audio vibration for liquid height estimation in robotic pouring. arXiv preprint arXiv:1903.00650 (2019)","DOI":"10.1109\/IROS40897.2019.8968303"},{"key":"33_CR19","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"740","DOI":"10.1007\/978-3-319-10602-1_48","volume-title":"Computer Vision \u2013 ECCV 2014","author":"TY Lin","year":"2014","unstructured":"Lin, T.Y., et al.: Microsoft COCO: common objects in context. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8693, pp. 740\u2013755. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-10602-1_48"},{"key":"33_CR20","doi-asserted-by":"crossref","unstructured":"McFee, B., et al.: librosa: audio and music signal analysis in python. In: Proceedings of the 14th Python in Science Conference, vol. 8, pp. 18\u201325 (2015)","DOI":"10.25080\/Majora-7b98e3ed-003"},{"key":"33_CR21","doi-asserted-by":"crossref","unstructured":"Muhlig, M., Gienger, M., Hellbach, S., Steil, J.J., Goerick, C.: Task-level imitation learning using variance-based movement optimization. In: 2009 IEEE International Conference on Robotics and Automation, pp. 1177\u20131184. IEEE (2009)","DOI":"10.1109\/ROBOT.2009.5152439"},{"key":"33_CR22","unstructured":"Nair, A., Bahl, S., Khazatsky, A., Pong, V., Berseth, G., Levine, S.: Contextual imagined goals for self-supervised robotic learning. In: Conference on Robot Learning, pp. 530\u2013539. PMLR (2020)"},{"key":"33_CR23","doi-asserted-by":"crossref","unstructured":"Pastor, P., Hoffmann, H., Asfour, T., Schaal, S.: Learning and generalization of motor skills by learning from demonstration. In: 2009 IEEE International Conference on Robotics and Automation, pp. 763\u2013768. IEEE (2009)","DOI":"10.1109\/ROBOT.2009.5152385"},{"key":"33_CR24","unstructured":"Paszke, A., et al.: Pytorch: an imperative style, high-performance deep learning library. In: Neural Information Processing Systems (NeurIPS), pp. 8024\u20138035 (2019)"},{"key":"33_CR25","doi-asserted-by":"crossref","unstructured":"Paulius, D., Huang, Y., Milton, R., Buchanan, W.D., Sam, J., Sun, Y.: Functional object-oriented network for manipulation learning. In: 2016 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 2655\u20132662. IEEE (2016)","DOI":"10.1109\/IROS.2016.7759413"},{"key":"33_CR26","doi-asserted-by":"crossref","unstructured":"Paulius, D., Jelodar, A.B., Sun, Y.: Functional object-oriented network: Construction & expansion. In: 2018 IEEE International Conference on Robotics and Automation (ICRA), pp. 1\u20137. IEEE (2018)","DOI":"10.1109\/ICRA.2018.8460200"},{"issue":"16","key":"33_CR27","doi-asserted-by":"publisher","first-page":"821","DOI":"10.1080\/01691864.2017.1365009","volume":"31","author":"HA Pierson","year":"2017","unstructured":"Pierson, H.A., Gashler, M.S.: Deep learning in robotics: a review of recent research. Adv. Rob. 31(16), 821\u2013835 (2017)","journal-title":"Adv. Rob."},{"key":"33_CR28","unstructured":"Pithadiya, K.J., Modi, C.K., Chauhan, J.D.: Selecting the most favourable edge detection technique for liquid level inspection in bottles. Int. J. Comput. Inf. Syst. Ind. Manag. Appl. (IJCISIM) ISSN, 2150\u20137988 (2011)"},{"key":"33_CR29","doi-asserted-by":"crossref","unstructured":"Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: unified, real-time object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 779\u2013788 (2016)","DOI":"10.1109\/CVPR.2016.91"},{"key":"33_CR30","doi-asserted-by":"crossref","unstructured":"Rozo, L., Jim\u00e9nez, P., Torras, C.: Force-based robot learning of pouring skills using parametric hidden Markov models. In: 9th International Workshop on Robot Motion and Control, pp. 227\u2013232. IEEE (2013)","DOI":"10.1109\/RoMoCo.2013.6614613"},{"key":"33_CR31","doi-asserted-by":"crossref","unstructured":"Saal, H.P., Ting, J.A., Vijayakumar, S.: Active estimation of object dynamics parameters with tactile sensors. In: 2010 IEEE\/RSJ International Conference on Intelligent Robots and Systems, pp. 916\u2013921. IEEE (2010)","DOI":"10.1109\/IROS.2010.5649191"},{"issue":"2","key":"33_CR32","doi-asserted-by":"publisher","first-page":"1642","DOI":"10.1109\/LRA.2020.2969200","volume":"5","author":"R Sanchez-Matilla","year":"2020","unstructured":"Sanchez-Matilla, R., et al.: Benchmark for human-to-robot handovers of unseen containers with unknown filling. IEEE Rob. Autom. Lett. 5(2), 1642\u20131649 (2020)","journal-title":"IEEE Rob. Autom. Lett."},{"key":"33_CR33","doi-asserted-by":"crossref","unstructured":"She, Q., et al.: Openloris-object: a robotic vision dataset and benchmark for lifelong deep learning. In: 2020 IEEE International Conference on Robotics and Automation (ICRA), pp. 4767\u20134773. IEEE (2020)","DOI":"10.1109\/ICRA40945.2020.9196887"},{"key":"33_CR34","doi-asserted-by":"crossref","unstructured":"Shi, X., et al.: Are we ready for service robots? the openloris-scene datasets for lifelong slam. In: 2020 IEEE International Conference on Robotics and Automation (ICRA), pp. 3139\u20133145. IEEE (2020)","DOI":"10.1109\/ICRA40945.2020.9196638"},{"key":"33_CR35","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. In: ICLR, vol. 2015 (2015)"},{"issue":"2","key":"33_CR36","doi-asserted-by":"publisher","first-page":"397","DOI":"10.1109\/LRA.2016.2633383","volume":"2","author":"PC Yang","year":"2016","unstructured":"Yang, P.C., Sasaki, K., Suzuki, K., Kase, K., Sugano, S., Ogata, T.: Repeatable folding task by humanoid robot worker using deep learning. IEEE Rob. Autom. Lett 2(2), 397\u2013403 (2016)","journal-title":"IEEE Rob. Autom. Lett"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition. ICPR International Workshops and Challenges"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-68793-9_33","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,2,20]],"date-time":"2021-02-20T17:16:27Z","timestamp":1613841387000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-030-68793-9_33"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"ISBN":["9783030687922","9783030687939"],"references-count":36,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-68793-9_33","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2021]]},"assertion":[{"value":"21 February 2021","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICPR","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Pattern Recognition","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2021","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"10 January 2021","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"11 January 2021","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ICPR2020","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.icpr2020.it\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}