{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,5,13]],"date-time":"2025-05-13T21:56:29Z","timestamp":1747173389423,"version":"3.40.5"},"reference-count":9,"publisher":"Cambridge University Press (CUP)","issue":"10","license":[{"start":{"date-parts":[[2020,10,12]],"date-time":"2020-10-12T00:00:00Z","timestamp":1602460800000},"content-version":"unspecified","delay-in-days":11,"URL":"https:\/\/www.cambridge.org\/core\/terms"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Robotica"],"published-print":{"date-parts":[[2020,10]]},"abstract":"<jats:p>Human\u2013robot interaction (HRI) is one of the most rapidly growing research fields in robotics and promising for the future of robotics technology. Despite the fact that numerous significant research results in HRI have been presented during the last years, there are still challenges in several critical topics of HRI, which could be summarized as: (i) collision and safety, (ii) virtual guides, (iii) cooperative manipulation, (iv) teleoperation and haptic interfaces, and (v) learning by observation or demonstration. In physical HRI research, the complementarity of the human and the robot capabilities is carefully considered for the advancement of their cooperation in a safe manner. New advanced control systems should be developed so the robot will acquire the ability to adapt easily to the human intentions and to the given task. The possible applications requiring co-manipulation are cooperative transportation of bulky and heavy objects, manufacturing processes such as assembly and surgery.<\/jats:p>","DOI":"10.1017\/s0263574720000946","type":"journal-article","created":{"date-parts":[[2020,10,12]],"date-time":"2020-10-12T05:38:57Z","timestamp":1602481137000},"page":"1715-1716","source":"Crossref","is-referenced-by-count":4,"title":["Special Issue on Human\u2013Robot Interaction (HRI)"],"prefix":"10.1017","volume":"38","author":[{"given":"Nikos","family":"Aspragathos","sequence":"first","affiliation":[]},{"given":"Vassilis","family":"Moulianitis","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3051-7321","authenticated-orcid":false,"given":"Panagiotis","family":"Koustoumpardis","sequence":"additional","affiliation":[]}],"member":"56","published-online":{"date-parts":[[2020,10,12]]},"reference":[{"key":"S0263574720000946_ref9","first-page":"1880","article-title":"Using a redundant user interface in teleoperated surgical systems for task performance enhancement","author":"Torabi","journal-title":"Robotica"},{"key":"S0263574720000946_ref7","doi-asserted-by":"crossref","first-page":"1842","DOI":"10.1017\/S0263574720000958","article-title":"Physical human-robot cooperation based on robust motion intention estimation","volume":"38","author":"Alevizos","year":"2020","journal-title":"Robotica"},{"key":"S0263574720000946_ref2","doi-asserted-by":"crossref","first-page":"1737","DOI":"10.1017\/S0263574719000985","article-title":"Neural network design for manipulator collision detection based only on the joint position sensors","volume":"38","author":"Sharkawy","year":"2020","journal-title":"Robotica"},{"key":"S0263574720000946_ref1","doi-asserted-by":"crossref","first-page":"1717","DOI":"10.1017\/S0263574719001425","article-title":"Collision detection on industrial robots in repetitive tasks using modified dynamic time warping","volume":"38","author":"Gordi\u0107","year":"2020","journal-title":"Robotica"},{"key":"S0263574720000946_ref3","doi-asserted-by":"crossref","first-page":"1756","DOI":"10.1017\/S0263574719001619","article-title":"An online trajectory generator on SE (3) for human\u2013robot collaboration","volume":"38","author":"Huber","year":"2020","journal-title":"Robotica"},{"key":"S0263574720000946_ref6","doi-asserted-by":"crossref","first-page":"1824","DOI":"10.1017\/S0263574719001437","article-title":"Kinesthetic guidance utilizing DMP synchronization and assistive virtual fixtures for progressive automation","volume":"38","author":"Papageorgiou","year":"2020","journal-title":"Robotica"},{"key":"S0263574720000946_ref8","doi-asserted-by":"crossref","first-page":"1867","DOI":"10.1017\/S0263574719001449","article-title":"Speed adaptation in learning from demonstration through latent space formulation","volume":"38","author":"Koskinopoulou","year":"2020","journal-title":"Robotica"},{"key":"S0263574720000946_ref5","doi-asserted-by":"crossref","first-page":"1807","DOI":"10.1017\/S0263574719000973","article-title":"Unified virtual guides framework for path tracking tasks","volume":"38","author":"\u017dlajpah","year":"2020","journal-title":"Robotica"},{"key":"S0263574720000946_ref4","doi-asserted-by":"crossref","first-page":"1778","DOI":"10.1017\/S0263574720000016","article-title":"Toward an intuitive and iterative 6D virtual guide programming framework for assisted human\u2013robot comanipulation","volume":"38","author":"Restrepo","year":"2020","journal-title":"Robotica"}],"container-title":["Robotica"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.cambridge.org\/core\/services\/aop-cambridge-core\/content\/view\/S0263574720000946","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,4,8]],"date-time":"2021-04-08T00:32:38Z","timestamp":1617841958000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.cambridge.org\/core\/product\/identifier\/S0263574720000946\/type\/journal_article"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,10]]},"references-count":9,"journal-issue":{"issue":"10","published-print":{"date-parts":[[2020,10]]}},"alternative-id":["S0263574720000946"],"URL":"https:\/\/doi.org\/10.1017\/s0263574720000946","relation":{},"ISSN":["0263-5747","1469-8668"],"issn-type":[{"type":"print","value":"0263-5747"},{"type":"electronic","value":"1469-8668"}],"subject":[],"published":{"date-parts":[[2020,10]]}}}