<?xml version="1.0" encoding="UTF-8"?><?xml-stylesheet type="text/xsl" href="static/style.xsl"?><OAI-PMH xmlns="http://www.openarchives.org/OAI/2.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/ http://www.openarchives.org/OAI/2.0/OAI-PMH.xsd"><responseDate>2026-03-10T06:15:24Z</responseDate><request verb="GetRecord" identifier="oai:uvadoc.uva.es:10324/69775" metadataPrefix="marc">https://uvadoc.uva.es/oai/request</request><GetRecord><record><header><identifier>oai:uvadoc.uva.es:10324/69775</identifier><datestamp>2025-02-20T09:27:42Z</datestamp><setSpec>com_10324_966</setSpec><setSpec>com_10324_952</setSpec><setSpec>com_10324_894</setSpec><setSpec>col_10324_967</setSpec></header><metadata><record xmlns="http://www.loc.gov/MARC21/slim" xmlns:doc="http://www.lyncode.com/xoai" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:dcterms="http://purl.org/dc/terms/" xsi:schemaLocation="http://www.loc.gov/MARC21/slim http://www.loc.gov/standards/marcxml/schema/MARC21slim.xsd">
<leader>00925njm 22002777a 4500</leader>
<datafield tag="042" ind1=" " ind2=" ">
<subfield code="a">dc</subfield>
</datafield>
<datafield tag="720" ind1=" " ind2=" ">
<subfield code="a">Calderón Sesmero, Raúl</subfield>
<subfield code="e">author</subfield>
</datafield>
<datafield tag="720" ind1=" " ind2=" ">
<subfield code="a">Duque Domingo, Jaime</subfield>
<subfield code="e">author</subfield>
</datafield>
<datafield tag="720" ind1=" " ind2=" ">
<subfield code="a">Gómez García-Bermejo, Jaime</subfield>
<subfield code="e">author</subfield>
</datafield>
<datafield tag="720" ind1=" " ind2=" ">
<subfield code="a">Zalama Casanova, Eduardo</subfield>
<subfield code="e">author</subfield>
</datafield>
<datafield tag="260" ind1=" " ind2=" ">
<subfield code="c">2024</subfield>
</datafield>
<datafield tag="520" ind1=" " ind2=" ">
<subfield code="a">The growing demand for projects with collaborative robots, known as “cobots”, underlines the need to efficiently address the execution of tasks with speed and flexibility, without neglecting safety in human–robot interaction. In general terms, this practice requires knowledge of robotics programming and skill in the use of hardware. The proposed solution consists of a mixed reality (MR) application integrated into a mixed reality head-mounted device (HMD) that accelerates the process of programming the complex manoeuvres of a cobot. This advancement is achieved through voice and gesture recognition, in addition to the use of digital panels. This allows any user, regardless of his or her robotics experience, to work more efficiently. The Robot Operating System (ROS) platform monitors the cobot and manages the transfer of data between the two. The system uses QR (Quick Response) codes to establish a precise frame of reference. This solution has proven its applicability in industrial processes, by automating manoeuvres and receiving positive feedback from users who have evaluated its performance. This solution promises to revolutionize the programming and operation of cobots, and pave the way for efficient and accessible collaborative robotics.</subfield>
</datafield>
<datafield tag="024" ind2=" " ind1="8">
<subfield code="a">Electronics, 2024, Vol. 13, Nº. 3, 571</subfield>
</datafield>
<datafield tag="024" ind2=" " ind1="8">
<subfield code="a">2079-9292</subfield>
</datafield>
<datafield tag="024" ind2=" " ind1="8">
<subfield code="a">https://uvadoc.uva.es/handle/10324/69775</subfield>
</datafield>
<datafield tag="024" ind2=" " ind1="8">
<subfield code="a">10.3390/electronics13030571</subfield>
</datafield>
<datafield tag="024" ind2=" " ind1="8">
<subfield code="a">571</subfield>
</datafield>
<datafield tag="024" ind2=" " ind1="8">
<subfield code="a">3</subfield>
</datafield>
<datafield tag="024" ind2=" " ind1="8">
<subfield code="a">Electronics</subfield>
</datafield>
<datafield tag="024" ind2=" " ind1="8">
<subfield code="a">13</subfield>
</datafield>
<datafield tag="024" ind2=" " ind1="8">
<subfield code="a">2079-9292</subfield>
</datafield>
<datafield ind1=" " ind2=" " tag="653">
<subfield code="a">Mixed reality</subfield>
</datafield>
<datafield ind1=" " ind2=" " tag="653">
<subfield code="a">Human-robot interaction</subfield>
</datafield>
<datafield ind1=" " ind2=" " tag="653">
<subfield code="a">Collaborative robots</subfield>
</datafield>
<datafield ind1=" " ind2=" " tag="653">
<subfield code="a">Robots</subfield>
</datafield>
<datafield ind1=" " ind2=" " tag="653">
<subfield code="a">Robotics and Automation</subfield>
</datafield>
<datafield ind1=" " ind2=" " tag="653">
<subfield code="a">Robótica</subfield>
</datafield>
<datafield ind1=" " ind2=" " tag="653">
<subfield code="a">Automatización</subfield>
</datafield>
<datafield ind1=" " ind2=" " tag="653">
<subfield code="a">Pattern recognition</subfield>
</datafield>
<datafield ind1=" " ind2=" " tag="653">
<subfield code="a">Reconocimiento de formas (Informática)</subfield>
</datafield>
<datafield ind1=" " ind2=" " tag="653">
<subfield code="a">Artificial intelligence</subfield>
</datafield>
<datafield ind1=" " ind2=" " tag="653">
<subfield code="a">Computer networks</subfield>
</datafield>
<datafield ind1=" " ind2=" " tag="653">
<subfield code="a">Redes de ordenadores</subfield>
</datafield>
<datafield ind1=" " ind2=" " tag="653">
<subfield code="a">Computer science</subfield>
</datafield>
<datafield tag="245" ind1="0" ind2="0">
<subfield code="a">Development of a human–robot interface for cobot trajectory planning using mixed reality</subfield>
</datafield>
</record></metadata></record></GetRecord></OAI-PMH>