{"id":76,"date":"2017-09-10T05:46:36","date_gmt":"2017-09-10T05:46:36","guid":{"rendered":"https:\/\/ryma.cinvestav.mx\/garechav\/?page_id=76"},"modified":"2017-09-11T15:36:06","modified_gmt":"2017-09-11T15:36:06","slug":"publicaciones","status":"publish","type":"page","link":"https:\/\/ryma.cinvestav.mx\/garechav\/publicaciones\/","title":{"rendered":"Publicaciones"},"content":{"rendered":"<p>[et_pb_section bb_built=\u00bb1&#8243; fullwidth=\u00bbon\u00bb specialty=\u00bboff\u00bb background_image=\u00bbhttps:\/\/ryma.cinvestav.mx\/garechav\/wp-content\/uploads\/sites\/9\/2017\/09\/IMAGEN-CAMPA\u00d1A1.jpg\u00bb _builder_version=\u00bb3.0.72&#8243; background_color_gradient_start=\u00bbrgba(0,0,0,0.56)\u00bb background_color_gradient_end=\u00bb#ffffff\u00bb background_color_gradient_direction=\u00bb122deg\u00bb background_color_gradient_direction_radial=\u00bbleft\u00bb background_color_gradient_start_position=\u00bb9%\u00bb locked=\u00bbon\u00bb global_module=\u00bb98&#8243;][et_pb_fullwidth_header admin_label=\u00bbNombre del Investigador\u00bb global_parent=\u00bb98&#8243; title=\u00bbDR. GUSTAVO ARECHAVALETA\u00bb subhead=\u00bbProfesor Investigador\u00bb background_layout=\u00bbdark\u00bb text_orientation=\u00bbleft\u00bb header_fullscreen=\u00bboff\u00bb header_scroll_down=\u00bboff\u00bb content_orientation=\u00bbcenter\u00bb image_orientation=\u00bbcenter\u00bb title_font=\u00bb|on|||\u00bb subhead_font=\u00bb|on|||\u00bb custom_button_one=\u00bboff\u00bb button_one_letter_spacing=\u00bb0&#8243; button_one_icon_placement=\u00bbright\u00bb button_one_letter_spacing_hover=\u00bb0&#8243; custom_button_two=\u00bboff\u00bb button_two_letter_spacing=\u00bb0&#8243; button_two_icon_placement=\u00bbright\u00bb button_two_letter_spacing_hover=\u00bb0&#8243; subhead_font_size=\u00bb20px\u00bb _builder_version=\u00bb3.0.72&#8243; background_overlay_color=\u00bbrgba(0,0,0,0.3)\u00bb parent_locked=\u00bbon\u00bb \/][et_pb_fullwidth_menu global_parent=\u00bb98&#8243; menu_id=\u00bb6&#8243; background_layout=\u00bblight\u00bb text_orientation=\u00bbleft\u00bb submenu_direction=\u00bbdownwards\u00bb fullwidth_menu=\u00bboff\u00bb active_link_color=\u00bbrgba(131,0,233,0.81)\u00bb menu_font=\u00bb|on|||\u00bb _builder_version=\u00bb3.0.72&#8243; background_color_gradient_start=\u00bb#009f93&#8243; background_color_gradient_end=\u00bbrgba(0,0,0,0.27)\u00bb background_color_gradient_start_position=\u00bb20%\u00bb background_color=\u00bb#d3d3d3&#8243; parent_locked=\u00bbon\u00bb \/][\/et_pb_section][et_pb_section bb_built=\u00bb1&#8243; _builder_version=\u00bb3.0.72&#8243; custom_css_main_element=\u00bbbox-shadow: inset 0px 3px 2px rgba(50, 50, 50, 0.75);\u00bb locked=\u00bboff\u00bb][et_pb_row parent_locked=\u00bboff\u00bb background_position=\u00bbtop_left\u00bb background_repeat=\u00bbrepeat\u00bb background_size=\u00bbinitial\u00bb][et_pb_column type=\u00bb4_4&#8243;][et_pb_text _builder_version=\u00bb3.0.72&#8243; background_layout=\u00bblight\u00bb text_orientation=\u00bbleft\u00bb border_style=\u00bbsolid\u00bb parent_locked=\u00bboff\u00bb]<\/p>\n<hr \/>\n<div class='et-box et-shadow'>\n\t\t\t\t\t<div class='et-box-content'><h2><strong>PUBLICACIONES<\/strong><\/h2><\/div><\/div>\n<p>Para ver las publicaciones de todo Rob\u00f3tica y Manufactura Avanzada, ver:\u00a0 <strong><a href=\"https:\/\/ryma.cinvestav.mx\/investigacion\/publicaciones\/\">Publicaciones RYMA<\/a><\/strong><\/p>\n<p><strong><div class=\"teachpress_pub_list\"><form name=\"tppublistform\" method=\"get\"><a name=\"tppubs\" id=\"tppubs\"><\/a><div class=\"teachpress_filter\"><select class=\"default\" name=\"yr\" id=\"yr\" tabindex=\"2\" onchange=\"teachpress_jumpMenu('parent',this, 'https:\/\/ryma.cinvestav.mx\/garechav\/publicaciones\/?')\">\r\n                   <option value=\"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=#tppubs\">Todos los a\u00f1os<\/option>\r\n                   <option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2020#tppubs\" >2020<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2019#tppubs\" >2019<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2018#tppubs\" >2018<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2017#tppubs\" >2017<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2016#tppubs\" >2016<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2015#tppubs\" >2015<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2014#tppubs\" >2014<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2013#tppubs\" >2013<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2012#tppubs\" >2012<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2011#tppubs\" >2011<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2010#tppubs\" >2010<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2009#tppubs\" >2009<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2008#tppubs\" >2008<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2007#tppubs\" >2007<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2006#tppubs\" >2006<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2005#tppubs\" >2005<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2004#tppubs\" >2004<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2003#tppubs\" >2003<\/option>\r\n                <\/select><select class=\"default\" name=\"type\" id=\"type\" tabindex=\"3\" onchange=\"teachpress_jumpMenu('parent',this, 'https:\/\/ryma.cinvestav.mx\/garechav\/publicaciones\/?')\">\r\n                   <option value=\"tgid=&amp;yr=&amp;auth=&amp;usr=&amp;type=#tppubs\">Todas las tipolog\u00edas<\/option>\r\n                   <option value = \"tgid=&amp;yr=&amp;auth=&amp;usr=&amp;type=article#tppubs\" >Art\u00edculos de revista<\/option><option value = \"tgid=&amp;yr=&amp;auth=&amp;usr=&amp;type=conference#tppubs\" >Conferencias<\/option><option value = \"tgid=&amp;yr=&amp;auth=&amp;usr=&amp;type=inbook#tppubs\" >Cap\u00edtulos de libros<\/option><option value = \"tgid=&amp;yr=&amp;auth=&amp;usr=&amp;type=inproceedings#tppubs\" >Proceedings Articles<\/option><option value = \"tgid=&amp;yr=&amp;auth=&amp;usr=&amp;type=proceedings#tppubs\" >Actas de congresos<\/option>\r\n                <\/select><select class=\"default\" name=\"usr\" id=\"usr\" tabindex=\"6\" onchange=\"teachpress_jumpMenu('parent',this, 'https:\/\/ryma.cinvestav.mx\/garechav\/publicaciones\/?')\">\r\n                   <option value=\"tgid=&amp;yr=&amp;type=&amp;auth=&amp;usr=#tppubs\">Todos los usuarios<\/option>\r\n                   <option value = \"tgid=&amp;yr=&amp;type=&amp;auth=&amp;usr=12#tppubs\" >mcastelan<\/option>\r\n                <\/select><\/div><\/form><div class=\"teachpress_publication_list\"><h3 class=\"tp_h3\" id=\"tp_h3_2017\">2017<\/h3><h3 class=\"tp_h3\" id=\"tp_h3_article\">Art\u00edculos de revista<\/h3><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Arechavaleta, Gustavo;  Morales-Diaz, America B.;  Perez-Villeda, Hector Manuel;  Castelan, Mario<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('2','tp_links')\" style=\"cursor:pointer;\">Hierarchical Task-Based Control of Multirobot Systems With Terminal Attractors<\/a> <span class=\"tp_pub_type tp_  article\">Art\u00edculo de revista<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">En: <\/span><span class=\"tp_pub_additional_journal\">IEEE Transactions on Control Systems Technology, <\/span><span class=\"tp_pub_additional_volume\">vol. 25, <\/span><span class=\"tp_pub_additional_number\">no 1, <\/span><span class=\"tp_pub_additional_pages\">pp. 334 - 341, <\/span><span class=\"tp_pub_additional_year\">2017<\/span>, <span class=\"tp_pub_additional_issn\">ISSN: 1063-6536<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_2\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('2','tp_abstract')\" title=\"Mostrar resumen\" style=\"cursor:pointer;\">Resumen<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_2\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('2','tp_links')\" title=\"Mostrar enlaces y recursos\" style=\"cursor:pointer;\">Enlaces<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_2\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('2','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_2\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{7454708,<br \/>\r\ntitle = {Hierarchical Task-Based Control of Multirobot Systems With Terminal Attractors},<br \/>\r\nauthor = {Arechavaleta, Gustavo and Morales-Diaz, America B. and Perez-Villeda, Hector Manuel and Castelan, Mario },<br \/>\r\nurl = {http:\/\/ieeexplore.ieee.org\/abstract\/document\/7454708\/},<br \/>\r\ndoi = {10.1109\/TCST.2016.2549279},<br \/>\r\nissn = {1063-6536},<br \/>\r\nyear  = {2017},<br \/>\r\ndate = {2017-01-01},<br \/>\r\njournal = {IEEE Transactions on Control Systems Technology},<br \/>\r\nvolume = {25},<br \/>\r\nnumber = {1},<br \/>\r\npages = {334 - 341},<br \/>\r\nabstract = {This brief proposes a hierarchical control scheme based on the definition of a set of multirobot task functions. To deal with the inherent conflicts between tasks, a strict hierarchy is imposed on them. We present a novel scheme that copes with two main difficulties shared in standard task-based controllers: 1) to impose a desired time convergence of tasks and 2) to avoid discontinuous task transitions occurred when a task is inserted or removed in the hierarchical structure. As a result, continuous input references are generated for the low-level control of the group. The validation is achieved in simulation and by performing an experiment with wheeled mobile robots.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('2','tp_bibtex')\">Cerrar<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_2\" style=\"display:none;\"><div class=\"tp_abstract_entry\">This brief proposes a hierarchical control scheme based on the definition of a set of multirobot task functions. To deal with the inherent conflicts between tasks, a strict hierarchy is imposed on them. We present a novel scheme that copes with two main difficulties shared in standard task-based controllers: 1) to impose a desired time convergence of tasks and 2) to avoid discontinuous task transitions occurred when a task is inserted or removed in the hierarchical structure. As a result, continuous input references are generated for the low-level control of the group. The validation is achieved in simulation and by performing an experiment with wheeled mobile robots.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('2','tp_abstract')\">Cerrar<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_2\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/ieeexplore.ieee.org\/abstract\/document\/7454708\/\" title=\"http:\/\/ieeexplore.ieee.org\/abstract\/document\/7454708\/\" target=\"_blank\">http:\/\/ieeexplore.ieee.org\/abstract\/document\/7454708\/<\/a><\/li><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/10.1109\/TCST.2016.2549279\" title=\"DOI de seguimiento:10.1109\/TCST.2016.2549279\" target=\"_blank\">doi:10.1109\/TCST.2016.2549279<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('2','tp_links')\">Cerrar<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_2016\">2016<\/h3><h3 class=\"tp_h3\" id=\"tp_h3_article\">Art\u00edculos de revista<\/h3><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Mart\u00ednez-Gonz\u00e1lez, Pablo Arturo;  Castelan, Mario;  Arechavaleta, Gustavo<\/p><p class=\"tp_pub_title\">Vision Based Persistent Localization of a Humanoid Robot for Locomotion Tasks <span class=\"tp_pub_type tp_  article\">Art\u00edculo de revista<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">En: <\/span><span class=\"tp_pub_additional_year\">2016<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_159\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('159','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_159\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{Mart\\'{i}nez-Gonz\\'{a}lez2016b,<br \/>\r\ntitle = {Vision Based Persistent Localization of a Humanoid Robot for Locomotion Tasks},<br \/>\r\nauthor = {Mart\\'{i}nez-Gonz\\'{a}lez, Pablo Arturo and Castelan, Mario and Arechavaleta, Gustavo},<br \/>\r\nyear  = {2016},<br \/>\r\ndate = {2016-06-06},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('159','tp_bibtex')\">Cerrar<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Martinez-Gonzalez, Pablo;  Castelan, Mario;  Arechavaleta, Gustavo<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('1','tp_links')\" style=\"cursor:pointer;\">Vision based persistent localization of a humanoid robot for locomotion Tasks<\/a> <span class=\"tp_pub_type tp_  article\">Art\u00edculo de revista<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">En: <\/span><span class=\"tp_pub_additional_journal\">International Journal of Applied Mathematics and Computer Science, <\/span><span class=\"tp_pub_additional_volume\">vol. 26, <\/span><span class=\"tp_pub_additional_number\">no 3, <\/span><span class=\"tp_pub_additional_year\">2016<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_resource_link\"><a id=\"tp_links_sh_1\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('1','tp_links')\" title=\"Mostrar enlaces y recursos\" style=\"cursor:pointer;\">Enlaces<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_1\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('1','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_1\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{Mart\\`{i}nez-Gonz\\'{a}lez2016,<br \/>\r\ntitle = {Vision based persistent localization of a humanoid robot for locomotion Tasks},<br \/>\r\nauthor = {Martinez-Gonzalez, Pablo and Castelan, Mario and Arechavaleta, Gustavo },<br \/>\r\nurl = {https:\/\/drive.google.com\/file\/d\/0B-7dVUdTjeJUNGdXd0N6UWRvdk0\/view},<br \/>\r\nyear  = {2016},<br \/>\r\ndate = {2016-03-26},<br \/>\r\njournal = {International Journal of Applied Mathematics and Computer Science},<br \/>\r\nvolume = {26},<br \/>\r\nnumber = {3},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('1','tp_bibtex')\">Cerrar<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_1\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/drive.google.com\/file\/d\/0B-7dVUdTjeJUNGdXd0N6UWRvdk0\/view\" title=\"https:\/\/drive.google.com\/file\/d\/0B-7dVUdTjeJUNGdXd0N6UWRvdk0\/view\" target=\"_blank\">https:\/\/drive.google.com\/file\/d\/0B-7dVUdTjeJUNGdXd0N6UWRvdk0\/view<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('1','tp_links')\">Cerrar<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Delfin, Josafat;  Becerra, Hector M;  Arechavaleta, Gustavo<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('5','tp_links')\" style=\"cursor:pointer;\">Visual Servo Walking Control for Humanoids with Finite-time Convergence and Smooth Robot Velocities<\/a> <span class=\"tp_pub_type tp_  article\">Art\u00edculo de revista<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">En: <\/span><span class=\"tp_pub_additional_journal\">International Journal of Control, <\/span><span class=\"tp_pub_additional_volume\">vol. 89, <\/span><span class=\"tp_pub_additional_number\">no 7, <\/span><span class=\"tp_pub_additional_pages\">pp. 1342-1358, <\/span><span class=\"tp_pub_additional_year\">2016<\/span>, <span class=\"tp_pub_additional_issn\">ISSN: 1366-5820<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_5\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('5','tp_abstract')\" title=\"Mostrar resumen\" style=\"cursor:pointer;\">Resumen<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_5\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('5','tp_links')\" title=\"Mostrar enlaces y recursos\" style=\"cursor:pointer;\">Enlaces<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_5\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('5','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_5\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{Delfin2016,<br \/>\r\ntitle = {Visual Servo Walking Control for Humanoids with Finite-time Convergence and Smooth Robot Velocities},<br \/>\r\nauthor = {Delfin, Josafat and Becerra, Hector M and Arechavaleta, Gustavo },<br \/>\r\nurl = {http:\/\/www.tandfonline.com\/doi\/abs\/10.1080\/00207179.2015.1129558},<br \/>\r\ndoi = {http:\/\/dx.doi.org\/10.1080\/00207179.2015.1129558},<br \/>\r\nissn = {1366-5820},<br \/>\r\nyear  = {2016},<br \/>\r\ndate = {2016-01-10},<br \/>\r\njournal = {International Journal of Control},<br \/>\r\nvolume = {89},<br \/>\r\nnumber = {7},<br \/>\r\npages = {1342-1358},<br \/>\r\nabstract = {In this paper, we address the problem of humanoid locomotion guided from information of a monocular camera. The goal of the robot is to reach a desired location defined in terms of a target image, i.e., a positioning task. The proposed approach allows us to introduce a desired time to complete the positioning task, which is advantageous in contrast to the classical exponential convergence. In particular, finite-time convergence is achieved while generating smooth robot velocities and considering the omnidirectional waking capability of the robot. In addition, we propose a hierarchical task-based control scheme, which can simultaneously handle the visual positioning and the obstacle avoidance tasks without affecting the desired time of convergence. The controller is able to activate or inactivate the obstacle avoidance task without generating discontinuous velocity references while the humanoid is walking. Stability of the closed loop for the two task-based control is demonstrated theoretically even during the transitions between the tasks. The proposed approach is generic in the sense that different visual control schemes are supported. We evaluate a homography-based visual servoing for position-based and image-based modalities, as well as for eye-in-hand and eye-to-hand configurations. The experimental evaluation is performed with the humanoid robot NAO.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('5','tp_bibtex')\">Cerrar<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_5\" style=\"display:none;\"><div class=\"tp_abstract_entry\">In this paper, we address the problem of humanoid locomotion guided from information of a monocular camera. The goal of the robot is to reach a desired location defined in terms of a target image, i.e., a positioning task. The proposed approach allows us to introduce a desired time to complete the positioning task, which is advantageous in contrast to the classical exponential convergence. In particular, finite-time convergence is achieved while generating smooth robot velocities and considering the omnidirectional waking capability of the robot. In addition, we propose a hierarchical task-based control scheme, which can simultaneously handle the visual positioning and the obstacle avoidance tasks without affecting the desired time of convergence. The controller is able to activate or inactivate the obstacle avoidance task without generating discontinuous velocity references while the humanoid is walking. Stability of the closed loop for the two task-based control is demonstrated theoretically even during the transitions between the tasks. The proposed approach is generic in the sense that different visual control schemes are supported. We evaluate a homography-based visual servoing for position-based and image-based modalities, as well as for eye-in-hand and eye-to-hand configurations. The experimental evaluation is performed with the humanoid robot NAO.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('5','tp_abstract')\">Cerrar<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_5\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/www.tandfonline.com\/doi\/abs\/10.1080\/00207179.2015.1129558\" title=\"http:\/\/www.tandfonline.com\/doi\/abs\/10.1080\/00207179.2015.1129558\" target=\"_blank\">http:\/\/www.tandfonline.com\/doi\/abs\/10.1080\/00207179.2015.1129558<\/a><\/li><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/http:\/\/dx.doi.org\/10.1080\/00207179.2015.1129558\" title=\"DOI de seguimiento:http:\/\/dx.doi.org\/10.1080\/00207179.2015.1129558\" target=\"_blank\">doi:http:\/\/dx.doi.org\/10.1080\/00207179.2015.1129558<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('5','tp_links')\">Cerrar<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Delfin, Josafat;  Becerra, Hector M;  Arechavaleta, Gustavo<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('160','tp_links')\" style=\"cursor:pointer;\">Visual servo walking control for humanoids with finite-time convergence and smooth robot velocities<\/a> <span class=\"tp_pub_type tp_  article\">Art\u00edculo de revista<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">En: <\/span><span class=\"tp_pub_additional_journal\">International Journal of Control, <\/span><span class=\"tp_pub_additional_volume\">vol. 89, <\/span><span class=\"tp_pub_additional_number\">no 7, <\/span><span class=\"tp_pub_additional_pages\">pp. 1342-1358, <\/span><span class=\"tp_pub_additional_year\">2016<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_160\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('160','tp_abstract')\" title=\"Mostrar resumen\" style=\"cursor:pointer;\">Resumen<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_160\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('160','tp_links')\" title=\"Mostrar enlaces y recursos\" style=\"cursor:pointer;\">Enlaces<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_160\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('160','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_160\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{doi:10.1080\/00207179.2015.1129558,<br \/>\r\ntitle = {Visual servo walking control for humanoids with finite-time convergence and smooth robot velocities},<br \/>\r\nauthor = {Delfin, Josafat and Becerra, Hector M and Arechavaleta, Gustavo},<br \/>\r\nurl = {http:\/\/dx.doi.org\/10.1080\/00207179.2015.1129558},<br \/>\r\ndoi = {10.1080\/00207179.2015.1129558},<br \/>\r\nyear  = {2016},<br \/>\r\ndate = {2016-01-01},<br \/>\r\njournal = {International Journal of Control},<br \/>\r\nvolume = {89},<br \/>\r\nnumber = {7},<br \/>\r\npages = {1342-1358},<br \/>\r\nabstract = {ABSTRACTIn this paper, we address the problem of humanoid locomotion guided from information of a monocular camera. The goal of the robot is to reach a desired location defined in terms of a target image, i.e., a positioning task. The proposed approach allows us to introduce a desired time to complete the positioning task, which is advantageous in contrast to the classical exponential convergence. In particular, finite-time convergence is achieved while generating smooth robot velocities and considering the omnidirectional waking capability of the robot. In addition, we propose a hierarchical task-based control scheme, which can simultaneously handle the visual positioning and the obstacle avoidance tasks without affecting the desired time of convergence. The controller is able to activate or inactivate the obstacle avoidance task without generating discontinuous velocity references while the humanoid is walking. Stability of the closed loop for the two task-based control is demonstrated theoretically even during the transitions between the tasks. The proposed approach is generic in the sense that different visual control schemes are supported. We evaluate a homography-based visual servoing for position-based and image-based modalities, as well as for eye-in-hand and eye-to-hand configurations. The experimental evaluation is performed with the humanoid robot NAO.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('160','tp_bibtex')\">Cerrar<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_160\" style=\"display:none;\"><div class=\"tp_abstract_entry\">ABSTRACTIn this paper, we address the problem of humanoid locomotion guided from information of a monocular camera. The goal of the robot is to reach a desired location defined in terms of a target image, i.e., a positioning task. The proposed approach allows us to introduce a desired time to complete the positioning task, which is advantageous in contrast to the classical exponential convergence. In particular, finite-time convergence is achieved while generating smooth robot velocities and considering the omnidirectional waking capability of the robot. In addition, we propose a hierarchical task-based control scheme, which can simultaneously handle the visual positioning and the obstacle avoidance tasks without affecting the desired time of convergence. The controller is able to activate or inactivate the obstacle avoidance task without generating discontinuous velocity references while the humanoid is walking. Stability of the closed loop for the two task-based control is demonstrated theoretically even during the transitions between the tasks. The proposed approach is generic in the sense that different visual control schemes are supported. We evaluate a homography-based visual servoing for position-based and image-based modalities, as well as for eye-in-hand and eye-to-hand configurations. The experimental evaluation is performed with the humanoid robot NAO.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('160','tp_abstract')\">Cerrar<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_160\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/dx.doi.org\/10.1080\/00207179.2015.1129558\" title=\"http:\/\/dx.doi.org\/10.1080\/00207179.2015.1129558\" target=\"_blank\">http:\/\/dx.doi.org\/10.1080\/00207179.2015.1129558<\/a><\/li><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/10.1080\/00207179.2015.1129558\" title=\"DOI de seguimiento:10.1080\/00207179.2015.1129558\" target=\"_blank\">doi:10.1080\/00207179.2015.1129558<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('160','tp_links')\">Cerrar<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_conference\">Conferencias<\/h3><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Delfin, Josafat;  Becerra, H\u00e9ctor M;  Arechavaleta, Gustavo<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('3','tp_links')\" style=\"cursor:pointer;\">Humanoid Localization and Navigation using a Visual Memory<\/a> <span class=\"tp_pub_type tp_  conference\">Conferencia<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">IEEE-RAS 16th International Conference on Humanoid Robots, <\/span><span class=\"tp_pub_additional_publisher\">IEEE, <\/span><span class=\"tp_pub_additional_year\">2016<\/span>, <span class=\"tp_pub_additional_issn\">ISSN: 2164-0580<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_3\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('3','tp_abstract')\" title=\"Mostrar resumen\" style=\"cursor:pointer;\">Resumen<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_3\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('3','tp_links')\" title=\"Mostrar enlaces y recursos\" style=\"cursor:pointer;\">Enlaces<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_3\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('3','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_3\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{conf:Delfin2016,<br \/>\r\ntitle = {Humanoid Localization and Navigation using a Visual Memory},<br \/>\r\nauthor = {Delfin, Josafat and Becerra, H\\'{e}ctor M and Arechavaleta, Gustavo },<br \/>\r\ndoi = {10.1109\/HUMANOIDS.2016.7803354},<br \/>\r\nissn = {2164-0580},<br \/>\r\nyear  = {2016},<br \/>\r\ndate = {2016-11-15},<br \/>\r\nbooktitle = {IEEE-RAS 16th International Conference on Humanoid Robots},<br \/>\r\npages = {725-731},<br \/>\r\npublisher = {IEEE},<br \/>\r\nabstract = {A visual memory (VM) is a topological map in which a set of key images organized in form of a graph represents an environment. In this paper, a navigation strategy for humanoid robots addressing the problems of localization, visual path planning and path following based on a VM is proposed. Assuming that the VM is given, the main contributions of the paper are: 1) A novel pure vision-based localization method. 2) The introduction of the estimated rotation between key images in the path planning stage to benefit paths with enough visual information and with less effort of robot rotation. 3) The integration of the complete navigation strategy and its experimental evaluation with a Nao robot in an unstructured environment. The humanoid robot is modeled as a holonomic system and the strategy might be used in different scenarios like corridors, uncluttered or cluttered environments.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('3','tp_bibtex')\">Cerrar<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_3\" style=\"display:none;\"><div class=\"tp_abstract_entry\">A visual memory (VM) is a topological map in which a set of key images organized in form of a graph represents an environment. In this paper, a navigation strategy for humanoid robots addressing the problems of localization, visual path planning and path following based on a VM is proposed. Assuming that the VM is given, the main contributions of the paper are: 1) A novel pure vision-based localization method. 2) The introduction of the estimated rotation between key images in the path planning stage to benefit paths with enough visual information and with less effort of robot rotation. 3) The integration of the complete navigation strategy and its experimental evaluation with a Nao robot in an unstructured environment. The humanoid robot is modeled as a holonomic system and the strategy might be used in different scenarios like corridors, uncluttered or cluttered environments.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('3','tp_abstract')\">Cerrar<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_3\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/10.1109\/HUMANOIDS.2016.7803354\" title=\"DOI de seguimiento:10.1109\/HUMANOIDS.2016.7803354\" target=\"_blank\">doi:10.1109\/HUMANOIDS.2016.7803354<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('3','tp_links')\">Cerrar<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_2014\">2014<\/h3><h3 class=\"tp_h3\" id=\"tp_h3_conference\">Conferencias<\/h3><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Martinez-Gonzalez, Pablo;  Varas, David;  Castelan, Mario;  Camacho, Margarita;  Marques, Ferran;  Arechavaleta, Gustavo<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('19','tp_links')\" style=\"cursor:pointer;\">3D shape reconstruction from a humanoid generated video sequence<\/a> <span class=\"tp_pub_type tp_  conference\">Conferencia<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">2014 IEEE-RAS International Conference on Humanoid Robots, <\/span><span class=\"tp_pub_additional_year\">2014<\/span>, <span class=\"tp_pub_additional_issn\">ISSN: 2164-0572<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_19\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('19','tp_abstract')\" title=\"Mostrar resumen\" style=\"cursor:pointer;\">Resumen<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_19\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('19','tp_links')\" title=\"Mostrar enlaces y recursos\" style=\"cursor:pointer;\">Enlaces<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_19\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('19','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_19\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{7041439,<br \/>\r\ntitle = {3D shape reconstruction from a humanoid generated video sequence},<br \/>\r\nauthor = {Martinez-Gonzalez, Pablo and Varas, David and Castelan, Mario and Camacho, Margarita and Marques, Ferran and Arechavaleta, Gustavo },<br \/>\r\nurl = {http:\/\/ieeexplore.ieee.org\/document\/7041439\/?arnumber=7041439\\&amp;tag=1},<br \/>\r\ndoi = {10.1109\/HUMANOIDS.2014.7041439},<br \/>\r\nissn = {2164-0572},<br \/>\r\nyear  = {2014},<br \/>\r\ndate = {2014-11-01},<br \/>\r\nbooktitle = {2014 IEEE-RAS International Conference on Humanoid Robots},<br \/>\r\npages = {699-706},<br \/>\r\nabstract = {This paper presents a strategy for estimating the geometry of an interest object from a monocular video sequence acquired by a walking humanoid robot. The problem is solved using a space carving algorithm, which relies on both the accurate extraction of the occluding boundaries of the object as well as the precise estimation of the camera pose for each video frame. For data acquisition, a monocular visual-based control has been developed that drives the trajectory of the robot around an object placed on a small table. Due to the stepping of the humanoid, the recorded sequence is contaminated with artefacts that affect the correct extraction of contours along the video frames. To overcome this issue, a method that assigns a fitness score for each frame is proposed, delivering a subset of camera poses and video frames that produce consistent 3D shape estimations of the objects used for experimental evaluation.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('19','tp_bibtex')\">Cerrar<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_19\" style=\"display:none;\"><div class=\"tp_abstract_entry\">This paper presents a strategy for estimating the geometry of an interest object from a monocular video sequence acquired by a walking humanoid robot. The problem is solved using a space carving algorithm, which relies on both the accurate extraction of the occluding boundaries of the object as well as the precise estimation of the camera pose for each video frame. For data acquisition, a monocular visual-based control has been developed that drives the trajectory of the robot around an object placed on a small table. Due to the stepping of the humanoid, the recorded sequence is contaminated with artefacts that affect the correct extraction of contours along the video frames. To overcome this issue, a method that assigns a fitness score for each frame is proposed, delivering a subset of camera poses and video frames that produce consistent 3D shape estimations of the objects used for experimental evaluation.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('19','tp_abstract')\">Cerrar<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_19\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/ieeexplore.ieee.org\/document\/7041439\/?arnumber=7041439&amp;amp;tag=1\" title=\"http:\/\/ieeexplore.ieee.org\/document\/7041439\/?arnumber=7041439&amp;amp;tag=1\" target=\"_blank\">http:\/\/ieeexplore.ieee.org\/document\/7041439\/?arnumber=7041439&amp;amp;tag=1<\/a><\/li><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/10.1109\/HUMANOIDS.2014.7041439\" title=\"DOI de seguimiento:10.1109\/HUMANOIDS.2014.7041439\" target=\"_blank\">doi:10.1109\/HUMANOIDS.2014.7041439<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('19','tp_links')\">Cerrar<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_inproceedings\">Proceedings Articles<\/h3><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Estopier-Castillo, Vicente;  Arechavaleta, Gustavo;  Olgu\u00edn-D\u00edaz, Ernesto<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('78','tp_links')\" style=\"cursor:pointer;\">Generacion de Movimientos Humanoides con Dinamica Inversa Jerarquica<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">En: <\/span><span class=\"tp_pub_additional_booktitle\">Generacion de Movimientos Humanoides con Dinamica Inversa Jerarquica, <\/span><span class=\"tp_pub_additional_publisher\">Congreso Latinoamericano de Control Autom\u00e1tico CLCA 2014, <\/span><span class=\"tp_pub_additional_year\">2014<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_resource_link\"><a id=\"tp_links_sh_78\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('78','tp_links')\" title=\"Mostrar enlaces y recursos\" style=\"cursor:pointer;\">Enlaces<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_78\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('78','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_78\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{Castillo2014,<br \/>\r\ntitle = {Generacion de Movimientos Humanoides con Dinamica Inversa Jerarquica},<br \/>\r\nauthor = {Estopier-Castillo, Vicente and Arechavaleta, Gustavo and Olgu\\'{i}n-D\\'{i}az, Ernesto},<br \/>\r\nurl = {http:\/\/amca.mx\/memorias\/amca2014\/articulos\/0112.pdf},<br \/>\r\nyear  = {2014},<br \/>\r\ndate = {2014-00-00},<br \/>\r\nbooktitle = {Generacion de Movimientos Humanoides con Dinamica Inversa Jerarquica},<br \/>\r\npublisher = {Congreso Latinoamericano de Control Autom\\'{a}tico CLCA 2014},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('78','tp_bibtex')\">Cerrar<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_78\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-file-pdf\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/amca.mx\/memorias\/amca2014\/articulos\/0112.pdf\" title=\"http:\/\/amca.mx\/memorias\/amca2014\/articulos\/0112.pdf\" target=\"_blank\">http:\/\/amca.mx\/memorias\/amca2014\/articulos\/0112.pdf<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('78','tp_links')\">Cerrar<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_2012\">2012<\/h3><h3 class=\"tp_h3\" id=\"tp_h3_conference\">Conferencias<\/h3><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Delfin, Josafat;  Mar, Oscar;  Hayet, Jean Bernard;  Castelan, Mario;  Arechavaleta, Gustavo<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('26','tp_links')\" style=\"cursor:pointer;\">An active strategy for the simultaneous localization and reconstruction of a 3D object from a humanoid platform<\/a> <span class=\"tp_pub_type tp_  conference\">Conferencia<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">2012 12th IEEE-RAS International Conference on Humanoid Robots (Humanoids 2012), <\/span><span class=\"tp_pub_additional_year\">2012<\/span>, <span class=\"tp_pub_additional_issn\">ISSN: 2164-0572<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_26\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('26','tp_abstract')\" title=\"Mostrar resumen\" style=\"cursor:pointer;\">Resumen<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_26\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('26','tp_links')\" title=\"Mostrar enlaces y recursos\" style=\"cursor:pointer;\">Enlaces<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_26\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('26','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_26\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{6651548,<br \/>\r\ntitle = {An active strategy for the simultaneous localization and reconstruction of a 3D object from a humanoid platform},<br \/>\r\nauthor = {Delfin, Josafat and Mar, Oscar and Hayet, Jean Bernard and Castelan, Mario and Arechavaleta, Gustavo },<br \/>\r\nurl = {http:\/\/ieeexplore.ieee.org\/document\/6651548\/?arnumber=6651548},<br \/>\r\ndoi = {10.1109\/HUMANOIDS.2012.6651548},<br \/>\r\nissn = {2164-0572},<br \/>\r\nyear  = {2012},<br \/>\r\ndate = {2012-11-01},<br \/>\r\nbooktitle = {2012 12th IEEE-RAS International Conference on Humanoid Robots (Humanoids 2012)},<br \/>\r\npages = {384-389},<br \/>\r\nabstract = {This paper proposes a novel approach for the simultaneous localization and reconstruction of a 3D object. This reconstruction is done with a small humanoid platform by the only means of monocular vision.We combine (1) a stochastic control approach to decide how to move the robot such that the object can be both localized with respect to the robot and correctly reconstructed, and (2) a space carving approach for the 3D reconstruction of the object that merges the acquired monocular views into a coherent model. We present a set of very promising results obtained with a NAO platform.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('26','tp_bibtex')\">Cerrar<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_26\" style=\"display:none;\"><div class=\"tp_abstract_entry\">This paper proposes a novel approach for the simultaneous localization and reconstruction of a 3D object. This reconstruction is done with a small humanoid platform by the only means of monocular vision.We combine (1) a stochastic control approach to decide how to move the robot such that the object can be both localized with respect to the robot and correctly reconstructed, and (2) a space carving approach for the 3D reconstruction of the object that merges the acquired monocular views into a coherent model. We present a set of very promising results obtained with a NAO platform.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('26','tp_abstract')\">Cerrar<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_26\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/ieeexplore.ieee.org\/document\/6651548\/?arnumber=6651548\" title=\"http:\/\/ieeexplore.ieee.org\/document\/6651548\/?arnumber=6651548\" target=\"_blank\">http:\/\/ieeexplore.ieee.org\/document\/6651548\/?arnumber=6651548<\/a><\/li><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/10.1109\/HUMANOIDS.2012.6651548\" title=\"DOI de seguimiento:10.1109\/HUMANOIDS.2012.6651548\" target=\"_blank\">doi:10.1109\/HUMANOIDS.2012.6651548<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('26','tp_links')\">Cerrar<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_2010\">2010<\/h3><h3 class=\"tp_h3\" id=\"tp_h3_conference\">Conferencias<\/h3><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Ramirez-Bejarano, Christian;  Castelan, Mario;  Arechavaleta, Gustavo<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('32','tp_links')\" style=\"cursor:pointer;\">Multilinear decomposition of human walking paths<\/a> <span class=\"tp_pub_type tp_  conference\">Conferencia<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">2010 10th IEEE-RAS International Conference on Humanoid Robots, <\/span><span class=\"tp_pub_additional_year\">2010<\/span>, <span class=\"tp_pub_additional_issn\">ISSN: 2164-0572<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_32\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('32','tp_abstract')\" title=\"Mostrar resumen\" style=\"cursor:pointer;\">Resumen<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_32\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('32','tp_links')\" title=\"Mostrar enlaces y recursos\" style=\"cursor:pointer;\">Enlaces<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_32\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('32','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_32\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{5686313,<br \/>\r\ntitle = {Multilinear decomposition of human walking paths},<br \/>\r\nauthor = {Ramirez-Bejarano, Christian and Castelan, Mario and Arechavaleta, Gustavo },<br \/>\r\nurl = {http:\/\/ieeexplore.ieee.org\/document\/5686313\/?arnumber=5686313},<br \/>\r\ndoi = {10.1109\/ICHR.2010.5686313},<br \/>\r\nissn = {2164-0572},<br \/>\r\nyear  = {2010},<br \/>\r\ndate = {2010-12-01},<br \/>\r\nbooktitle = {2010 10th IEEE-RAS International Conference on Humanoid Robots},<br \/>\r\npages = {492-497},<br \/>\r\nabstract = {In a previous work, the authors have shown how the Principal Components Analysis (PCA) of a set of human walking paths provides sufficient information to derive a linear human-like path generator based on examples. The present work aims to provide an analysis of human walking paths from the perspective of multilinear algebra, using the n-mode Singular Value Decomposition (SVD). This decomposition subsumes the conventional PCA and extends its capabilities to study the interaction between the different vector spaces attached to the data. The training set consists of sixty routes performed three times by five different subjects for a total of 900 examples. The outcome of this decomposition is a novel multilinear model capable of revealing the global shape variations between the different vector spaces, particularly, exhibiting the planning strategies that each subject seemed to use to approach to the different targets. In addition, we sketch how the multilinear model may be used to synthesize new routes for the hypothetical case of a subject trying to reach a moving target.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('32','tp_bibtex')\">Cerrar<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_32\" style=\"display:none;\"><div class=\"tp_abstract_entry\">In a previous work, the authors have shown how the Principal Components Analysis (PCA) of a set of human walking paths provides sufficient information to derive a linear human-like path generator based on examples. The present work aims to provide an analysis of human walking paths from the perspective of multilinear algebra, using the n-mode Singular Value Decomposition (SVD). This decomposition subsumes the conventional PCA and extends its capabilities to study the interaction between the different vector spaces attached to the data. The training set consists of sixty routes performed three times by five different subjects for a total of 900 examples. The outcome of this decomposition is a novel multilinear model capable of revealing the global shape variations between the different vector spaces, particularly, exhibiting the planning strategies that each subject seemed to use to approach to the different targets. In addition, we sketch how the multilinear model may be used to synthesize new routes for the hypothetical case of a subject trying to reach a moving target.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('32','tp_abstract')\">Cerrar<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_32\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/ieeexplore.ieee.org\/document\/5686313\/?arnumber=5686313\" title=\"http:\/\/ieeexplore.ieee.org\/document\/5686313\/?arnumber=5686313\" target=\"_blank\">http:\/\/ieeexplore.ieee.org\/document\/5686313\/?arnumber=5686313<\/a><\/li><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/10.1109\/ICHR.2010.5686313\" title=\"DOI de seguimiento:10.1109\/ICHR.2010.5686313\" target=\"_blank\">doi:10.1109\/ICHR.2010.5686313<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('32','tp_links')\">Cerrar<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_2009\">2009<\/h3><h3 class=\"tp_h3\" id=\"tp_h3_conference\">Conferencias<\/h3><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Castelan, Mario;  Arechavaleta, Gustavo<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('34','tp_links')\" style=\"cursor:pointer;\">Approximating the reachable space of human walking paths: a low dimensional linear approach<\/a> <span class=\"tp_pub_type tp_  conference\">Conferencia<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">2009 9th IEEE-RAS International Conference on Humanoid Robots, <\/span><span class=\"tp_pub_additional_year\">2009<\/span>, <span class=\"tp_pub_additional_issn\">ISSN: 2164-0572<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_34\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('34','tp_abstract')\" title=\"Mostrar resumen\" style=\"cursor:pointer;\">Resumen<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_34\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('34','tp_links')\" title=\"Mostrar enlaces y recursos\" style=\"cursor:pointer;\">Enlaces<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_34\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('34','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_34\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{5379595,<br \/>\r\ntitle = {Approximating the reachable space of human walking paths: a low dimensional linear approach},<br \/>\r\nauthor = {Castelan, Mario and Arechavaleta, Gustavo },<br \/>\r\nurl = {http:\/\/ieeexplore.ieee.org\/document\/5379595\/?arnumber=5379595},<br \/>\r\ndoi = {10.1109\/ICHR.2009.5379595},<br \/>\r\nissn = {2164-0572},<br \/>\r\nyear  = {2009},<br \/>\r\ndate = {2009-12-01},<br \/>\r\nbooktitle = {2009 9th IEEE-RAS International Conference on Humanoid Robots},<br \/>\r\npages = {81-86},<br \/>\r\nabstract = {In this work, we aim to exhibit the geometrical shape primitives of human walking trajectories using a statistical model constructed through Principal Component Analysis. This analysis provides sufficient information to derive a linear human-like path generator based on examples. The examples are provided by a motion capture database of human walking trajectories. The proposed model captures the shape of trajectories in terms of path length and deformation. We have successfully applied our model to compute a good approximation of the reachable space of human walking. This can be done with a negligible computational cost since it is based on a linear combination of basis human paths.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('34','tp_bibtex')\">Cerrar<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_34\" style=\"display:none;\"><div class=\"tp_abstract_entry\">In this work, we aim to exhibit the geometrical shape primitives of human walking trajectories using a statistical model constructed through Principal Component Analysis. This analysis provides sufficient information to derive a linear human-like path generator based on examples. The examples are provided by a motion capture database of human walking trajectories. The proposed model captures the shape of trajectories in terms of path length and deformation. We have successfully applied our model to compute a good approximation of the reachable space of human walking. This can be done with a negligible computational cost since it is based on a linear combination of basis human paths.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('34','tp_abstract')\">Cerrar<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_34\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/ieeexplore.ieee.org\/document\/5379595\/?arnumber=5379595\" title=\"http:\/\/ieeexplore.ieee.org\/document\/5379595\/?arnumber=5379595\" target=\"_blank\">http:\/\/ieeexplore.ieee.org\/document\/5379595\/?arnumber=5379595<\/a><\/li><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/10.1109\/ICHR.2009.5379595\" title=\"DOI de seguimiento:10.1109\/ICHR.2009.5379595\" target=\"_blank\">doi:10.1109\/ICHR.2009.5379595<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('34','tp_links')\">Cerrar<\/a><\/p><\/div><\/div><\/div><\/div><\/div><\/strong><\/p>\n<p>[\/et_pb_text][\/et_pb_column][\/et_pb_row][\/et_pb_section][et_pb_section bb_built=\u00bb1&#8243; fullwidth=\u00bboff\u00bb specialty=\u00bboff\u00bb background_color=\u00bbrgba(0,0,0,0.32)\u00bb inner_shadow=\u00bbon\u00bb custom_css_main_element=\u00bbbox-shadow: inset 0px 3px 2px rgba(150, 150, 150, 0.85);\u00bb _builder_version=\u00bb3.0.72&#8243; locked=\u00bbon\u00bb global_module=\u00bb321&#8243;][et_pb_row global_parent=\u00bb321&#8243; make_fullwidth=\u00bboff\u00bb use_custom_width=\u00bboff\u00bb width_unit=\u00bbon\u00bb use_custom_gutter=\u00bboff\u00bb allow_player_pause=\u00bboff\u00bb parallax=\u00bboff\u00bb parallax_method=\u00bbon\u00bb make_equal=\u00bboff\u00bb parallax_1=\u00bboff\u00bb parallax_method_1=\u00bboff\u00bb custom_margin=\u00bb-40px|||\u00bb background_position=\u00bbtop_left\u00bb background_repeat=\u00bbrepeat\u00bb background_size=\u00bbinitial\u00bb parent_locked=\u00bbon\u00bb][et_pb_column type=\u00bb4_4&#8243;][et_pb_image admin_label=\u00bbLogoCINVESTAV del Pie de p\u00e1gina\u00bb global_parent=\u00bb321&#8243; src=\u00bbhttps:\/\/ryma.cinvestav.mx\/wp-content\/uploads\/2014\/08\/roboticaCinvestavOK_transparencia_white.png\u00bb alt=\u00bbRob\u00f3tica y Manufactura Avanzada, Cinvestav\u00bb show_in_lightbox=\u00bboff\u00bb url_new_window=\u00bboff\u00bb use_overlay=\u00bboff\u00bb animation=\u00bboff\u00bb sticky=\u00bbon\u00bb align=\u00bbcenter\u00bb max_width=\u00bb95px\u00bb max_width_last_edited=\u00bbon|desktop\u00bb force_fullwidth=\u00bboff\u00bb always_center_on_mobile=\u00bbon\u00bb border_style=\u00bbsolid\u00bb custom_margin=\u00bb||15px|\u00bb _builder_version=\u00bb3.0.72&#8243; parent_locked=\u00bbon\u00bb \/][et_pb_text global_parent=\u00bb321&#8243; _builder_version=\u00bb3.0.72&#8243; background_layout=\u00bbdark\u00bb text_orientation=\u00bbcenter\u00bb border_style=\u00bbsolid\u00bb custom_margin=\u00bb||-50px|\u00bb parent_locked=\u00bbon\u00bb]<\/p>\n<hr \/>\n<p style=\"text-align: center;\">Av. Industrial\u00a0Metalurgia\u00a0#1062,\u00a0Parque Ind. Ramos Arizpe,\u00a0Ramos Arizpe, Coah.\u00a0C.P. 25900, M\u00e9xico. \u00a0Tel. +52 (844) 438-9600<\/p>\n<p>[\/et_pb_text][\/et_pb_column][\/et_pb_row][\/et_pb_section]<\/p>\n","protected":false},"excerpt":{"rendered":"<p><div class='et-box et-shadow'>\n\t\t\t\t\t<div class='et-box-content'>PUBLICACIONES<\/div><\/div> Para ver las publicaciones de todo Rob\u00f3tica y Manufactura Avanzada, ver:\u00a0 Publicaciones RYMA <div class=\"teachpress_pub_list\"><form name=\"tppublistform\" method=\"get\"><a name=\"tppubs\" id=\"tppubs\"><\/a><div class=\"teachpress_filter\"><select class=\"default\" name=\"yr\" id=\"yr\" tabindex=\"2\" onchange=\"teachpress_jumpMenu('parent',this, 'https:\/\/ryma.cinvestav.mx\/garechav\/publicaciones\/?')\">\r\n                   <option value=\"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=#tppubs\">Todos los a\u00f1os<\/option>\r\n                   <option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2020#tppubs\" >2020<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2019#tppubs\" >2019<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2018#tppubs\" >2018<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2017#tppubs\" >2017<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2016#tppubs\" >2016<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2015#tppubs\" >2015<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2014#tppubs\" >2014<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2013#tppubs\" >2013<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2012#tppubs\" >2012<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2011#tppubs\" >2011<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2010#tppubs\" >2010<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2009#tppubs\" >2009<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2008#tppubs\" >2008<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2007#tppubs\" >2007<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2006#tppubs\" >2006<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2005#tppubs\" >2005<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2004#tppubs\" >2004<\/option><option value = \"tgid=&amp;type=&amp;auth=&amp;usr=&amp;yr=2003#tppubs\" >2003<\/option>\r\n                <\/select><select class=\"default\" name=\"type\" id=\"type\" tabindex=\"3\" onchange=\"teachpress_jumpMenu('parent',this, 'https:\/\/ryma.cinvestav.mx\/garechav\/publicaciones\/?')\">\r\n                   <option value=\"tgid=&amp;yr=&amp;auth=&amp;usr=&amp;type=#tppubs\">Todas las tipolog\u00edas<\/option>\r\n                   <option value = \"tgid=&amp;yr=&amp;auth=&amp;usr=&amp;type=article#tppubs\" >Art\u00edculos de revista<\/option><option value = \"tgid=&amp;yr=&amp;auth=&amp;usr=&amp;type=conference#tppubs\" >Conferencias<\/option><option value = \"tgid=&amp;yr=&amp;auth=&amp;usr=&amp;type=inbook#tppubs\" >Cap\u00edtulos de libros<\/option><option value = \"tgid=&amp;yr=&amp;auth=&amp;usr=&amp;type=inproceedings#tppubs\" >Proceedings Articles<\/option><option value = \"tgid=&amp;yr=&amp;auth=&amp;usr=&amp;type=proceedings#tppubs\" >Actas de congresos<\/option>\r\n                <\/select><select class=\"default\" name=\"usr\" id=\"usr\" tabindex=\"6\" onchange=\"teachpress_jumpMenu('parent',this, 'https:\/\/ryma.cinvestav.mx\/garechav\/publicaciones\/?')\">\r\n                   <option value=\"tgid=&amp;yr=&amp;type=&amp;auth=&amp;usr=#tppubs\">Todos los usuarios<\/option>\r\n                   <option value = \"tgid=&amp;yr=&amp;type=&amp;auth=&amp;usr=12#tppubs\" >mcastelan<\/option>\r\n                <\/select><\/div><\/form><div class=\"teachpress_publication_list\"><h3 class=\"tp_h3\" id=\"tp_h3_2017\">2017<\/h3><h3 class=\"tp_h3\" id=\"tp_h3_article\">Art\u00edculos de revista<\/h3><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Arechavaleta, Gustavo;  Morales-Diaz, America B.;  Perez-Villeda, Hector Manuel;  Castelan, Mario<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('2','tp_links')\" style=\"cursor:pointer;\">Hierarchical Task-Based Control of Multirobot Systems With Terminal Attractors<\/a> <span class=\"tp_pub_type tp_  article\">Art\u00edculo de revista<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">En: <\/span><span class=\"tp_pub_additional_journal\">IEEE Transactions on Control Systems Technology, <\/span><span class=\"tp_pub_additional_volume\">vol. 25, <\/span><span class=\"tp_pub_additional_number\">no 1, <\/span><span class=\"tp_pub_additional_pages\">pp. 334 - 341, <\/span><span class=\"tp_pub_additional_year\">2017<\/span>, <span class=\"tp_pub_additional_issn\">ISSN: 1063-6536<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_2\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('2','tp_abstract')\" title=\"Mostrar resumen\" style=\"cursor:pointer;\">Resumen<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_2\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('2','tp_links')\" title=\"Mostrar enlaces y recursos\" style=\"cursor:pointer;\">Enlaces<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_2\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('2','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_2\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{7454708,<br \/>\r\ntitle = {Hierarchical Task-Based Control of Multirobot Systems With Terminal Attractors},<br \/>\r\nauthor = {Arechavaleta, Gustavo and Morales-Diaz, America B. and Perez-Villeda, Hector Manuel and Castelan, Mario },<br \/>\r\nurl = {http:\/\/ieeexplore.ieee.org\/abstract\/document\/7454708\/},<br \/>\r\ndoi = {10.1109\/TCST.2016.2549279},<br \/>\r\nissn = {1063-6536},<br \/>\r\nyear  = {2017},<br \/>\r\ndate = {2017-01-01},<br \/>\r\njournal = {IEEE Transactions on Control Systems Technology},<br \/>\r\nvolume = {25},<br \/>\r\nnumber = {1},<br \/>\r\npages = {334 - 341},<br \/>\r\nabstract = {This brief proposes a hierarchical control scheme based on the definition of a set of multirobot task functions. To deal with the inherent conflicts between tasks, a strict hierarchy is imposed on them. We present a novel scheme that copes with two main difficulties shared in standard task-based controllers: 1) to impose a desired time convergence of tasks and 2) to avoid discontinuous task transitions occurred when a task is inserted or removed in the hierarchical structure. As a result, continuous input references are generated for the low-level control of the group. The validation is achieved in simulation and by performing an experiment with wheeled mobile robots.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('2','tp_bibtex')\">Cerrar<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_2\" style=\"display:none;\"><div class=\"tp_abstract_entry\">This brief proposes a hierarchical control scheme based on the definition of a set of multirobot task functions. To deal with the inherent conflicts between tasks, a strict hierarchy is imposed on them. We present a novel scheme that copes with two main difficulties shared in standard task-based controllers: 1) to impose a desired time convergence of tasks and 2) to avoid discontinuous task transitions occurred when a task is inserted or removed in the hierarchical structure. As a result, continuous input references are generated for the low-level control of the group. The validation is achieved in simulation and by performing an experiment with wheeled mobile robots.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('2','tp_abstract')\">Cerrar<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_2\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/ieeexplore.ieee.org\/abstract\/document\/7454708\/\" title=\"http:\/\/ieeexplore.ieee.org\/abstract\/document\/7454708\/\" target=\"_blank\">http:\/\/ieeexplore.ieee.org\/abstract\/document\/7454708\/<\/a><\/li><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/10.1109\/TCST.2016.2549279\" title=\"DOI de seguimiento:10.1109\/TCST.2016.2549279\" target=\"_blank\">doi:10.1109\/TCST.2016.2549279<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('2','tp_links')\">Cerrar<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_2016\">2016<\/h3><h3 class=\"tp_h3\" id=\"tp_h3_article\">Art\u00edculos de revista<\/h3><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Mart\u00ednez-Gonz\u00e1lez, Pablo Arturo;  Castelan, Mario;  Arechavaleta, Gustavo<\/p><p class=\"tp_pub_title\">Vision Based Persistent Localization of a Humanoid Robot for Locomotion Tasks <span class=\"tp_pub_type tp_  article\">Art\u00edculo de revista<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">En: <\/span><span class=\"tp_pub_additional_year\">2016<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_159\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('159','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_159\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{Mart\\'{i}nez-Gonz\\'{a}lez2016b,<br \/>\r\ntitle = {Vision Based Persistent Localization of a Humanoid Robot for Locomotion Tasks},<br \/>\r\nauthor = {Mart\\'{i}nez-Gonz\\'{a}lez, Pablo Arturo and Castelan, Mario and Arechavaleta, Gustavo},<br \/>\r\nyear  = {2016},<br \/>\r\ndate = {2016-06-06},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('159','tp_bibtex')\">Cerrar<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Martinez-Gonzalez, Pablo;  Castelan, Mario;  Arechavaleta, Gustavo<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('1','tp_links')\" style=\"cursor:pointer;\">Vision based persistent localization of a humanoid robot for locomotion Tasks<\/a> <span class=\"tp_pub_type tp_  article\">Art\u00edculo de revista<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">En: <\/span><span class=\"tp_pub_additional_journal\">International Journal of Applied Mathematics and Computer Science, <\/span><span class=\"tp_pub_additional_volume\">vol. 26, <\/span><span class=\"tp_pub_additional_number\">no 3, <\/span><span class=\"tp_pub_additional_year\">2016<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_resource_link\"><a id=\"tp_links_sh_1\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('1','tp_links')\" title=\"Mostrar enlaces y recursos\" style=\"cursor:pointer;\">Enlaces<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_1\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('1','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_1\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{Mart\\`{i}nez-Gonz\\'{a}lez2016,<br \/>\r\ntitle = {Vision based persistent localization of a humanoid robot for locomotion Tasks},<br \/>\r\nauthor = {Martinez-Gonzalez, Pablo and Castelan, Mario and Arechavaleta, Gustavo },<br \/>\r\nurl = {https:\/\/drive.google.com\/file\/d\/0B-7dVUdTjeJUNGdXd0N6UWRvdk0\/view},<br \/>\r\nyear  = {2016},<br \/>\r\ndate = {2016-03-26},<br \/>\r\njournal = {International Journal of Applied Mathematics and Computer Science},<br \/>\r\nvolume = {26},<br \/>\r\nnumber = {3},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('1','tp_bibtex')\">Cerrar<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_1\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/drive.google.com\/file\/d\/0B-7dVUdTjeJUNGdXd0N6UWRvdk0\/view\" title=\"https:\/\/drive.google.com\/file\/d\/0B-7dVUdTjeJUNGdXd0N6UWRvdk0\/view\" target=\"_blank\">https:\/\/drive.google.com\/file\/d\/0B-7dVUdTjeJUNGdXd0N6UWRvdk0\/view<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('1','tp_links')\">Cerrar<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Delfin, Josafat;  Becerra, Hector M;  Arechavaleta, Gustavo<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('5','tp_links')\" style=\"cursor:pointer;\">Visual Servo Walking Control for Humanoids with Finite-time Convergence and Smooth Robot Velocities<\/a> <span class=\"tp_pub_type tp_  article\">Art\u00edculo de revista<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">En: <\/span><span class=\"tp_pub_additional_journal\">International Journal of Control, <\/span><span class=\"tp_pub_additional_volume\">vol. 89, <\/span><span class=\"tp_pub_additional_number\">no 7, <\/span><span class=\"tp_pub_additional_pages\">pp. 1342-1358, <\/span><span class=\"tp_pub_additional_year\">2016<\/span>, <span class=\"tp_pub_additional_issn\">ISSN: 1366-5820<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_5\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('5','tp_abstract')\" title=\"Mostrar resumen\" style=\"cursor:pointer;\">Resumen<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_5\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('5','tp_links')\" title=\"Mostrar enlaces y recursos\" style=\"cursor:pointer;\">Enlaces<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_5\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('5','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_5\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{Delfin2016,<br \/>\r\ntitle = {Visual Servo Walking Control for Humanoids with Finite-time Convergence and Smooth Robot Velocities},<br \/>\r\nauthor = {Delfin, Josafat and Becerra, Hector M and Arechavaleta, Gustavo },<br \/>\r\nurl = {http:\/\/www.tandfonline.com\/doi\/abs\/10.1080\/00207179.2015.1129558},<br \/>\r\ndoi = {http:\/\/dx.doi.org\/10.1080\/00207179.2015.1129558},<br \/>\r\nissn = {1366-5820},<br \/>\r\nyear  = {2016},<br \/>\r\ndate = {2016-01-10},<br \/>\r\njournal = {International Journal of Control},<br \/>\r\nvolume = {89},<br \/>\r\nnumber = {7},<br \/>\r\npages = {1342-1358},<br \/>\r\nabstract = {In this paper, we address the problem of humanoid locomotion guided from information of a monocular camera. The goal of the robot is to reach a desired location defined in terms of a target image, i.e., a positioning task. The proposed approach allows us to introduce a desired time to complete the positioning task, which is advantageous in contrast to the classical exponential convergence. In particular, finite-time convergence is achieved while generating smooth robot velocities and considering the omnidirectional waking capability of the robot. In addition, we propose a hierarchical task-based control scheme, which can simultaneously handle the visual positioning and the obstacle avoidance tasks without affecting the desired time of convergence. The controller is able to activate or inactivate the obstacle avoidance task without generating discontinuous velocity references while the humanoid is walking. Stability of the closed loop for the two task-based control is demonstrated theoretically even during the transitions between the tasks. The proposed approach is generic in the sense that different visual control schemes are supported. We evaluate a homography-based visual servoing for position-based and image-based modalities, as well as for eye-in-hand and eye-to-hand configurations. The experimental evaluation is performed with the humanoid robot NAO.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('5','tp_bibtex')\">Cerrar<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_5\" style=\"display:none;\"><div class=\"tp_abstract_entry\">In this paper, we address the problem of humanoid locomotion guided from information of a monocular camera. The goal of the robot is to reach a desired location defined in terms of a target image, i.e., a positioning task. The proposed approach allows us to introduce a desired time to complete the positioning task, which is advantageous in contrast to the classical exponential convergence. In particular, finite-time convergence is achieved while generating smooth robot velocities and considering the omnidirectional waking capability of the robot. In addition, we propose a hierarchical task-based control scheme, which can simultaneously handle the visual positioning and the obstacle avoidance tasks without affecting the desired time of convergence. The controller is able to activate or inactivate the obstacle avoidance task without generating discontinuous velocity references while the humanoid is walking. Stability of the closed loop for the two task-based control is demonstrated theoretically even during the transitions between the tasks. The proposed approach is generic in the sense that different visual control schemes are supported. We evaluate a homography-based visual servoing for position-based and image-based modalities, as well as for eye-in-hand and eye-to-hand configurations. The experimental evaluation is performed with the humanoid robot NAO.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('5','tp_abstract')\">Cerrar<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_5\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/www.tandfonline.com\/doi\/abs\/10.1080\/00207179.2015.1129558\" title=\"http:\/\/www.tandfonline.com\/doi\/abs\/10.1080\/00207179.2015.1129558\" target=\"_blank\">http:\/\/www.tandfonline.com\/doi\/abs\/10.1080\/00207179.2015.1129558<\/a><\/li><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/http:\/\/dx.doi.org\/10.1080\/00207179.2015.1129558\" title=\"DOI de seguimiento:http:\/\/dx.doi.org\/10.1080\/00207179.2015.1129558\" target=\"_blank\">doi:http:\/\/dx.doi.org\/10.1080\/00207179.2015.1129558<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('5','tp_links')\">Cerrar<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Delfin, Josafat;  Becerra, Hector M;  Arechavaleta, Gustavo<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('160','tp_links')\" style=\"cursor:pointer;\">Visual servo walking control for humanoids with finite-time convergence and smooth robot velocities<\/a> <span class=\"tp_pub_type tp_  article\">Art\u00edculo de revista<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">En: <\/span><span class=\"tp_pub_additional_journal\">International Journal of Control, <\/span><span class=\"tp_pub_additional_volume\">vol. 89, <\/span><span class=\"tp_pub_additional_number\">no 7, <\/span><span class=\"tp_pub_additional_pages\">pp. 1342-1358, <\/span><span class=\"tp_pub_additional_year\">2016<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_160\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('160','tp_abstract')\" title=\"Mostrar resumen\" style=\"cursor:pointer;\">Resumen<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_160\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('160','tp_links')\" title=\"Mostrar enlaces y recursos\" style=\"cursor:pointer;\">Enlaces<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_160\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('160','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_160\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{doi:10.1080\/00207179.2015.1129558,<br \/>\r\ntitle = {Visual servo walking control for humanoids with finite-time convergence and smooth robot velocities},<br \/>\r\nauthor = {Delfin, Josafat and Becerra, Hector M and Arechavaleta, Gustavo},<br \/>\r\nurl = {http:\/\/dx.doi.org\/10.1080\/00207179.2015.1129558},<br \/>\r\ndoi = {10.1080\/00207179.2015.1129558},<br \/>\r\nyear  = {2016},<br \/>\r\ndate = {2016-01-01},<br \/>\r\njournal = {International Journal of Control},<br \/>\r\nvolume = {89},<br \/>\r\nnumber = {7},<br \/>\r\npages = {1342-1358},<br \/>\r\nabstract = {ABSTRACTIn this paper, we address the problem of humanoid locomotion guided from information of a monocular camera. The goal of the robot is to reach a desired location defined in terms of a target image, i.e., a positioning task. The proposed approach allows us to introduce a desired time to complete the positioning task, which is advantageous in contrast to the classical exponential convergence. In particular, finite-time convergence is achieved while generating smooth robot velocities and considering the omnidirectional waking capability of the robot. In addition, we propose a hierarchical task-based control scheme, which can simultaneously handle the visual positioning and the obstacle avoidance tasks without affecting the desired time of convergence. The controller is able to activate or inactivate the obstacle avoidance task without generating discontinuous velocity references while the humanoid is walking. Stability of the closed loop for the two task-based control is demonstrated theoretically even during the transitions between the tasks. The proposed approach is generic in the sense that different visual control schemes are supported. We evaluate a homography-based visual servoing for position-based and image-based modalities, as well as for eye-in-hand and eye-to-hand configurations. The experimental evaluation is performed with the humanoid robot NAO.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('160','tp_bibtex')\">Cerrar<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_160\" style=\"display:none;\"><div class=\"tp_abstract_entry\">ABSTRACTIn this paper, we address the problem of humanoid locomotion guided from information of a monocular camera. The goal of the robot is to reach a desired location defined in terms of a target image, i.e., a positioning task. The proposed approach allows us to introduce a desired time to complete the positioning task, which is advantageous in contrast to the classical exponential convergence. In particular, finite-time convergence is achieved while generating smooth robot velocities and considering the omnidirectional waking capability of the robot. In addition, we propose a hierarchical task-based control scheme, which can simultaneously handle the visual positioning and the obstacle avoidance tasks without affecting the desired time of convergence. The controller is able to activate or inactivate the obstacle avoidance task without generating discontinuous velocity references while the humanoid is walking. Stability of the closed loop for the two task-based control is demonstrated theoretically even during the transitions between the tasks. The proposed approach is generic in the sense that different visual control schemes are supported. We evaluate a homography-based visual servoing for position-based and image-based modalities, as well as for eye-in-hand and eye-to-hand configurations. The experimental evaluation is performed with the humanoid robot NAO.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('160','tp_abstract')\">Cerrar<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_160\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/dx.doi.org\/10.1080\/00207179.2015.1129558\" title=\"http:\/\/dx.doi.org\/10.1080\/00207179.2015.1129558\" target=\"_blank\">http:\/\/dx.doi.org\/10.1080\/00207179.2015.1129558<\/a><\/li><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/10.1080\/00207179.2015.1129558\" title=\"DOI de seguimiento:10.1080\/00207179.2015.1129558\" target=\"_blank\">doi:10.1080\/00207179.2015.1129558<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('160','tp_links')\">Cerrar<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_conference\">Conferencias<\/h3><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Delfin, Josafat;  Becerra, H\u00e9ctor M;  Arechavaleta, Gustavo<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('3','tp_links')\" style=\"cursor:pointer;\">Humanoid Localization and Navigation using a Visual Memory<\/a> <span class=\"tp_pub_type tp_  conference\">Conferencia<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">IEEE-RAS 16th International Conference on Humanoid Robots, <\/span><span class=\"tp_pub_additional_publisher\">IEEE, <\/span><span class=\"tp_pub_additional_year\">2016<\/span>, <span class=\"tp_pub_additional_issn\">ISSN: 2164-0580<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_3\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('3','tp_abstract')\" title=\"Mostrar resumen\" style=\"cursor:pointer;\">Resumen<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_3\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('3','tp_links')\" title=\"Mostrar enlaces y recursos\" style=\"cursor:pointer;\">Enlaces<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_3\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('3','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_3\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{conf:Delfin2016,<br \/>\r\ntitle = {Humanoid Localization and Navigation using a Visual Memory},<br \/>\r\nauthor = {Delfin, Josafat and Becerra, H\\'{e}ctor M and Arechavaleta, Gustavo },<br \/>\r\ndoi = {10.1109\/HUMANOIDS.2016.7803354},<br \/>\r\nissn = {2164-0580},<br \/>\r\nyear  = {2016},<br \/>\r\ndate = {2016-11-15},<br \/>\r\nbooktitle = {IEEE-RAS 16th International Conference on Humanoid Robots},<br \/>\r\npages = {725-731},<br \/>\r\npublisher = {IEEE},<br \/>\r\nabstract = {A visual memory (VM) is a topological map in which a set of key images organized in form of a graph represents an environment. In this paper, a navigation strategy for humanoid robots addressing the problems of localization, visual path planning and path following based on a VM is proposed. Assuming that the VM is given, the main contributions of the paper are: 1) A novel pure vision-based localization method. 2) The introduction of the estimated rotation between key images in the path planning stage to benefit paths with enough visual information and with less effort of robot rotation. 3) The integration of the complete navigation strategy and its experimental evaluation with a Nao robot in an unstructured environment. The humanoid robot is modeled as a holonomic system and the strategy might be used in different scenarios like corridors, uncluttered or cluttered environments.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('3','tp_bibtex')\">Cerrar<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_3\" style=\"display:none;\"><div class=\"tp_abstract_entry\">A visual memory (VM) is a topological map in which a set of key images organized in form of a graph represents an environment. In this paper, a navigation strategy for humanoid robots addressing the problems of localization, visual path planning and path following based on a VM is proposed. Assuming that the VM is given, the main contributions of the paper are: 1) A novel pure vision-based localization method. 2) The introduction of the estimated rotation between key images in the path planning stage to benefit paths with enough visual information and with less effort of robot rotation. 3) The integration of the complete navigation strategy and its experimental evaluation with a Nao robot in an unstructured environment. The humanoid robot is modeled as a holonomic system and the strategy might be used in different scenarios like corridors, uncluttered or cluttered environments.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('3','tp_abstract')\">Cerrar<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_3\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/10.1109\/HUMANOIDS.2016.7803354\" title=\"DOI de seguimiento:10.1109\/HUMANOIDS.2016.7803354\" target=\"_blank\">doi:10.1109\/HUMANOIDS.2016.7803354<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('3','tp_links')\">Cerrar<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_2014\">2014<\/h3><h3 class=\"tp_h3\" id=\"tp_h3_conference\">Conferencias<\/h3><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Martinez-Gonzalez, Pablo;  Varas, David;  Castelan, Mario;  Camacho, Margarita;  Marques, Ferran;  Arechavaleta, Gustavo<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('19','tp_links')\" style=\"cursor:pointer;\">3D shape reconstruction from a humanoid generated video sequence<\/a> <span class=\"tp_pub_type tp_  conference\">Conferencia<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">2014 IEEE-RAS International Conference on Humanoid Robots, <\/span><span class=\"tp_pub_additional_year\">2014<\/span>, <span class=\"tp_pub_additional_issn\">ISSN: 2164-0572<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_19\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('19','tp_abstract')\" title=\"Mostrar resumen\" style=\"cursor:pointer;\">Resumen<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_19\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('19','tp_links')\" title=\"Mostrar enlaces y recursos\" style=\"cursor:pointer;\">Enlaces<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_19\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('19','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_19\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{7041439,<br \/>\r\ntitle = {3D shape reconstruction from a humanoid generated video sequence},<br \/>\r\nauthor = {Martinez-Gonzalez, Pablo and Varas, David and Castelan, Mario and Camacho, Margarita and Marques, Ferran and Arechavaleta, Gustavo },<br \/>\r\nurl = {http:\/\/ieeexplore.ieee.org\/document\/7041439\/?arnumber=7041439\\&amp;tag=1},<br \/>\r\ndoi = {10.1109\/HUMANOIDS.2014.7041439},<br \/>\r\nissn = {2164-0572},<br \/>\r\nyear  = {2014},<br \/>\r\ndate = {2014-11-01},<br \/>\r\nbooktitle = {2014 IEEE-RAS International Conference on Humanoid Robots},<br \/>\r\npages = {699-706},<br \/>\r\nabstract = {This paper presents a strategy for estimating the geometry of an interest object from a monocular video sequence acquired by a walking humanoid robot. The problem is solved using a space carving algorithm, which relies on both the accurate extraction of the occluding boundaries of the object as well as the precise estimation of the camera pose for each video frame. For data acquisition, a monocular visual-based control has been developed that drives the trajectory of the robot around an object placed on a small table. Due to the stepping of the humanoid, the recorded sequence is contaminated with artefacts that affect the correct extraction of contours along the video frames. To overcome this issue, a method that assigns a fitness score for each frame is proposed, delivering a subset of camera poses and video frames that produce consistent 3D shape estimations of the objects used for experimental evaluation.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('19','tp_bibtex')\">Cerrar<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_19\" style=\"display:none;\"><div class=\"tp_abstract_entry\">This paper presents a strategy for estimating the geometry of an interest object from a monocular video sequence acquired by a walking humanoid robot. The problem is solved using a space carving algorithm, which relies on both the accurate extraction of the occluding boundaries of the object as well as the precise estimation of the camera pose for each video frame. For data acquisition, a monocular visual-based control has been developed that drives the trajectory of the robot around an object placed on a small table. Due to the stepping of the humanoid, the recorded sequence is contaminated with artefacts that affect the correct extraction of contours along the video frames. To overcome this issue, a method that assigns a fitness score for each frame is proposed, delivering a subset of camera poses and video frames that produce consistent 3D shape estimations of the objects used for experimental evaluation.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('19','tp_abstract')\">Cerrar<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_19\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/ieeexplore.ieee.org\/document\/7041439\/?arnumber=7041439&amp;amp;tag=1\" title=\"http:\/\/ieeexplore.ieee.org\/document\/7041439\/?arnumber=7041439&amp;amp;tag=1\" target=\"_blank\">http:\/\/ieeexplore.ieee.org\/document\/7041439\/?arnumber=7041439&amp;amp;tag=1<\/a><\/li><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/10.1109\/HUMANOIDS.2014.7041439\" title=\"DOI de seguimiento:10.1109\/HUMANOIDS.2014.7041439\" target=\"_blank\">doi:10.1109\/HUMANOIDS.2014.7041439<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('19','tp_links')\">Cerrar<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_inproceedings\">Proceedings Articles<\/h3><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Estopier-Castillo, Vicente;  Arechavaleta, Gustavo;  Olgu\u00edn-D\u00edaz, Ernesto<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('78','tp_links')\" style=\"cursor:pointer;\">Generacion de Movimientos Humanoides con Dinamica Inversa Jerarquica<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">En: <\/span><span class=\"tp_pub_additional_booktitle\">Generacion de Movimientos Humanoides con Dinamica Inversa Jerarquica, <\/span><span class=\"tp_pub_additional_publisher\">Congreso Latinoamericano de Control Autom\u00e1tico CLCA 2014, <\/span><span class=\"tp_pub_additional_year\">2014<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_resource_link\"><a id=\"tp_links_sh_78\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('78','tp_links')\" title=\"Mostrar enlaces y recursos\" style=\"cursor:pointer;\">Enlaces<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_78\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('78','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_78\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{Castillo2014,<br \/>\r\ntitle = {Generacion de Movimientos Humanoides con Dinamica Inversa Jerarquica},<br \/>\r\nauthor = {Estopier-Castillo, Vicente and Arechavaleta, Gustavo and Olgu\\'{i}n-D\\'{i}az, Ernesto},<br \/>\r\nurl = {http:\/\/amca.mx\/memorias\/amca2014\/articulos\/0112.pdf},<br \/>\r\nyear  = {2014},<br \/>\r\ndate = {2014-00-00},<br \/>\r\nbooktitle = {Generacion de Movimientos Humanoides con Dinamica Inversa Jerarquica},<br \/>\r\npublisher = {Congreso Latinoamericano de Control Autom\\'{a}tico CLCA 2014},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('78','tp_bibtex')\">Cerrar<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_78\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-file-pdf\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/amca.mx\/memorias\/amca2014\/articulos\/0112.pdf\" title=\"http:\/\/amca.mx\/memorias\/amca2014\/articulos\/0112.pdf\" target=\"_blank\">http:\/\/amca.mx\/memorias\/amca2014\/articulos\/0112.pdf<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('78','tp_links')\">Cerrar<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_2012\">2012<\/h3><h3 class=\"tp_h3\" id=\"tp_h3_conference\">Conferencias<\/h3><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Delfin, Josafat;  Mar, Oscar;  Hayet, Jean Bernard;  Castelan, Mario;  Arechavaleta, Gustavo<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('26','tp_links')\" style=\"cursor:pointer;\">An active strategy for the simultaneous localization and reconstruction of a 3D object from a humanoid platform<\/a> <span class=\"tp_pub_type tp_  conference\">Conferencia<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">2012 12th IEEE-RAS International Conference on Humanoid Robots (Humanoids 2012), <\/span><span class=\"tp_pub_additional_year\">2012<\/span>, <span class=\"tp_pub_additional_issn\">ISSN: 2164-0572<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_26\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('26','tp_abstract')\" title=\"Mostrar resumen\" style=\"cursor:pointer;\">Resumen<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_26\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('26','tp_links')\" title=\"Mostrar enlaces y recursos\" style=\"cursor:pointer;\">Enlaces<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_26\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('26','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_26\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{6651548,<br \/>\r\ntitle = {An active strategy for the simultaneous localization and reconstruction of a 3D object from a humanoid platform},<br \/>\r\nauthor = {Delfin, Josafat and Mar, Oscar and Hayet, Jean Bernard and Castelan, Mario and Arechavaleta, Gustavo },<br \/>\r\nurl = {http:\/\/ieeexplore.ieee.org\/document\/6651548\/?arnumber=6651548},<br \/>\r\ndoi = {10.1109\/HUMANOIDS.2012.6651548},<br \/>\r\nissn = {2164-0572},<br \/>\r\nyear  = {2012},<br \/>\r\ndate = {2012-11-01},<br \/>\r\nbooktitle = {2012 12th IEEE-RAS International Conference on Humanoid Robots (Humanoids 2012)},<br \/>\r\npages = {384-389},<br \/>\r\nabstract = {This paper proposes a novel approach for the simultaneous localization and reconstruction of a 3D object. This reconstruction is done with a small humanoid platform by the only means of monocular vision.We combine (1) a stochastic control approach to decide how to move the robot such that the object can be both localized with respect to the robot and correctly reconstructed, and (2) a space carving approach for the 3D reconstruction of the object that merges the acquired monocular views into a coherent model. We present a set of very promising results obtained with a NAO platform.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('26','tp_bibtex')\">Cerrar<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_26\" style=\"display:none;\"><div class=\"tp_abstract_entry\">This paper proposes a novel approach for the simultaneous localization and reconstruction of a 3D object. This reconstruction is done with a small humanoid platform by the only means of monocular vision.We combine (1) a stochastic control approach to decide how to move the robot such that the object can be both localized with respect to the robot and correctly reconstructed, and (2) a space carving approach for the 3D reconstruction of the object that merges the acquired monocular views into a coherent model. We present a set of very promising results obtained with a NAO platform.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('26','tp_abstract')\">Cerrar<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_26\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/ieeexplore.ieee.org\/document\/6651548\/?arnumber=6651548\" title=\"http:\/\/ieeexplore.ieee.org\/document\/6651548\/?arnumber=6651548\" target=\"_blank\">http:\/\/ieeexplore.ieee.org\/document\/6651548\/?arnumber=6651548<\/a><\/li><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/10.1109\/HUMANOIDS.2012.6651548\" title=\"DOI de seguimiento:10.1109\/HUMANOIDS.2012.6651548\" target=\"_blank\">doi:10.1109\/HUMANOIDS.2012.6651548<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('26','tp_links')\">Cerrar<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_2010\">2010<\/h3><h3 class=\"tp_h3\" id=\"tp_h3_conference\">Conferencias<\/h3><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Ramirez-Bejarano, Christian;  Castelan, Mario;  Arechavaleta, Gustavo<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('32','tp_links')\" style=\"cursor:pointer;\">Multilinear decomposition of human walking paths<\/a> <span class=\"tp_pub_type tp_  conference\">Conferencia<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">2010 10th IEEE-RAS International Conference on Humanoid Robots, <\/span><span class=\"tp_pub_additional_year\">2010<\/span>, <span class=\"tp_pub_additional_issn\">ISSN: 2164-0572<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_32\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('32','tp_abstract')\" title=\"Mostrar resumen\" style=\"cursor:pointer;\">Resumen<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_32\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('32','tp_links')\" title=\"Mostrar enlaces y recursos\" style=\"cursor:pointer;\">Enlaces<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_32\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('32','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_32\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{5686313,<br \/>\r\ntitle = {Multilinear decomposition of human walking paths},<br \/>\r\nauthor = {Ramirez-Bejarano, Christian and Castelan, Mario and Arechavaleta, Gustavo },<br \/>\r\nurl = {http:\/\/ieeexplore.ieee.org\/document\/5686313\/?arnumber=5686313},<br \/>\r\ndoi = {10.1109\/ICHR.2010.5686313},<br \/>\r\nissn = {2164-0572},<br \/>\r\nyear  = {2010},<br \/>\r\ndate = {2010-12-01},<br \/>\r\nbooktitle = {2010 10th IEEE-RAS International Conference on Humanoid Robots},<br \/>\r\npages = {492-497},<br \/>\r\nabstract = {In a previous work, the authors have shown how the Principal Components Analysis (PCA) of a set of human walking paths provides sufficient information to derive a linear human-like path generator based on examples. The present work aims to provide an analysis of human walking paths from the perspective of multilinear algebra, using the n-mode Singular Value Decomposition (SVD). This decomposition subsumes the conventional PCA and extends its capabilities to study the interaction between the different vector spaces attached to the data. The training set consists of sixty routes performed three times by five different subjects for a total of 900 examples. The outcome of this decomposition is a novel multilinear model capable of revealing the global shape variations between the different vector spaces, particularly, exhibiting the planning strategies that each subject seemed to use to approach to the different targets. In addition, we sketch how the multilinear model may be used to synthesize new routes for the hypothetical case of a subject trying to reach a moving target.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('32','tp_bibtex')\">Cerrar<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_32\" style=\"display:none;\"><div class=\"tp_abstract_entry\">In a previous work, the authors have shown how the Principal Components Analysis (PCA) of a set of human walking paths provides sufficient information to derive a linear human-like path generator based on examples. The present work aims to provide an analysis of human walking paths from the perspective of multilinear algebra, using the n-mode Singular Value Decomposition (SVD). This decomposition subsumes the conventional PCA and extends its capabilities to study the interaction between the different vector spaces attached to the data. The training set consists of sixty routes performed three times by five different subjects for a total of 900 examples. The outcome of this decomposition is a novel multilinear model capable of revealing the global shape variations between the different vector spaces, particularly, exhibiting the planning strategies that each subject seemed to use to approach to the different targets. In addition, we sketch how the multilinear model may be used to synthesize new routes for the hypothetical case of a subject trying to reach a moving target.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('32','tp_abstract')\">Cerrar<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_32\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/ieeexplore.ieee.org\/document\/5686313\/?arnumber=5686313\" title=\"http:\/\/ieeexplore.ieee.org\/document\/5686313\/?arnumber=5686313\" target=\"_blank\">http:\/\/ieeexplore.ieee.org\/document\/5686313\/?arnumber=5686313<\/a><\/li><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/10.1109\/ICHR.2010.5686313\" title=\"DOI de seguimiento:10.1109\/ICHR.2010.5686313\" target=\"_blank\">doi:10.1109\/ICHR.2010.5686313<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('32','tp_links')\">Cerrar<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_2009\">2009<\/h3><h3 class=\"tp_h3\" id=\"tp_h3_conference\">Conferencias<\/h3><div class=\"tp_publication tp_publication_conference\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\"> Castelan, Mario;  Arechavaleta, Gustavo<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('34','tp_links')\" style=\"cursor:pointer;\">Approximating the reachable space of human walking paths: a low dimensional linear approach<\/a> <span class=\"tp_pub_type tp_  conference\">Conferencia<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_booktitle\">2009 9th IEEE-RAS International Conference on Humanoid Robots, <\/span><span class=\"tp_pub_additional_year\">2009<\/span>, <span class=\"tp_pub_additional_issn\">ISSN: 2164-0572<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_34\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('34','tp_abstract')\" title=\"Mostrar resumen\" style=\"cursor:pointer;\">Resumen<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_34\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('34','tp_links')\" title=\"Mostrar enlaces y recursos\" style=\"cursor:pointer;\">Enlaces<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_34\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('34','tp_bibtex')\" title=\"Mostrar entrada BibTeX \" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_34\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@conference{5379595,<br \/>\r\ntitle = {Approximating the reachable space of human walking paths: a low dimensional linear approach},<br \/>\r\nauthor = {Castelan, Mario and Arechavaleta, Gustavo },<br \/>\r\nurl = {http:\/\/ieeexplore.ieee.org\/document\/5379595\/?arnumber=5379595},<br \/>\r\ndoi = {10.1109\/ICHR.2009.5379595},<br \/>\r\nissn = {2164-0572},<br \/>\r\nyear  = {2009},<br \/>\r\ndate = {2009-12-01},<br \/>\r\nbooktitle = {2009 9th IEEE-RAS International Conference on Humanoid Robots},<br \/>\r\npages = {81-86},<br \/>\r\nabstract = {In this work, we aim to exhibit the geometrical shape primitives of human walking trajectories using a statistical model constructed through Principal Component Analysis. This analysis provides sufficient information to derive a linear human-like path generator based on examples. The examples are provided by a motion capture database of human walking trajectories. The proposed model captures the shape of trajectories in terms of path length and deformation. We have successfully applied our model to compute a good approximation of the reachable space of human walking. This can be done with a negligible computational cost since it is based on a linear combination of basis human paths.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {conference}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('34','tp_bibtex')\">Cerrar<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_34\" style=\"display:none;\"><div class=\"tp_abstract_entry\">In this work, we aim to exhibit the geometrical shape primitives of human walking trajectories using a statistical model constructed through Principal Component Analysis. This analysis provides sufficient information to derive a linear human-like path generator based on examples. The examples are provided by a motion capture database of human walking trajectories. The proposed model captures the shape of trajectories in terms of path length and deformation. We have successfully applied our model to compute a good approximation of the reachable space of human walking. This can be done with a negligible computational cost since it is based on a linear combination of basis human paths.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('34','tp_abstract')\">Cerrar<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_34\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/ieeexplore.ieee.org\/document\/5379595\/?arnumber=5379595\" title=\"http:\/\/ieeexplore.ieee.org\/document\/5379595\/?arnumber=5379595\" target=\"_blank\">http:\/\/ieeexplore.ieee.org\/document\/5379595\/?arnumber=5379595<\/a><\/li><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/10.1109\/ICHR.2009.5379595\" title=\"DOI de seguimiento:10.1109\/ICHR.2009.5379595\" target=\"_blank\">doi:10.1109\/ICHR.2009.5379595<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('34','tp_links')\">Cerrar<\/a><\/p><\/div><\/div><\/div><\/div><\/div> Av. Industrial\u00a0Metalurgia\u00a0#1062,\u00a0Parque Ind. Ramos Arizpe,\u00a0Ramos Arizpe, Coah.\u00a0C.P. 25900, M\u00e9xico. \u00a0Tel. +52 (844) 438-9600<\/p>\n","protected":false},"author":14,"featured_media":0,"parent":0,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"","meta":{"_et_pb_use_builder":"on","_et_pb_old_content":"","_et_gb_content_width":"","footnotes":""},"class_list":["post-76","page","type-page","status-publish","hentry"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v27.3 - https:\/\/yoast.com\/product\/yoast-seo-wordpress\/ -->\n<title>Publicaciones - Prof. Dr. Gustavo Arechavaleta<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/ryma.cinvestav.mx\/garechav\/publicaciones\/\" \/>\n<meta property=\"og:locale\" content=\"es_ES\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Publicaciones - Prof. Dr. Gustavo Arechavaleta\" \/>\n<meta property=\"og:description\" content=\"Para ver las publicaciones de todo Rob\u00f3tica y Manufactura Avanzada, ver:\u00a0 Publicaciones RYMA Av. Industrial\u00a0Metalurgia\u00a0#1062,\u00a0Parque Ind. Ramos Arizpe,\u00a0Ramos Arizpe, Coah.\u00a0C.P. 25900, M\u00e9xico. \u00a0Tel. +52 (844) 438-9600\" \/>\n<meta property=\"og:url\" content=\"https:\/\/ryma.cinvestav.mx\/garechav\/publicaciones\/\" \/>\n<meta property=\"og:site_name\" content=\"Prof. Dr. Gustavo Arechavaleta\" \/>\n<meta property=\"article:modified_time\" content=\"2017-09-11T15:36:06+00:00\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"Tiempo de lectura\" \/>\n\t<meta name=\"twitter:data1\" content=\"3 minutos\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\\\/\\\/schema.org\",\"@graph\":[{\"@type\":\"WebPage\",\"@id\":\"https:\\\/\\\/ryma.cinvestav.mx\\\/garechav\\\/publicaciones\\\/\",\"url\":\"https:\\\/\\\/ryma.cinvestav.mx\\\/garechav\\\/publicaciones\\\/\",\"name\":\"Publicaciones - Prof. Dr. Gustavo Arechavaleta\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/ryma.cinvestav.mx\\\/garechav\\\/#website\"},\"datePublished\":\"2017-09-10T05:46:36+00:00\",\"dateModified\":\"2017-09-11T15:36:06+00:00\",\"breadcrumb\":{\"@id\":\"https:\\\/\\\/ryma.cinvestav.mx\\\/garechav\\\/publicaciones\\\/#breadcrumb\"},\"inLanguage\":\"es\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\\\/\\\/ryma.cinvestav.mx\\\/garechav\\\/publicaciones\\\/\"]}]},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\\\/\\\/ryma.cinvestav.mx\\\/garechav\\\/publicaciones\\\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Portada\",\"item\":\"https:\\\/\\\/ryma.cinvestav.mx\\\/garechav\\\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Publicaciones\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\\\/\\\/ryma.cinvestav.mx\\\/garechav\\\/#website\",\"url\":\"https:\\\/\\\/ryma.cinvestav.mx\\\/garechav\\\/\",\"name\":\"Prof. Dr. Gustavo Arechavaleta\",\"description\":\"Miembro de Rob\u00f3tica y Manufactura Avanzada - Cinvestav\",\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\\\/\\\/ryma.cinvestav.mx\\\/garechav\\\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"es\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Publicaciones - Prof. Dr. Gustavo Arechavaleta","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/ryma.cinvestav.mx\/garechav\/publicaciones\/","og_locale":"es_ES","og_type":"article","og_title":"Publicaciones - Prof. Dr. Gustavo Arechavaleta","og_description":"Para ver las publicaciones de todo Rob\u00f3tica y Manufactura Avanzada, ver:\u00a0 Publicaciones RYMA Av. Industrial\u00a0Metalurgia\u00a0#1062,\u00a0Parque Ind. Ramos Arizpe,\u00a0Ramos Arizpe, Coah.\u00a0C.P. 25900, M\u00e9xico. \u00a0Tel. +52 (844) 438-9600","og_url":"https:\/\/ryma.cinvestav.mx\/garechav\/publicaciones\/","og_site_name":"Prof. Dr. Gustavo Arechavaleta","article_modified_time":"2017-09-11T15:36:06+00:00","twitter_card":"summary_large_image","twitter_misc":{"Tiempo de lectura":"3 minutos"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"WebPage","@id":"https:\/\/ryma.cinvestav.mx\/garechav\/publicaciones\/","url":"https:\/\/ryma.cinvestav.mx\/garechav\/publicaciones\/","name":"Publicaciones - Prof. Dr. Gustavo Arechavaleta","isPartOf":{"@id":"https:\/\/ryma.cinvestav.mx\/garechav\/#website"},"datePublished":"2017-09-10T05:46:36+00:00","dateModified":"2017-09-11T15:36:06+00:00","breadcrumb":{"@id":"https:\/\/ryma.cinvestav.mx\/garechav\/publicaciones\/#breadcrumb"},"inLanguage":"es","potentialAction":[{"@type":"ReadAction","target":["https:\/\/ryma.cinvestav.mx\/garechav\/publicaciones\/"]}]},{"@type":"BreadcrumbList","@id":"https:\/\/ryma.cinvestav.mx\/garechav\/publicaciones\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Portada","item":"https:\/\/ryma.cinvestav.mx\/garechav\/"},{"@type":"ListItem","position":2,"name":"Publicaciones"}]},{"@type":"WebSite","@id":"https:\/\/ryma.cinvestav.mx\/garechav\/#website","url":"https:\/\/ryma.cinvestav.mx\/garechav\/","name":"Prof. Dr. Gustavo Arechavaleta","description":"Miembro de Rob\u00f3tica y Manufactura Avanzada - Cinvestav","potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/ryma.cinvestav.mx\/garechav\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"es"}]}},"_links":{"self":[{"href":"https:\/\/ryma.cinvestav.mx\/garechav\/wp-json\/wp\/v2\/pages\/76","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/ryma.cinvestav.mx\/garechav\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/ryma.cinvestav.mx\/garechav\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/ryma.cinvestav.mx\/garechav\/wp-json\/wp\/v2\/users\/14"}],"replies":[{"embeddable":true,"href":"https:\/\/ryma.cinvestav.mx\/garechav\/wp-json\/wp\/v2\/comments?post=76"}],"version-history":[{"count":25,"href":"https:\/\/ryma.cinvestav.mx\/garechav\/wp-json\/wp\/v2\/pages\/76\/revisions"}],"predecessor-version":[{"id":402,"href":"https:\/\/ryma.cinvestav.mx\/garechav\/wp-json\/wp\/v2\/pages\/76\/revisions\/402"}],"wp:attachment":[{"href":"https:\/\/ryma.cinvestav.mx\/garechav\/wp-json\/wp\/v2\/media?parent=76"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}