{"id":78,"date":"2018-11-16T21:15:08","date_gmt":"2018-11-17T02:15:08","guid":{"rendered":"https:\/\/carl-unh.com\/\/?page_id=78"},"modified":"2023-09-29T14:29:27","modified_gmt":"2023-09-29T14:29:27","slug":"research","status":"publish","type":"page","link":"https:\/\/carl.cs.unh.edu\/index.php\/research\/","title":{"rendered":"Research &#038; Publications"},"content":{"rendered":"\n<p><\/p>\n\n\n<h3><strong>Learning Temporal Features from Video Demonstrations<\/strong><\/h3>\n<hr \/>\n<p>Temporal features, relationships between spatial features across time, are integral to understanding many of the tasks that humans engage in daily. Unfortunately, learning temporal features directly from video is a complex problem and contemporary deep learning-based models are duration-specific making them vulnerable to variations in the duration over which activities are expressed and limiting the scale of the data they can model. I address an approach that can overcome these challenges when capturing temporal features from video of human-led demonstrations of sequential tasks. Inference using these temporal features is then leveraged for better performance on policy learning and activity recognition tasks.<\/p>\n<p><a href=\"https:\/\/youtu.be\/seVCV3UGPUQ\">https:\/\/youtu.be\/seVCV3UGPUQ<\/a><\/p>\n<p><strong>Sponsor:<\/strong> National Science Foundation IIS 1664554<\/p>\n<p><strong style=\"font-size: inherit;\">Investigator: <\/strong><span style=\"font-size: inherit;\">\u00a0Madison Clark-Turner<\/span><\/p>\n<p><strong>Publications:<\/strong><\/p>\n<ol>\n<li>Hussein Mostafa*, Madison Clark-Turner*, and Begum Momotaz. <a href=\"https:\/\/www.cs.unh.edu\/~mbegum\/pub\/BC_ITR_2022.pdf\">A Hierarchical Approach for Learning Multi-Step Sequential Tasks from Video Demonstrations<\/a> (in preparation).<\/li>\n<li>Madison Clark-Turner and Begum Momotaz. <a href=\"https:\/\/www.cs.unh.edu\/~mbegum\/pub\/ITR_2022_Madison.pdf\">Understanding Temporal Relations from Video: A Pathway to Learning Sequential Tasks from Visual Demonstrations<\/a> (in preparation).<\/li>\n<\/ol>\n<h3><strong>Policy Learning for Sequential Tasks from Demonstrations<\/strong><\/h3>\n<hr \/>\n<p>We are interested in learning the high-level policy of multi-step sequential (MSS) tasks, such as activities of daily living, from video demonstrations. Videos of MSS tasks are typically long in duration and exhibit large feature variance, especially when captured in non-engineered settings. Learning task policy from such videos using state-of-the-art end-to-end approaches is sample inefficient due to a reliance on pixel-level information. Understanding the unique temporal structures of MSS tasks can make policy learning easier and sample efficient. However, understanding this temporal structure requires analyzing the entire content of the video or a task which is a complex and under-explored area in the current literature. We propose a hierarchical solution to this problem where i) an automated feature selection process extrapolates temporally grounded task-relevant features from the video and then ii) a stochastic policy learning model learns a feature-constrained task policy. The proposed model is sample efficient as a result of substituting selected temporally-grounded features for pixel-level information. We demonstrate the efficacy of our proposed framework by teaching a YuMi robot a tea making task from videos and then comparing our approach&#8217;s performance to a behavioral cloning baseline.<\/p>\n<p>https:\/\/youtu.be\/DjUgI2-aMmo<\/p>\n<p><strong>Sponsor:<\/strong> National Science Foundation IIS 1830597<\/p>\n<p><strong>Investigator: <\/strong>\u00a0Mostafa Hussein<\/p>\n<p><strong>Publications:<\/strong><\/p>\n<ol>\n<li style=\"text-align: left;\">Mostafa Hussein, Brendan Crowe, Madison Clark-Turner, Petrik Marek, and Begum Momotaz. <a href=\"https:\/\/www.cs.unh.edu\/~mbegum\/pub\/BC_IROS_2021.pdf\">Robust Behavior Cloning with Adversarial Demonstration Detection<\/a><a href=\"https:\/\/www.cs.unh.edu\/~mbegum\/pub\/ BC_IROS_2021.pdf\" data-wplink-url-error=\"true\">,<\/a> IROS 2021.<\/li>\n<li style=\"text-align: left;\">Hussein Mostafa, Brendan Crowe, Madison Clark-Turner, Petrik Marek, and Begum Momotaz. <a href=\"https:\/\/arxiv.org\/abs\/2101.01251\">Robust Maximum Entropy Behavior cloning<\/a>, NeurIPS Workshop on Robot Learning, 2020).<\/li>\n<li style=\"text-align: left;\">Hussein Mostafa, Madison Clark-Turner, and Begum Momotaz. <a href=\"https:\/\/openreview.net\/forum?id=tyXMrpEMkDb\"><span dir=\"ltr\" role=\"presentation\">Detecting Incorrect Visual Demonstrations for <\/span><\/a><span dir=\"ltr\" role=\"presentation\">Improved Policy Learning, <\/span>CoRL 2022, NZ.<\/li>\n<\/ol>\n<h3><strong>Trajectory Learning from Demonstrations<\/strong><\/h3>\n<hr \/>\n<p>This project is focused on trajectory learning from demonstration (LfD) for assistive robotics. The primary objectives of trajectory LfD are to learn motion tasks from demonstrations and exhibit robustness to spatial and temporal perturbations. Depending on the operational environment, the specific objectives of LfD algorithms may very. For rehabilitation robotics, a robot therapist must demonstrate proper exercise movements and provide feedback on a human\u2019s exercise execution. In a workplace setting, a robot collaborator should fluently hand over objects to its human counterpart. In all applications, the LfD algorithm must provide a feasible trajectory for the robot to execute, ideally in the form of a controller. The key to success in multi-purpose trajectory LfD lies in 1) controller expressivity, which is the set of trajectories that can be reproduced, and 2) constraint enforcement, which eliminates infeasible trajectories.<\/p>\n<p>Three different learning strategies\u00a0 have been developed, namely controller learning via regression, controller learning via optimization, and controller learning via inverse optimal control. These approaches have been published in\u00a0 ICRA 2019, IROS 2020, IROS 2021. Videos from each publication are shown below.<\/p>\n<p><iframe loading=\"lazy\" title=\"ICRA 2019 Video\" width=\"640\" height=\"360\" src=\"https:\/\/www.youtube.com\/embed\/t306yQfLc_c?start=51&#038;feature=oembed\" frameborder=\"0\" allow=\"accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share\" referrerpolicy=\"strict-origin-when-cross-origin\" allowfullscreen><\/iframe><\/p>\n<p><iframe loading=\"lazy\" title=\"IROS 2020\" width=\"640\" height=\"360\" src=\"https:\/\/www.youtube.com\/embed\/-QhBj1Dnhac?feature=oembed\" frameborder=\"0\" allow=\"accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share\" referrerpolicy=\"strict-origin-when-cross-origin\" allowfullscreen><\/iframe><\/p>\n<p><iframe loading=\"lazy\" title=\"IROS 2021\" width=\"640\" height=\"360\" src=\"https:\/\/www.youtube.com\/embed\/w4Wafe9I_V0?feature=oembed\" frameborder=\"0\" allow=\"accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share\" referrerpolicy=\"strict-origin-when-cross-origin\" allowfullscreen><\/iframe><\/p>\n<p><strong>Sponsor:<\/strong> National Science Foundation IIS\u00a01830597, <a href=\"https:\/\/www.unh.edu\/research\/collaborative-research-excellence-core-initiative\">UNH CoRE initiative 2018, 2017<\/a><\/p>\n<p><strong>Investigator: <\/strong>\u00a0Paul Gesel<\/p>\n<p><strong>Publications:<\/strong><\/p>\n<div>1. Paul Gesel, Noushad Sojib and Momotaz Begum, Self-Supervised Visual Motor Skills Via Neural Radiance Fields, IROS 2023<\/div>\n<div>\u00a0<\/div>\n<div>2. Paul Gesel and Momotaz Begum. <a href=\"https:\/\/ieeexplore.ieee.org\/document\/10161237\">Learning Stable Dynamics via Iterative\u00a0 \u00a0 \u00a0 \u00a0 Quadratic Programming<\/a>, under review, ICRA 2023<\/div>\n<div>\u00a0<\/div>\n<div>3. Paul Gesel, D. LaRoche, S. Arthanat, and M. Begum, <a href=\"https:\/\/www.cs.unh.edu\/~mbegum\/pub\/PSM_IROS_2021.pdf\">Learning to Optimize\u00a0 \u00a0Control Policies and Evaluate Reproduction Performance from Human Demonstration<\/a><em>,\u00a0<\/em>IROS 2021\u00a0<\/div>\n<div>\u00a0<\/div>\n<div>4. Paul Gesel, D. LaRoche, S. Arthanat, and M. Begum, <a href=\"https:\/\/www.cs.unh.edu\/~mbegum\/pub\/PSM_IROS2020.pdf\">Learning Adaptive\u00a0 Human Motion viaPhase Space Analysis of Demonstrated Trajectories<\/a> \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 IROS 2020\u00a0<\/div>\n<div>\u00a0<\/div>\n<div>5. Paul Gesel, M. Begum, and D. LaRoche, <a href=\"https:\/\/www.cs.unh.edu\/~mbegum\/pub\/PSM_ICRA_2019.pdf\">Learning Motion Trajectories\u00a0 \u00a0from Phase Space Analysis of the Demonstration<\/a><em>,\u00a0<\/em>ICRA 2019\u00a0<\/div>\n<h3><strong>Care Robot for Older Adults with Dementia<\/strong><\/h3>\n<hr \/>\n<p>This project is focused on designing an autonomous mobile robot capable of helping older adults with dementia to age in place through helping them in health-care and home safety related activities. We designed a preliminary prototype that leverages PDDL to generate a set of care plans on the fly based on real-time sensor inputs.\u00a0<\/p>\n<p><iframe loading=\"lazy\" title=\"UNH Smart Home Socially Assistive Robot\" width=\"640\" height=\"360\" src=\"https:\/\/www.youtube.com\/embed\/YRrHj6lVp2A?feature=oembed\" frameborder=\"0\" allow=\"accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share\" referrerpolicy=\"strict-origin-when-cross-origin\" allowfullscreen><\/iframe><\/p>\n<p><strong>Investigators: <\/strong>\u00a0This is a multidisciplinary project involving Tianyi Gu (CS), Sajay Arthanat (OT), Dain LaRoche (ES), and Dongpeng Xu (CS)<\/p>\n<p><strong>Publications:<\/strong><\/p>\n<ol>\n<li>S. Arthanat, M. Begum, T. Gu, D. P LaRoche, D. Xu, and N.Zhang, <a href=\"https:\/\/www.cs.unh.edu\/~mbegum\/pub\/SH_DR_2020.pdf\"><em>Caregiver perspectives on a smart home-based socially assistive robot for individuals with Alzheimer\u2019s disease and related dementia<\/em><\/a>, Disability and Rehabilitation: Assistive Technology, 1-10, 2020<\/li>\n<li>T. Gu, M. Begum, N. Zhang, D. Xu, S. Arthanat and D. LaRoche, <a href=\"https:\/\/icaps20subpages.icaps-conference.org\/wp-content\/uploads\/2020\/10\/10-PlanRob_2020_paper_8.pdf\"><em>Adaptive Software Framework for Dementia-care Robot<\/em><\/a>, Workshop on Planning and Robotics, International Conference on Automated Planning and Scheduling, 2020<\/li>\n<\/ol>\n<p><strong>Sponsor:<\/strong> \u00a0<a href=\"https:\/\/www.unh.edu\/research\/collaborative-research-excellence-core-initiative\">UNH CoRE initiative 2019<\/a><\/p>\n<h3><strong>Deep Q-learning of Human-Robot Interaction for Robot-mediated Behavior Intervention<\/strong><\/h3>\n<hr \/>\n<p>This research is focused on merging and transitioning deep learning-based Learning from Demonstrations (LfD) techniques to real world human-robot interaction (HRI) applications. We have designed a deep Q network that can learn a structured sequential task from demonstration data. We have tested the network to learn a behavioral intervention where the goal is to teach children with developmental delays (such as autism spectrum disorders) how to respond to a greeting in a socially appropriate manner, i.e. a <em>social greeting intervention<\/em>. The intervention follows the principle of applied behavior analysis (ABA). The video below shows an example of this work (published in HRI 2018)<\/p>\n<p>Since deep networks require a massive amount of data for training, a criteria difficult to meet by most HRI problems, we are currently investigating on designing a compact representation of long duration video data using interval algebraic relationships among events observed in the video.<\/p>\n<h4>\u00a0<\/h4>\n<p><a href=\"https:\/\/youtu.be\/qAJtr0AShbI\">https:\/\/youtu.be\/qAJtr0AShbI<\/a><\/p>\n<p><strong>Investigator:\u00a0<\/strong>\u00a0Madison Clark-Turner<\/p>\n<p><strong>Publications:<\/strong><\/p>\n<ol>\n<li>M. Turner and M. Begum, <a href=\"https:\/\/www.cs.unh.edu\/~mbegum\/pub\/DQN_HRI_2018.pdf\"><em>Deep Reinforcement Learning of Abstract Reasoning from Demonstration<\/em><\/a>, ACM\/IEEE International Conference on Human Robot Interaction, 2018<\/li>\n<li>M. Turner and M. Begum, <a href=\"https:\/\/www.cs.unh.edu\/~mbegum\/pub\/DQN_ROMAN_2017.pdf\"><em>Deep Recurrent Q-Learning of Behavioral Intervention Delivery by a Robot from Demonstration Data<\/em><\/a>, IEEE International Symposium on Robot and Human Interactive Communications, 2017<\/li>\n<li>M. Turner and M. Begum, <a href=\"http:\/\/users.eecs.northwestern.edu\/~sjq751\/hcr_crv\/Clark-Turner_HCR_RSS17.pdf\">Learning to Deliver Robot-Mediated Behavioral Intervention<\/a>, Workshop on Human-centered Robotics, Robotics Science and Systems 2017<\/li>\n<\/ol>\n<p><strong>Sponsor:<\/strong> National Science Foundation IIS\u00a01664554<\/p>\n<h3><strong>Reward Function Learning from Demonstrations of Sequential Tasks<\/strong><\/h3>\n<hr \/>\n<p>One way to learn long duration sequential tasks from demonstrations is to learn the reward function of the demonstrator. The goal of this project is to learn the reward function in partially observable environments. Learning the reward function of a Partially Observable Markov Decision Process (POMDP) is not a well understood problem. Alternatively, we propose to reduce the underlying POMDP to an MDP and and extract the reward function using an efficient MDP-IRL algorithm. Our extensive experiments suggest that the reward function learned this way generates POMDP policies that mimic the policies of the demonstrator well. We tested this approach to learn the reward function of a mock therapist as s\/he delivers an ABA-style social greeting intervention. The video below shows a robot delivering different steps of the intervention learned through learning the reward structures from a set of demonstrations.<\/p>\n<p><strong>Investigator:<\/strong> Mostafa Hussein<\/p>\n<p><strong>Publications:<\/strong><\/p>\n<p>Mostafa Hussein, Momotaz Begum, Marek Petrik, <a href=\"https:\/\/www.cs.unh.edu\/~mbegum\/pub\/IRL_ICRA_2019.pdf\">Inverse Reinforcement Learning of Interaction Dynamics from Demonstrations<\/a>, ICRA 2019.<\/p>\n<p><strong>Sponsor:<\/strong> National Science Foundation IIS\u00a01830597<\/p>\n<h3><strong>Activity Recognition using Interval Temporal Relations among Shallow Features<\/strong><\/h3>\n<hr \/>\n<p>This project investigates the possibility of replacing pixel-based representation of state space in deep learning with a novel interval temporal relations-based representation among observed events in the video. We are currently conducting rigorous testing of the performance of Interval Attribute Descriptors (IAD) generated by 3D Convolutional Neural Networks as a method to represent the interval temporal relationships of human activities in publicly available video data-sets such as UCF101.<\/p>\n<p><strong>Investigator: <\/strong>\u00a0Jordan Chadwick<\/p>\n<h3><strong>Temporal Context Graph for Policy Selection in Sequential Tasks<\/strong><\/h3>\n<hr \/>\n<p>Complex everyday activities often possess rich temporal patterns, along with spatial features, that can be exploited to learn such tasks from a limited number of demonstrations. This projects develops a temporal context graph (TCG) based on Allen&#8217;s interval temporal Algebra that helps to select the correct policy based on the learned temporal structure of a task. TCG is a capable of dealing with tasks with cyclical atomic actions and consisting of sequential and parallel temporal relations. The video below shows the performance of TCG in learning two sequential tasks from demonstrations.<\/p>\n<p><iframe loading=\"lazy\" title=\"Leveraging Temporal Reasoning for Policy Selection in Learning from Demonstration\" width=\"640\" height=\"360\" src=\"https:\/\/www.youtube.com\/embed\/y4olN-pOOMk?feature=oembed\" frameborder=\"0\" allow=\"accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share\" referrerpolicy=\"strict-origin-when-cross-origin\" allowfullscreen><\/iframe><\/p>\n<p><strong>Investigator: <\/strong>\u00a0Estuardo Carpio, Madison Clark-Turner<\/p>\n<p><strong>Publications:<\/strong><\/p>\n<div>Estuardo Carpio, Madison Clark-Turner, Paul Gesel, and M. Begum, <em><a href=\"https:\/\/www.cs.unh.edu\/~mbegum\/pub\/TCG_ICRA_2019.pdf\">Leveraging Temporal Reasoning for Policy Selection in Learning from Demonstration<\/a>,\u00a0<\/em>ICRA 2019\u00a0<\/div>\n<div>\u00a0<\/div>\n<div>Estuardo Carpio, Madison Clark-Turner, Paul Gesel, and M. Begum, <a href=\"https:\/\/www.cs.unh.edu\/~mbegum\/pub\/TCG_RoMan_2019.pdf\">Learning Sequential Human-Robot Interaction Tasks from Demonstrations: The Role of Temporal Reasoning<\/a>, RoMan 2019<\/div>\n<p><strong>Sponsor:<\/strong> National Science Foundation IIS\u00a01664554,\u00a0<a href=\"https:\/\/www.unh.edu\/research\/collaborative-research-excellence-core-initiative\">UNH CoRE initiative<\/a><\/p>\n<h3><strong>Exercise Sequence Learning from Demonstrations<\/strong><\/h3>\n<hr \/>\n<p><strong>Publications:<\/strong><\/p>\n<ol>\n<li>A. Lydakis, Y. Meng, C. Munroe, Yi-Ning Wu, M. Begum, <a href=\"https:\/\/www.cs.unh.edu\/~mbegum\/pub\/MDP_ICORR_2017.pdf\">A Learning-based Agent for Home Neurorehabilitation<\/a>, IEEE International Conference or Rehabilitation Robotics, ICORR 2017, Pubmed ID: 28813990<\/li>\n<li>A. Lydakis, Pei-Chun Kao, and M. Begum, <a href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3056540.3056555\">Irregular Gait Detection using Wearable Sensors<\/a>,\u00a0 PErvasive Technologies Related to Assistive Environments (PETRA) 2017<\/li>\n<li>Y. Meng, C. Munroe, Y. Wu, and M. Begum, <a href=\"https:\/\/www.cs.unh.edu\/~mbegum\/pub\/MDP_RoMAN_2016.pdf\">Learning from Demonstration Framework to Promote Home-based Neuromotor Rehabilitation<\/a>, IEEE International Symposium on Robot and Human Interactive Communication, RoMAN 2016<\/li>\n<li>C. Munroe, Y. Meng, H. Yanco, and M. Begum, <a href=\"https:\/\/www.youtube.com\/watch?v=vuemkioDDhI\">Augmented Reality Eyeglasses for Promoting Home-Based Rehabilitation for<\/a><br \/>Children with Cerebral Palsy, ACM\/IEEE International Conference on Human-Robot Interaction, HRI 2016 (video submission)<\/li>\n<\/ol>\n\n\n<h2 class=\"wp-block-heading\">Publications from <strong>Momotaz&#8217;s Previous <a href=\"https:\/\/momotazbegumonu.wixsite.com\/mypage\/cv\">Research<\/a><\/strong><\/h2>\n\n\n\n<p><a href=\"https:\/\/momotazbegumonu.wixsite.com\/mypage\/publications\">Publications<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Learning Temporal Features from Video Demonstrations Temporal features, relationships between spatial features across time, are integral to understanding many of the tasks that humans engage in daily. Unfortunately, learning temporal features directly from video is a complex problem and contemporary [&hellip;]<\/p>\n","protected":false},"author":8,"featured_media":158,"parent":0,"menu_order":2,"comment_status":"closed","ping_status":"closed","template":"","meta":{"footnotes":""},"_links":{"self":[{"href":"https:\/\/carl.cs.unh.edu\/index.php\/wp-json\/wp\/v2\/pages\/78"}],"collection":[{"href":"https:\/\/carl.cs.unh.edu\/index.php\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/carl.cs.unh.edu\/index.php\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/carl.cs.unh.edu\/index.php\/wp-json\/wp\/v2\/users\/8"}],"replies":[{"embeddable":true,"href":"https:\/\/carl.cs.unh.edu\/index.php\/wp-json\/wp\/v2\/comments?post=78"}],"version-history":[{"count":34,"href":"https:\/\/carl.cs.unh.edu\/index.php\/wp-json\/wp\/v2\/pages\/78\/revisions"}],"predecessor-version":[{"id":410,"href":"https:\/\/carl.cs.unh.edu\/index.php\/wp-json\/wp\/v2\/pages\/78\/revisions\/410"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/carl.cs.unh.edu\/index.php\/wp-json\/wp\/v2\/media\/158"}],"wp:attachment":[{"href":"https:\/\/carl.cs.unh.edu\/index.php\/wp-json\/wp\/v2\/media?parent=78"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}