대구한의대학교 향산도서관

상세정보

부가기능

Structured Deep Visual Dynamics Models for Robot Manipulation

상세 프로파일

상세정보
자료유형학위논문
서명/저자사항Structured Deep Visual Dynamics Models for Robot Manipulation.
개인저자Byravan, Arunkumar.
단체저자명University of Washington. Computer Science and Engineering.
발행사항[S.l.]: University of Washington., 2019.
발행사항Ann Arbor: ProQuest Dissertations & Theses, 2019.
형태사항178 p.
기본자료 저록Dissertations Abstracts International 81-05B.
Dissertation Abstract International
ISBN9781687955777
학위논문주기Thesis (Ph.D.)--University of Washington, 2019.
일반주기 Source: Dissertations Abstracts International, Volume: 81-05, Section: B.
Advisor: Fox, Dieter.
이용제한사항This item must not be sold to any third party vendors.This item must not be added to any third party search indexes.
요약The emergence of deep learning, access to large amounts of data and powerful computing hardware have led to great strides in the state-of-the-art in robotics, computer vision, and AI. Unlike traditional methods that are strongly model-based with priors and explicit structural constraints, these newer learning approaches tend to be data-driven and often neglect the underlying problem structure. As a consequence, while they usually outperform their traditional counterparts on many problems, achieving good generalisation, interpretability, task transfer and data-efficiency has been challenging. Combining the strengths of the two paradigms, the flexibility of modern learning techniques, and the domain knowledge and structure of traditional methods should help bridge this gap.In this thesis, we will present work that combines these two paradigms, specifically in the context of learning visual dynamics models for robot manipulation tasks. This thesis is divided into two parts. In the first part, we discuss a structured approach to designing visual dynamics models for manipulation tasks. We propose a specific class of deep visual dynamics models (SE3-Nets) that explicitly encode strong physical and 3D geometric priors (specifically, rigid body physics) in their structure. As opposed to deep models that reason about motion a pixel level, SE3-Nets model the dynamics of observed scenes at the object level - they identify objects in the scene and predict rigid body rotation and translation per object. This leads to an interpretable architecture that can robustly model the dynamics of complex interactions. Next, we discuss SE3-Pose-Nets, an extension of SE3-Nets that additionally learns to estimate a latent, globally-consistent pose representation for objects and use the corresponding representation for real-time closed-loop visuomotor control of a Baxter robot. We show that the structure inherent in SE3-Pose-Nets allows them to be robust to visual perturbations and noise, generalizing to settings significantly different than seen during training. We also briefly discuss Dynamics-Nets, a recurrent extension to SE3-Pose-Nets that can be used for the control of dynamical systems.In the second part of the thesis, we present an approach towards solving long-horizon manipulation tasks, using reinforcement learning
일반주제명Robotics.
언어영어
바로가기URL : 이 자료의 원문은 한국교육학술정보원에서 제공합니다.

서평(리뷰)

  • 서평(리뷰)

태그

  • 태그

나의 태그

나의 태그 (0)

모든 이용자 태그

모든 이용자 태그 (0) 태그 목록형 보기 태그 구름형 보기
 
로그인폼