UNZUETA IRURTIA, Luis (2009) Markerless full-body human motion capture and combined motor action recognition for human-computer interaction. PhD thesis, UNA.
|PDF (Markerless full-body human motion capture and combined motor action recognition for human-computer interaction) - Published Version |
Download (6Mb) | Preview
Typically, people interact with computers by means of devices such as the keyboard and the mouse. Computers can detect the events coming from these devices, such as pushing down or releasing isolated or combined keyboard and mouse buttons, or the mouse motions, and then react according to the interpretation assigned to them. This communication procedure has been satisfactorily used for a wide range of applications. However, this approach lacks naturalness with respect to face-to-face human communication.
This thesis project presents a method for markerless real-time capture and automatic interpretation of full-body human movements for human-computer interaction (HCI).
Three stages can be distinguished in order to reach this objective: (1) the markerless tracking of as many of the user’s body parts as possible, (2) the reconstruction of the kinematical 3D skeleton that represents the user’s pose from the tracked body parts, and (3) the recognition of the movement patterns in order to make the computer “understand” the user’s will and then react according to it. These three processes must be solved in real time in order to attain a satisfactory HCI.
The first stage can be solved by means of cameras focusing on the user and computer vision algorithms that extract and track the user’s relevant characteristics from the images. This project proposes a method that combines color probabilities and optical flow. The second one requires to situate the kinematical 3D skeleton in plausible biomechanical poses fitted to the detected body parts, considering previous poses in order to obtain smooth motions. This project proposes an analytic-iterative inverse kinematics method that situates the body parts sequentially, from the torso to the upper and lower limbs, taking into account the biomechanical limits and the most relevant collisions. Finally, the last stage requires analyzing which are the significant features of motion in order to interpret patterns with artificial intelligence techniques. This project proposes a method to automatically extract potential gestures from the data flow and then label them, allowing the performance of combined actions.
|Item Type:||Thesis (PhD)|
|Subjects:||Technological Sciences > Computer technology|
|Divisions:||UN > E.T.S. de Ing. Industriales y de Telecomunicación > Ingeniería Mecánica|
|Date Deposited:||24 May 2010 16:59|
|Last Modified:||21 Mar 2011 16:22|
- ASCII Citation
- Dublin Core
- EP3 XML
- Eprints Application Profile
- HTML Citation
- OAI-ORE Resource Map (Atom Format)
- OAI-ORE Resource Map (RDF Format)
- OpenURL ContextObject
- OpenURL ContextObject in Span
- Reference Manager
- Simple Metadata
Personal del repositorio solamente: página de control del documento
¹ Data hau katalogazio datuei dagokie, ez dokumentuari berari.