.comment-link {margin-left:.6em;}

Tech Scribes

This is a forum for posting articles discussing your Master's or PhD research. The aim of this website is to help you get intelligent feedback and help you view the problem from a different perspective. If you wish to post on the site, please e-mail caveman(at)gmail(dot)com or mailsunildsouza(at)yahoo(dot)com with the article which you want to post.

Monday, August 23, 2004

Gesture Recognition

Humans have always wanted to interact with their environment. Humans communicate with other people as well as objects daily to get work done. Computers and machines have become important elements in our society. We use them to perform various tasks. Humans still rely on keyboards, mouse or buttons to interact with machines. This has tremendously affected the naturalness of interaction. Humans interact with each other using speech and gestures and it would be great if we can do the same with machines. Recently a lot of research has been done for using speech as input for systems. Interaction using only speech has some drawbacks. The speech recognition system has to be trained by the user and even then they are not accurate enough. Also designing speech input based systems that can be operated by just one user don’t make sense.

Gesture recognition has been used to complement speech recognition. Gesture recognition can be done by interpreting the motion of a human arm. E.g. a hello wave. Gesture recognition was initially done using glove based techniques. These gloves had mechanical sensors to measure the arm and finger joint angles. The user has to wear a mechanical glove and carry the load of wires attached to it. This method of gesture recognition defeats the purpose (naturalness of interaction) itself. Researchers have now tried gesture recognition using visual sensors. Presently common human hand gestures like pointing, wave, and stop as well as American Sign Language gestures can be recognized using vision. Further research is needed to understand more complex human gestures and use both hand tracking for gesture recognition. I am presently working on a gesture recognition system.

PROPOSAL:

We plan to develop a gesture recognition system which permits a human user wearing a specially marked glove to command a computer system to carry out predefined gesture action commands. Here they basically get the initial and the final position of the points of interest (markers) on the glove and extract their motion vector, since different gestures have different sets of motion vectors for the markers they can be discriminated. After solving the static gesture recognition problem we plan to form a library of simple dynamic human gestures which can be recognized and be used for intelligent control of machines. This would require tracking the human hand. Such gestures would consist of discrete abrupt motions and a multi modal particle filter would be needed to track it using a library of motion models. Bandwidth limitation is a major concern in video tele-conferencing. Our system can be used to for video coding: image sequence are described by states (position, scale and orientation) of all objects in the scene and only change in model parameters are sent instead of the whole sequence of images.

8 Comments:

At 7:57 PM, Blogger BD said...

Intel OpenCV libraries have some Gesture recognition utilities. are you using them?

 
At 6:32 AM, Blogger Sunil said...

I am not using OpenCV. Till now its simple blob tracking. Eventually though I would be using OpenCV.

 
At 3:06 PM, Anonymous Anonymous said...

Speech and gestures are a great way for humans to interact with other humans. It is not clear at all that it is the best way for humans to interact with technology. For example, violinists interact quite expressively and subtly with their hands directly touching the instrument. It's hard to imagine playing an instrument by talking to it and gesturing (conductors not withstanding). Why do you believe that mouses and keyboards are not efficient ways to interact with computers?

 
At 4:46 PM, Blogger The Tobacconist said...

I agree with anon man on the need to justify the need for gesture based recognition when the mouse and keyboard are just fine. I think u are targetting the wrong market. A generalized tool that can be applied using simple hardware in a public environment will be of greater significance. To give u an example, Speech Tech has been trying big time to get a Large Vocab System that can recognize any relevant sequence of phonemes u can think of. But now everyone has realized that the potential to actually apply Speech Tech is in the public domain of Interactive Telephone Systems that provide an genuine alternative to the keyboard and mouse. eg. Railway/Airline Inquiry/Booking.
I am guessing u work primarily on HCI. And that would explain ur interest in using Gestures to "simplify" life. I personally would be uncomfy wearing a glove when I work on my comp. I think where u can really innovate is by thinking of better applications than on Personal Computers.
This just an opinion.

 
At 7:22 PM, Blogger Sunil said...

Thanks for your comments guys. I have nothing against using keyboard and a mouse for working on a computer. This work is aimed more at interacting with robots. Robots I believe in 5-7 years would be commonplace. Interacting effectively with robots for humans will be very essential. I personally would not like to hook my robot to my workstation and give it a set of commands. A robot that is equipped with gesture/speech recognition would be very handy. Gesture recognition with gloves is just a start.

This group (Computational Vision and Active Perception Laboratory at KTH, Stockholm, Sweden) http://www.nada.kth.se/cvap/gvmdi/

is involved in interesting gesture recognition research.

 
At 12:01 PM, Blogger The Tobacconist said...

Ok now I see the point. That would be cool wouldn't it. Anyway what is the approach u guys are using. Is it rule based? Neural Nets?

 
At 12:28 PM, Blogger Sunil said...

I am presently just doing blob tracking to obtain the motion vector and the comparing them to the library of gestures with similar motion vectors.

 
At 11:31 PM, Anonymous Anonymous said...

the most important differences between inter-person communication and person-computer communciation lie in speech and gesture. we are more comfortable communicating with another human being as a result of these two traits. equipping a computer with the ability to recognize speech and gesture is I think the right step forward.

 

Post a Comment

<< Home