2017
|
Shipman, F; Duggina, S; Monteiro, C; Gutierrez-Osuna, R Speed-Accuracy Tradeoffs for Detecting Sign Language Content in Video Sharing Sites Proceedings Article In: Proceedings of ACM SIGACCESS Conference on Computers and Accessibility (ASSETS 2017), pp. 185-189, 2017. @inproceedings{shipman2017assets,
title = {Speed-Accuracy Tradeoffs for Detecting Sign Language Content in Video Sharing Sites},
author = {F Shipman and S Duggina and C Monteiro and R Gutierrez-Osuna},
url = {https://psi.engr.tamu.edu/wp-content/uploads/2018/01/shipman2017assets.pdf},
year = {2017},
date = {2017-11-21},
booktitle = {Proceedings of ACM SIGACCESS Conference on Computers and Accessibility (ASSETS 2017)},
pages = {185-189},
keywords = {Computer vision, Gestures, Speech},
pubstate = {published},
tppubtype = {inproceedings}
}
|
2016
|
Monteiro, C; Mathew, C; Shipman, F; Gutierrez-Osuna, R Detecting and Identifying Sign Languages through Visual Features Proceedings Article In: 2016 IEEE International Symposium on Multimedia (ISM), 2016. @inproceedings{monteiro2016ism,
title = {Detecting and Identifying Sign Languages through Visual Features},
author = {C Monteiro and C Mathew and F Shipman and R Gutierrez-Osuna},
url = {https://psi.engr.tamu.edu/wp-content/uploads/2018/01/monteiro2016ism.pdf},
year = {2016},
date = {2016-12-11},
booktitle = {2016 IEEE International Symposium on Multimedia (ISM)},
keywords = {Computer vision, Gestures, Pattern recognition},
pubstate = {published},
tppubtype = {inproceedings}
}
|
2015
|
Shipman, F; Gutierrez-Osuna, R; Shipman, T; Monteiro, C; Karappa, V Towards a Distributed Digital Library for Sign Language Content Proceedings Article In: Proc. ACM/IEEE Joint Conference on Digital Libraries, 2015. @inproceedings{shipman2015jcdl,
title = {Towards a Distributed Digital Library for Sign Language Content},
author = {F Shipman and R Gutierrez-Osuna and T Shipman and C Monteiro and V Karappa},
url = {https://psi.engr.tamu.edu/wp-content/uploads/2018/01/shipman2015jcdl.pdf},
year = {2015},
date = {2015-04-02},
booktitle = {Proc. ACM/IEEE Joint Conference on Digital Libraries},
volume = {in press},
keywords = {Computer vision, Gestures, Pattern recognition},
pubstate = {published},
tppubtype = {inproceedings}
}
|
2014
|
Karappa, V; Monteiro, C; Shipman, F; Gutierrez-Osuna, R Detection of sign-language content in video through polar motion profiles Proceedings Article In: Proc. 39th International Conference on Acoustics, Speech, and Signal Processing (ICASSP), pp. 1299-1303, 2014. @inproceedings{virendraasl2014icassp,
title = {Detection of sign-language content in video through polar motion profiles},
author = {V Karappa and C Monteiro and F Shipman and R Gutierrez-Osuna},
url = {https://psi.engr.tamu.edu/wp-content/uploads/2018/01/virendraasl2014icassp.pdf},
year = {2014},
date = {2014-05-09},
booktitle = {Proc. 39th International Conference on Acoustics, Speech, and Signal Processing (ICASSP)},
pages = {1299-1303},
keywords = {Computer vision, Gestures},
pubstate = {published},
tppubtype = {inproceedings}
}
|
2013
|
Shipman, F; Gutierrez-Osuna, R; Monteiro, C Identifying Sign Language Videos in Video Sharing Sites Journal Article In: ACM Transactions on Accessible Computing, vol. in press, 2013. @article{Shipman2013,
title = {Identifying Sign Language Videos in Video Sharing Sites},
author = {F Shipman and R Gutierrez-Osuna and C Monteiro},
url = {https://psi.engr.tamu.edu/wp-content/uploads/2018/01/Shipman2013.pdf},
year = {2013},
date = {2013-11-04},
journal = {ACM Transactions on Accessible Computing},
volume = {in press},
keywords = {Computer vision, Gestures},
pubstate = {published},
tppubtype = {article}
}
|
2012
|
Monteiro, C; Gutierrez-Osuna, R; Shipman, F Design and Evaluation of Classifier for Identifying Sign Language Videos in Video Sharing Sites Conference 13th International ACM SIGACCESS Conference on Computers and Accessibility (ASSETS 2012)., 2012. @conference{monteiro2012assets,
title = {Design and Evaluation of Classifier for Identifying Sign Language Videos in Video Sharing Sites},
author = {C Monteiro and R Gutierrez-Osuna and F Shipman},
url = {https://psi.engr.tamu.edu/wp-content/uploads/2018/01/monteiro2012assets.pdf},
year = {2012},
date = {2012-10-22},
booktitle = {13th International ACM SIGACCESS Conference on Computers and Accessibility (ASSETS 2012).},
abstract = {Video sharing sites provide an opportunity for the collection and use of sign language presentations about a wide range of topics. Currently, locating sign language videos (SL videos) in such sharing sites relies on the existence and accuracy of tags, titles or other metadata indicating the content is in sign language. In this paper, we describe the design and evaluation of a classifier for distinguishing between sign language videos and other videos. A test collection of SL videos and videos likely to be incorrectly recognized as SL videos (likely false positives) was created for evaluating alternative classifiers. Five video features thought to be potentially valuable for this task were developed based on common video analysis techniques. A comparison of the relative value of the five video features shows that a measure of the symmetry of movement relative to the face is the best feature for distinguishing sign language videos. Overall, an SVM classifier provided with all five features achieves 82% precision and 90% recall when tested on the challenging test collection. The performance would be considerably higher when applied to the more varied collections of large video sharing sites.},
keywords = {Computer vision, Gestures},
pubstate = {published},
tppubtype = {conference}
}
Video sharing sites provide an opportunity for the collection and use of sign language presentations about a wide range of topics. Currently, locating sign language videos (SL videos) in such sharing sites relies on the existence and accuracy of tags, titles or other metadata indicating the content is in sign language. In this paper, we describe the design and evaluation of a classifier for distinguishing between sign language videos and other videos. A test collection of SL videos and videos likely to be incorrectly recognized as SL videos (likely false positives) was created for evaluating alternative classifiers. Five video features thought to be potentially valuable for this task were developed based on common video analysis techniques. A comparison of the relative value of the five video features shows that a measure of the symmetry of movement relative to the face is the best feature for distinguishing sign language videos. Overall, an SVM classifier provided with all five features achieves 82% precision and 90% recall when tested on the challenging test collection. The performance would be considerably higher when applied to the more varied collections of large video sharing sites. |
Lucchese, G; Field, M; Ho, J; Gutierrez-Osuna, R; Hammond, T GestureCommander: continuous touch-based gesture prediction Conference Proceedings of the 2012 ACM annual conference extended abstracts on Human Factors in Computing Systems Extended Abstracts, ACM 2012. @conference{lucchese2012chi,
title = {GestureCommander: continuous touch-based gesture prediction},
author = {G Lucchese and M Field and J Ho and R Gutierrez-Osuna and T Hammond},
url = {https://psi.engr.tamu.edu/wp-content/uploads/2018/01/lucchese2012chi.pdf},
year = {2012},
date = {2012-01-01},
booktitle = {Proceedings of the 2012 ACM annual conference extended abstracts on Human Factors in Computing Systems Extended Abstracts},
pages = {1925--1930},
organization = {ACM},
abstract = {GestureCommander is a touch-based gesture control system for mobile devices that is able to recognize gestures as they are being performed. Continuous recognition allows the system to provide visual feedback to the user and to anticipate user commands to possibly decrease perceived response time. To achieve this goal we employ two Hidden Markov Model (HMM) systems, one for recognition and another for generating visual feedback. We analyze a set of geometric features used in other gesture recognition systems and determine a subset that works best for HMMs. Finally we demonstrate the practicality of our recognition HMMs in a proof of concept mobile application for Google's Android mobile platform that has a recognition accuracy rate of 96% over 15 distinct gestures.},
keywords = {Gestures, Mobile computing},
pubstate = {published},
tppubtype = {conference}
}
GestureCommander is a touch-based gesture control system for mobile devices that is able to recognize gestures as they are being performed. Continuous recognition allows the system to provide visual feedback to the user and to anticipate user commands to possibly decrease perceived response time. To achieve this goal we employ two Hidden Markov Model (HMM) systems, one for recognition and another for generating visual feedback. We analyze a set of geometric features used in other gesture recognition systems and determine a subset that works best for HMMs. Finally we demonstrate the practicality of our recognition HMMs in a proof of concept mobile application for Google's Android mobile platform that has a recognition accuracy rate of 96% over 15 distinct gestures. |
2011
|
Boulos, M N K; Blanchard, B J; Walker, C; Montero, J; Tripathy, A; Gutierrez-Osuna, R Web GIS in practice X: a Microsoft Kinect natural user interface for Google Earth navigation Journal Article In: International Journal of Health Geographics, vol. 10, no. 1, pp. 45, 2011. @article{boulos2011web,
title = {Web GIS in practice X: a Microsoft Kinect natural user interface for Google Earth navigation},
author = {M N K Boulos and B J Blanchard and C Walker and J Montero and A Tripathy and R Gutierrez-Osuna},
url = {https://psi.engr.tamu.edu/wp-content/uploads/2018/01/boulos2011web.pdf},
year = {2011},
date = {2011-01-01},
journal = {International Journal of Health Geographics},
volume = {10},
number = {1},
pages = {45},
publisher = {BioMed Central Ltd},
abstract = {This paper covers the use of depth sensors such as Microsoft Kinect and ASUS Xtion to provide a natural user interface (NUI) for controlling 3-D (three-dimensional) virtual globes such as Google Earth (including its Street View mode), Bing Maps 3D, and NASA World Wind. The paper introduces the Microsoft Kinect device, briefly describing how it works (the underlying technology by PrimeSense), as well as its market uptake and application potential beyond its original intended purpose as a home entertainment and video game controller. The different software drivers available for connecting the Kinect device to a PC (Personal Computer) are also covered, and their comparative pros and cons briefly discussed. We survey a number of approaches and application examples for controlling 3-D virtual globes using the Kinect sensor, then describe Kinoogle, a Kinect interface for natural interaction with Google Earth, developed by students at Texas A&M University. Readers interested in trying out the application on their own hardware can download a Zip archive (included with the manuscript as additional files 1, 2, & 3) that contains a 'Kinnogle installation package for Windows PCs'. Finally, we discuss some usability aspects of Kinoogle and similar NUIs for controlling 3-D virtual globes (including possible future improvements), and propose a number of unique, practical 'use scenarios' where such NUIs could prove useful in navigating a 3-D virtual globe, compared to conventional mouse/3-D mouse and keyboard-based interfaces.},
keywords = {Gestures},
pubstate = {published},
tppubtype = {article}
}
This paper covers the use of depth sensors such as Microsoft Kinect and ASUS Xtion to provide a natural user interface (NUI) for controlling 3-D (three-dimensional) virtual globes such as Google Earth (including its Street View mode), Bing Maps 3D, and NASA World Wind. The paper introduces the Microsoft Kinect device, briefly describing how it works (the underlying technology by PrimeSense), as well as its market uptake and application potential beyond its original intended purpose as a home entertainment and video game controller. The different software drivers available for connecting the Kinect device to a PC (Personal Computer) are also covered, and their comparative pros and cons briefly discussed. We survey a number of approaches and application examples for controlling 3-D virtual globes using the Kinect sensor, then describe Kinoogle, a Kinect interface for natural interaction with Google Earth, developed by students at Texas A&M University. Readers interested in trying out the application on their own hardware can download a Zip archive (included with the manuscript as additional files 1, 2, & 3) that contains a 'Kinnogle installation package for Windows PCs'. Finally, we discuss some usability aspects of Kinoogle and similar NUIs for controlling 3-D virtual globes (including possible future improvements), and propose a number of unique, practical 'use scenarios' where such NUIs could prove useful in navigating a 3-D virtual globe, compared to conventional mouse/3-D mouse and keyboard-based interfaces. |
2008
|
Paulson, B; Rajan, P; Davalos, P; Gutierrez-Osuna, R; Hammond, T What!?! no Rubine features?: using geometric-based features to produce normalized confidence values for sketch recognition Conference HCC Workshop: Sketch Tools for Diagramming, 2008. @conference{paulson2008no,
title = {What!?! no Rubine features?: using geometric-based features to produce normalized confidence values for sketch recognition},
author = {B Paulson and P Rajan and P Davalos and R Gutierrez-Osuna and T Hammond},
url = {https://psi.engr.tamu.edu/wp-content/uploads/2018/01/paulson2008no.pdf},
year = {2008},
date = {2008-01-01},
booktitle = {HCC Workshop: Sketch Tools for Diagramming},
pages = {57--63},
abstract = {As pen-based interfaces become more popular in today’s applications, the need for algorithms to accurately recognize hand-drawn sketches and shapes has increased. In many cases, complex shapes can be constructed hierarchically as a combination of smaller primitive shapes meeting certain geometric constraints.However, in order to construct higher level shapes, it is imperative to accurately recognize the lower-level primitives. Two approaches have become widespread in the sketch recognition field for recognizing lower-level primitives: gesture-based recognition and geometric-based recognition. Our goal is to use a hybrid approach that combines features from both traditional gesture based recognition systems and geometric-based recognition systems. In this paper, we show that we can produce a system with high recognition rates while providing the added benefit of being able to produce normalized confidence values for alternative interpretations;something most geometric-based recognizers lack. Moresignificantly, results from feature subset selection indicate that geometric features aid the recognition process more than gesture-based features when given naturally sketched data.},
keywords = {Gestures},
pubstate = {published},
tppubtype = {conference}
}
As pen-based interfaces become more popular in today’s applications, the need for algorithms to accurately recognize hand-drawn sketches and shapes has increased. In many cases, complex shapes can be constructed hierarchically as a combination of smaller primitive shapes meeting certain geometric constraints.However, in order to construct higher level shapes, it is imperative to accurately recognize the lower-level primitives. Two approaches have become widespread in the sketch recognition field for recognizing lower-level primitives: gesture-based recognition and geometric-based recognition. Our goal is to use a hybrid approach that combines features from both traditional gesture based recognition systems and geometric-based recognition systems. In this paper, we show that we can produce a system with high recognition rates while providing the added benefit of being able to produce normalized confidence values for alternative interpretations;something most geometric-based recognizers lack. Moresignificantly, results from feature subset selection indicate that geometric features aid the recognition process more than gesture-based features when given naturally sketched data. |