Matthew Turk

ORCID: 0000-0002-4198-8401
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Advanced Vision and Imaging
  • Advanced Image and Video Retrieval Techniques
  • Robotics and Sensor-Based Localization
  • Hand Gesture Recognition Systems
  • Human Pose and Action Recognition
  • Face and Expression Recognition
  • Face recognition and analysis
  • Computer Graphics and Visualization Techniques
  • Optical measurement and interference techniques
  • Gaze Tracking and Assistive Technology
  • Augmented Reality Applications
  • Interactive and Immersive Displays
  • Tactile and Sensory Interactions
  • Image Processing Techniques and Applications
  • Robot Manipulation and Learning
  • Image Retrieval and Classification Techniques
  • Virtual Reality Applications and Impacts
  • Video Surveillance and Tracking Methods
  • Visual Attention and Saliency Detection
  • Biometric Identification and Security
  • Video Analysis and Summarization
  • Handwritten Text Recognition Techniques
  • Image and Object Detection Techniques
  • Industrial Vision Systems and Defect Detection
  • 3D Surveying and Cultural Heritage

Toyota Technological Institute at Chicago
2020-2024

University of California, Santa Barbara
2010-2023

University of California System
2005-2016

Institute of Electrical and Electronics Engineers
2005

University of California, Berkeley
2004

Massachusetts Institute of Technology
1991-2003

University of Illinois Urbana-Champaign
1999-2002

Microsoft (United States)
2002

Human Media
1991

We have developed a near-real-time computer system that can locate and track subject's head, then recognize the person by comparing characteristics of face to those known individuals. The computational approach taken in this is motivated both physiology information theory, as well practical requirements performance accuracy. Our treats recognition problem an intrinsically two-dimensional (2-D) rather than requiring recovery three-dimensional geometry, taking advantage fact faces are normally...

10.1162/jocn.1991.3.1.71 article EN Journal of Cognitive Neuroscience 1991-01-01

An approach to the detection and identification of human faces is presented, a working, near-real-time face recognition system which tracks subject's head then recognizes person by comparing characteristics those known individuals described. This treats as two-dimensional problem, taking advantage fact that are normally upright thus may be described small set 2-D characteristic views. Face images projected onto feature space ('face space') best encodes variation among images. The defined...

10.1109/cvpr.1991.139758 article EN 2002-12-10

Vision-based hand gesture interfaces require fast and extremely robust detection. Here, we study view-specific posture detection with an object recognition method proposed by Viola Jones. Training this is computationally very expensive, prohibiting the evaluation of many appearances for their suitability to In paper, present a frequency analysis-based instantaneous estimation class separability, without need any training. We built detectors most promising candidates, receiver operating...

10.1109/afgr.2004.1301601 article EN 2004-06-10

Computer-mediated communication systems known as collaborative virtual environments (CVEs) allow geographically separated individuals to interact verbally and nonverbally in a shared space real time. We discuss CVE-based research paradigm that transforms (i.e., filters modifies) nonverbal behaviors during social interaction. Because the technology underlying CVEs allows strategic decoupling of rendered behavior from actual interactants, conceptual perceptual constraints inherent face-to-face...

10.1162/1054746041944803 article EN PRESENCE Virtual and Augmented Reality 2004-08-01

The performance of face recognition methods using subspace projection is directly related to the characteristics their basis images, especially in cases local distortion or partial occlusion. In order for a method be robust and occlusion, images generated by should exhibit part-based representation. We propose an effective representation named locally salient ICA (LS-ICA) that LS-ICA only employs information from important facial parts maximize benefit applying idea "recognition parts". It...

10.1109/tpami.2005.242 article EN IEEE Transactions on Pattern Analysis and Machine Intelligence 2005-11-01

We present a non-photorealistic rendering approach to capture and convey shape features of real-world scenes. use camera with multiple flashes that are strategically positioned cast shadows along depth discontinuities in the scene. The projective-geometric relationship camera-flash setup is then exploited detect distinguish them from intensity edges due material discontinuities.We introduce depiction methods utilize detected edge generate stylized static animated images. can highlight...

10.1145/1015706.1015779 article EN ACM Transactions on Graphics 2004-08-01

We propose a state-based approach to gesture learning and recognition. Using spatial clustering temporal alignment, each is defined be an ordered sequence of states in spatial-temporal space. The 2D image positions the centers head both hands user are used as features; these located by color-based tracking method. From training data given gesture, we first learn information then group into segments that automatically aligned temporally. further integrated build finite state machine (FSM)...

10.1109/afgr.2000.840667 article EN 2002-11-07

This paper introduces "Flocks of Features," a fast tracking method for non-rigid and highly articulated objects such as hands. It combines KLT features learned foreground color distribution to facilitate 2D position from monocular view. The tracker's benefits lie in its speed, robustness against background noise, ability track that undergo arbitrary rotations vast rapid deformations. We demonstrate tracker performance on hand with non-stationary camera unconstrained indoor outdoor...

10.1109/cvpr.2004.345 article EN 2005-04-01

We describe a framework and prototype implementation for unobtrusive mobile remote collaboration on tasks that involve the physical environment. Our system uses Augmented Reality paradigm model-free, markerless visual tracking to facilitate decoupled, live updated views of environment world-stabilized annotations while supporting moving camera unknown, unprepared environments. In order evaluate our concept prototype, we conducted user study with 48 participants in which expert instructed...

10.1145/2371574.2371610 article EN 2012-09-21

Immersive collaborative virtual environments (CVEs) are simulations in which geographically separated individuals interact a shared, three-dimensional, digital space using immersive environment technology. Unlike videoconference technology, transmits direct video streams, CVEs accurately track movements of interactants and render them nearly simultaneously (i.e., real time) onto avatars, three-dimensional representations the interactants. Nonverbal behaviors can be rendered veridically or...

10.1111/j.1468-2958.2005.tb00881.x article EN Human Communication Research 2005-10-01

Advances in imaging technology have made the capture and display of digital images ubiquitous. A variety displays are used to view them, ranging from high-resolution computer monitors low-resolution mobile devices, often undergo changes size aspect ratio adapt different screens. Also, displaying printing documents with embedded frequently entail resizing comply overall layout. Straightforward image operators, such as scaling, do not produce satisfactory results, since they oblivious content....

10.1117/12.862419 article EN Proceedings of SPIE, the International Society for Optical Engineering/Proceedings of SPIE 2010-08-19

We present a mobile augmented reality (AR) translation system, using smartphone's camera and touchscreen, that requires the user to simply tap on word of interest once in order produce translation, presented as an AR overlay. The seamlessly replaces original text live stream, matching background foreground colors estimated from source images. For this purpose, we developed efficient algorithm for accurately detecting location orientation stream is robust perspective distortion, combine it...

10.1109/wacv.2011.5711545 article EN 2011-01-01

Algorithms based on RANSAC that estimate models using feature correspondences between images can slow down tremendously when the percentage of correct (inliers) is small. In this paper, we present a probabilistic parametric model allows us to assign confidence values for each matching correspondence and therefore accelerates generation hypothesis under these conditions. Our framework leverages Extreme Value Theory accurately statistics scores produced by nearest-neighbor matcher. Using new...

10.1109/iccv.2013.307 article EN 2013-12-01

The motivations for organizing this special section were to better address the challenges of face recognition in real-world scenarios, promote systematic research and evaluation promising methods systems, provide a snapshot where we are domain, stimulate discussion about future directions. We solicited original contributions on all aspects recognition, including: design robust similarity features metrics; clustering sorting algorithms; novel user interaction models algorithms tagging;...

10.1109/tpami.2011.182 article EN IEEE Transactions on Pattern Analysis and Machine Intelligence 2011-08-25

The computational capability of mobile phones has been rapidly increasing, to the point where augmented reality become feasible on cell phones. We present an approach indoor localization and pose estimation in order support applications a phone platform. Using embedded camera, application localizes device familiar environment determines its orientation. Once 6 DOF is determined, 3D virtual objects from database can be projected into image displayed for user. Off-line data acquisition...

10.1109/cvprw.2010.5543249 article EN IEEE Computer Society Conference on Computer Vision and Pattern Recognition workshops 2010-06-01

In this paper we present a user study evaluating the benefits of geometrically correct user-perspective rendering using an Augmented Reality (AR) magic lens. simulation compared lens against common device-perspective on both phone-sized and tablet-sized displays. Our results indicate that display allows for significantly faster performance selection task has over task. Based these promising results, created proof-of-concept prototype, engineered with current off-the-shelf devices software....

10.1109/ismar.2012.6402557 article EN 2012-11-01

We propose a novel formulation of minimal case solutions for determining the relative pose perspective and generalized cameras given partially known rotation, namely, axis rotation. An rotation may be easily obtained by detecting vertical vanishing points with computer vision techniques, or aid sensor measurements from smart phone. Given our algorithms solve angle around along unknown translation. formulate these problems as Quadratic Eigen value Problems which are very simple to construct....

10.1109/3dv.2014.66 article EN 2014-12-01

Producing sports highlights is a labor-intensive work that requires some degree of specialization. We propose model capable automatically generating with focus on cricket. Cricket sport complex set rules and played for longer time than most other sports. In this paper we considers both event-based excitement-based features to recognize clip important events in cricket match. Replays, audio intensity, player celebration, playfield scenarios are examples cues used capture such events. To...

10.1109/cvprw.2018.00233 article EN 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) 2018-06-01

A simple imaging range sensor is described, based on the measurement of focal error, as described by A. Pentland (1982 and 1987). The current implementation can produce over a 1 m/sup 3/ workspace with measured standard error 2.5% (4.5 significant bits data). system implemented using relatively inexpensive commercial image-processing equipment. Experience shows that this ranging technique be both economical practical for tasks which require quick reliable but coarse estimates range. Examples...

10.1109/cvpr.1989.37858 article EN 2003-01-07

The research described in this paper analyzes the in-plane rotational robustness of Viola-Jones object detection method when used for hand appearance detection. We determine bounds training and achieving undiminished performance without an increase classifier complexity. result - up to 15/spl deg/ total differs from method's on faces (30/spl total). found that randomly rotating data within these allows rates about one order magnitude better than those trained strictly aligned data....

10.1109/icpr.2004.1334480 article EN Proceedings of the 17th International Conference on Pattern Recognition, 2004. ICPR 2004. 2004-01-01

We propose a novel approach for modeling, tracking and recognizing facial expressions. Our method works on low dimensional expression manifold, which is obtained by Isomap embedding. In this space, contour features are first clustered, using mixture model. Then, dynamics learned classification. use ICondensation to track in the embedded while expressions cooperative manner, within common probabilistic framework. The image observation likelihood derived from variation of Active Shape Model...

10.1109/cvpr.2004.390 article EN 2005-04-05
Coming Soon ...