Following is a list of a few of the projects underway
with CPL. Most of the projects listed below have a lot of overlap
with each other.
of Captured Experiences project
is aimed at reducing substantially the human input necessary for creating
and accessing large collections of multimedia, particularly multimedia
created by capturing what is happening in an environment.
project is a an attempt at a building a large scale living laboratory that
is aware of its inhabitants and their activities.
is HCI research in the domain of ballet. While good dance instructors are
masters in their own right, we are collaborating with the Psychology department
to build and evaluate a computer system that could be used to teach dance.
The system will provide users with views of each dance motion from various
angles and at different speeds, and is being tested against video-instruction
tapes and books.
Capture your whole classroom experience, a room that takes notes.
A research project aimed at the recognition, modeling, and realistic animation
of facial expressions. Various methods for recognition of facial
expressions have been developed. At present our interest in creating
a database for testing our system and towards realistic animation of facial
The Bacchae is a PTRL production of a greek
tragedy performed at the Dramatech theater in March, 1998. As our first
step toward applying our research in the world of entertainment f/x, we
provided parts of a computer vision system, in the form of video-augmenting
software, to represent the world as seen from the eyes of a blind soothsayer.
a. e. Intelligent and Aware
Spaces) are spaces
that have been transformed into smart work areas where color CCD cameras,
big screen displays, microphones, and other sensors are fused with computers.
Real-time analysis and tracking of lab inhabitants takes place. An intelligent
agent interfaces with lab dwellers and other devices in the room.
A Robust method for tracking heads from video is developed. In this
method the head is tracked by finding the six translation and rotation
parameters to register a rendered image of the textured model with the
head in the video image.
is a vision-based methodology for detecting and recognizing physical interactions
between a person and objects in the surroundings. Human action recognition
as well as object classification is performed using an object-oriented
framework. The goal is to make computers aware of people and their activities.
This research has applications in automatic video annotation and surveillance
as well as in embedded environments.
(PErsonal PEt) project
is a long-term research project to build intelligent, adaptive, user-friendly
agent that displays different emotional states and awide range of behaviors.
More specifically, we are trying to build an agent that exhibits pet-like
behaviors and emotions. Our goal is to make the interaction between the
user and the agent as natural as possible thus making the user perceive
the agent more as a friend or a pet than as a toy.
We are creating a system that identifies and tracks people based on their
footfalls. We are instrumenting a space in order to track and identify
multiple people across a large indoor area. We are also integrating video
indetification and tracking technologies into the project. Applications
we are exploring include home activities, art and dance performance, and
and Audio: The aim of this project is to explore
and prototype a spectrum of applications for ubiquitous video and audio
processing in our daily environments. We are developing methodologies
to instrument spaces with video and audio sensors and studying modeling
techniques for interpretation and analysis of signals that are captured
by these sensors.