forInternational Conference on Distributed Smart Cameras (ICDSC)

Call for Participation


2nd ACM/IEEE International Conference on
Distributed Smart Cameras (ICDSC-08)

September 7-11, 2008
Stanford University, California, USA

Plenary Speakers

Alex (Sandy) Pentland, MIT

Sensible Organizations: how distributed sensor data is allowing organizations to reinvent themselves

Speaker's Biography
Professor Alex ("Sandy") Pentland is a pioneer in organizational engineering, mobile information systems, and perceptual computing. Sandy's focus is the development of human-centered technology, and the creation of ventures that take this technology into the real world. His work provides people with a clearer picture of their social environment, and helps companies and communities to reinvent themselves to be both more human and productive.

He directs the Digital Life Consortium within the MIT Media Laboratory, a group of more than twenty multinational corporations exploring new ways to innovate, and is Founder of MIT's Legatum Center for Development and Entrepreneurship, established to support aspiring entrepreneurs in emerging markets. In 1997, Newsweek magazine named him one of the 100 Americans likely to shape this century.

Link to speaker's webpage

David Forsyth, UIUC

Looking at People

There is a great need for programs that can describe what people are doing from video. This is difficult to do, because it is hard to identify and track people in video sequences, because we have no canonical vocabulary for describing what people are doing, and because phenomena such as aspect and individual variation greatly affect the appearance of what people are doing. Recent work in kinematic tracking has produced methods that can report the kinematic configuration of the body fairly accurately and fully automatically.

The problem of vocabulary is more difficult. I will discuss a generative activity model that allows activities to be assembled from a set of distinct spatial and temporal components. The models themselves are learned from labelled motion capture data and are assembled in a way that makes it possible to learn very complex finite automata without estimating large numbers of parameters. The advantage of such a model is that one can search videos for examples of activities specified with a simple query language, without possessing any example of the activity sought. In this case, aspect is dealt with by explicit 3D reasoning.

An alternative strategy for dealing with aspect and individual variation is to build discriminative methods applied to appearance features. The difficulty here is that activities look different when seen from different directions. I will describe recent methods that make it possible to transfer models --- that is, to learn a model of an activity from one view, then recognize it in a completely different view.

Speaker's Biography
David Forsyth holds a BSc and an MSc in Electrical Engineering from the University of the Witwatersrand, Johannesburg, and an MA and D.Phil from Oxford University. He is currently a full professor at U. Illinois Urbana-Champaign, having served 10 years on the faculty at UC Berkeley. He has published over 100 papers on computer vision, computer graphics and machine learning. He served as program co-chair for IEEE Computer Vision and Pattern Recognition in 2000, general co-chair for CVPR 2006, program co-chair for ECCV 2008, and is a regular member of the program committee of all major international conferences on computer vision. He has received best paper awards at the International Conference on Computer Vision and at the European Conference on Computer Vision, and an IEEE Technical Achievement award.

Link to speaker's webpage

Yiannis Aloimonos, Univ. of Maryland

Languages of Human Activity

Speaker's Biography
Yiannis Aloimonos (PhD 1987, Univ. of Rochester) is a Professor of Computational Vision and Intelligence in the Dept. of Computer Science at the University of Maryland, College Park and the Director of the Computer Vision Laboratory at the Institute for Advanced Computer Studies. He is also affiliated with the Cognitive Science Program. He is known for his work on Active Vision and his study of vision as a dynamic process. He has contributed to the theory of Computational Vision in various ways, including the discovery of the trilinear constraints (with M. Spetsakis), and the mathematics of stability in motion analysis as a function of the field of view (with C. Fermuller), which led to the development of omni-directional sensors. He has received several awards for his work (including the Marr Prize for his work on Active Vision, the Presidential Young Investigator Award from President Bush (1990) and the Bodossaki Prize in Artificial Intelligence). He has coauthored four books, including Active Perception and Visual Navigation. He is interested in cognitive systems, specifically the integration of visual cues and the integration of vision, action and language.

Link to speaker's webpage

J. K. Aggarwal, Univ. of Texas at Austin

Computer Recognition of Human Activities and Objects

Computer Vision has graduated from a research tool in early 1960s to a mature discipline today. The developments in cameras, computers and memory have contributed in part to this maturing of computer vision. Namely, there is an explosive growth in the number of cameras in public places, the speed of computers has increased significantly and the price of memory has spectacularly decreased. The word camera may be used in a very broad sense since the imaging modalities vary from the usual cameras imaging a visual intensity image to thermal image, x-ray image and laser range image. The applications cover a broad spectrum of topics from recognizing faces and activities of persons, discovering abandoned baggage, inspecting baggage, imaging internal organs and fusing images from multiple modalities. The present talk will focus on the recognition of human activities and objects.

At The University of Texas at Austin, we are pursuing a number of projects on human motion understanding. Professor Aggarwal will present his research on modeling and recognition of actions and interactions, and human and object interactions. We consider atomic actions, composite actions and interactions, and continued and recursive activities. In addition, we consider the interactions between a person and an object. The object may be a piece of luggage, a car or an unmovable object like a fence. The applications considered include monitoring of: human activities in public places, abandoned baggage, parking of cars on roads and in parking lots. The issues considered in these problems will illustrate the richness of ideas involved and the difficulties associated with understanding human activities. Application of the above research to monitoring and surveillance will be discussed together with actual examples and their solutions. The use of multiple cameras possibly equipped with PTZ capabilities will be discussed.

Speaker's Biography
J.K. Aggarwal has served on the faculty of The University of Texas at Austin College of Engineering in the Department of Electrical and Computer Engineering since 1964. He is currently one of the Cullen Professors of Electrical and Computer Engineering. Professor Aggarwal earned his B.Sc. from University of Bombay, India in 1957, B. Eng. from University of Liverpool, Liverpool, England, 1960, M.S. and Ph.D. from University of Illinois, Urbana, Illinois, in 1961 and 1964 respectively. His research interests include image processing, computer vision and pattern recognition. The current focus of research is on the automatic recognition of human activity and interactions in video sequences, and on the use of perceptual grouping for the automatic recognition and retrieval of images and videos from databases.

A fellow of IEEE (1976) and IAPR (1998), Professor Aggarwal received the Best Paper Award of the Pattern Recognition Society in 1975, the Senior Research Award of the American Society of Engineering Education in 1992 and the IEEE Computer Society Technical Achievement Award in 1996. He is the recipient of the 2004 K. S. Fu Prize of the IAPR and the 2005 Leon K. Kirchmayer Graduate Teaching Award of the IEEE. He is the author or editor of 7 books and 52 book chapters, author of over 200 journal papers, as well as numerous proceeding papers and technical reports. He has served as the Chairman of the IEEE Computer Society Technical Committee on Pattern Analysis and Machine Intelligence (1987-1989), Director of the NATO Advanced Research Workshop on Multisensor Fusion for Computer Vision, Grenoble, France (1989), Chairman of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (1993), and the President of the International Association for Pattern Recognition (1992-1994). He is a life fellow of IEEE and Golden Core Member of IEEE Computer Society.

Link to speaker's webpage