: Paolo Remagnino, Gian L.uca Foresti, Tim Ellis
: Gian Luca Foresti, Tim Ellis
: Ambient Intelligence A Novel Paradigm
: Springer-Verlag
: 9780387229911
: 1
: CHF 85.40
:
: Betriebssysteme, Benutzeroberflächen
: English
: 240
: Wasserzeichen/DRM
: PC/MAC/eReader/Tablet
: PDF

Ambient Intelligence (AmI) is an integrating technology for supporting a pervasive and transparent infrastructure for implementing smart environments. Such technology is used to enable environments for detecting events and behaviors of people and for responding in a contextually relevant fashion. AmI proposes a multi-disciplinary approach for enhancing human machine interaction.

Am ient Intelligence: A Novel Paradigm is a compilation of edited chapters describing current state-of-the-art and new research techniques including those related to intelligent visual monitoring, face and speech recognition, innovative education methods, as well as smart and cognitive environments.

The authors start with a description of the iDorm as an example of a smart environment conforming to the AmI paradigm, and introduces computer vision as an important component of the system. Other computer vision examples describe visual monitoring for the elderly, classic and novel surveillance techniques using clusters of cameras installed in indoor and outdoor application domains, and the monitoring of public spaces. Face and speech recognition systems are also covered as well as enhanced LEGO blocks for novel educational purposes. The book closes with a provocative chapter on how a cybernetic system can be designed as the backbone of a human machine interaction.

Chapter 6

A DISTRIBUTED MULTIPLE CAMERA SURVEILLANCE SYSTEM
(p.107-108)

T.Ellis, J.Black, M.Xu and D.Makris
Digital Imaging Research Centre (DIRC), Kingston University, UK
{t.ellis,j.black,m.xu,d.makris}@kingston.ac.uk

1. Introduction

An important capability of an ambient intelligent environment is the capacity to detect, locate and identify objects of interest. In many cases interesting can move, and in order to provide meaningful interaction, capturing and tracking the motion creates a perceptively-enabled interface, capable of understanding and reacting to a wide range of actions and activities. CCTV systems fulfill an increasingly important role in the modern world, providing live video access to remote environments. Whilst the role of CCTV has been primarily focused on rather specific surveillance and monitoring tasks (i.e. security and traffic monitoring), the potential uses cover a much wider range.

The proliferation of video security surveillance systems over the past 5-10 years, in both public and commercial environments, is extensively used to remotely monitor activity in sensitive locations and publicly accessible spaces. In town and city centres, surveillance has been acknowledged to result in significant reductions in crime. However, in order to provide comprehensive and large area coverage of anything but the simplest environments, a large number of cameras must be employed.

In complex and cluttered environments with even moderate numbers of moving objects (e.g. 10-20) the problem of tracking individual objects is significantly complicated by occlusions in the scene, where an object may be partially occluded or totally disappear from camera view for both short or extended periods of time. Static occlusion results from objects moving behind (with respect to the camera) fixed elements in the scene (e.g. walls, bushes), whilst dynamic occlusion occurs as a result of moving objects in the scene occluding each other, where targets may merge or separate (e.g. a group of people walking together).

Information can be combined from multiple viewpoints to improve reliability, particularly taking advantage of the additional information where it minimises occlusion within the field-of-view (FOV). We treat the non-visible regions between camera views as simply another type of occlusion, and employ spatio-temporal reasoning to match targets moving between cameras that are spatially adjacent. The"boundaries" of the system represent locations from which previously unseen targets can enter the network.

To aid robust tracking across the camera network requires the system to maintain a record of each target entering the system and throughout its duration. When a target disappears from any camera FOV, motion prediction, colour identification, and learnt route patterns are used to re-establish tracking when the target reappears. Each target is maintained as a persistent object in the active database and spatial and temporal reasoning are used to detect these activities and ensure that entries are not retained for indefinite periods.

This chapter describes a multi-camera surveillance network that can detect and track objects (principally pedestrians and vehicles) moving through an outdoor environment. The remainder of this chapter is divided into four sections. The first describes the architecture of our multi-camera surveillance system. The second considers the image analysis methods for detecting and tracking objects within a single camera view. The next section deals with the integration of information from multiple cameras. The final section describes the structure of the database.
Contents7
Preface9
Foreword11
1 AMBIENT INTELLIGENCE14
1. Introduction14
2. The Essex approach15
2.1 The iDorm - A Testbed for Ubiquitous Computing and Ambient Intelligence15
2.2 The iDorm Embedded Computational Artifacts17
3. Integrating Computer Vision21
3.1 User Detection22
3.2 Estimating reliability of detection24
3.3 Vision in the iDorm26
4. Conclusions26
References26
2 TOWARDS AMBIENT INTELLIGENCE FOR THE DOMESTIC CARE OF THE ELDERLY28
1. Introduction28
2. An Integrated Supervision System29
2.1 E-service Based Integration Schemata32
3. People and Robot Localization and Tracking System34
3.1 System architecture and implementation36
4. The Plan Execution Monitoring System39
4.1 Representing Contingencies42
4.2 The Execution Monitor43
5. Integrating Sensing and Execution Monitoring: a Running Example46
6. Conclusions and Future Work49
References51
3 SCALING AMBIENT INTELLIGENCE52
1. Ambient Intelligence: the contribution of different disciplines52
2. I-BLOCKS technology55
3. Design process57
4. Scaling Ambient Intelligence at level of compositional devices: predefined activities58
4.1 Arithmetic training59
4.2 Storytelling Play Scenario60
4.3 Linguistic scenario63
5. Scaling Ambient Intelligence at level of compositional devices: free activities65
6. Scaling Ambient Intelligence at the level of configurable environments: future scenarios67
6.1 The Augmented Playground67
6.2 Self-reconfigurable Robots70
7. Discussion and conclusions71
References73
4 VIDEO AND RADIO ATTRIBUTES EXTRACTION FOR HETEROGENEOUS LOCATION ESTIMATION76
1. Introduction76
2. Related work77
3. Main tasks of Ambient Intelligence systems78
4. Architecture design79
4.1 Inspiration79
4.2 Mapping the Model into an AmI Architecture81
4.3 Artificial Sensing82
4.4 Proposed structure82
5. Context aware systems84
5.1 Location feature85
5.2 The formalism86
5.3 Alignment and Extraction of Video and Radio Object Reports88
6. Results93
6.1 The environment93
6.2 Results for video object extraction93
6.3 Results for radio object extraction93
6.4 Alignment results95
7. Conclusions95
8. Acknowledgments96
References96
5 DISTRIBUTED ACTIVE MULTICAMERA NETWORKS102
1. Introduction102
2. Sensing modalities102
3. Vision for Ambient Intelligence103
4. Architecture104
5. Tracking and object detection105
5.1 Object detection105
5.2 Tracking106
5.3 Appearance models107
5.4 Track data108
6. Normalization108
7. Multi-camera coordination110
8. Multi-scale image acquisition111
8.1 Active Head Tracking and Face Cataloging112
8.2 Uncalibrated, multiscale data acquisition114
8.3 Extensions115
9. Indexing Surveillance Data115
9.1 Visualization116
10. Privacy116
11. Conclusions117
References117
6 A DISTRIBUTED MULTIPLE CAMERA SURVEILLANCE SYSTEM120
1. Introduction120
2. System architecture121
3. Motion detection and single-view tracking121
3.1 Motion Detection122
3.2 Scene Models124
3.3 Target Tracking125
3.4 Partial Observation126
3.5 Target Reasoning129
4. Multi view tracking133
4.1 Homography Estimation133
4.2 Least Median of Squares134
4.3 Feature Matching Between Overlapping Views135
4.4 3D Measurements136
4.5 Tracking in 3D137
4.6 Non-Overlapping Views139
5. System architecture142
5.1 Surveillance Database143
6. Summary145
7. Appendix147
7.1 Kalman Filter147
7.2 Homography Estimation148
7.3 3D Measurement Estimation149
References150
7 LEARNING AND INTEGRATING INFORMATION FROM MULTIPLE CAMERA VIEWS152
1. Introduction152
1.1 Semantic Scene Model154
2. Learning point-based regions156
3. Learning traj