Preface |
List of Contributors |
Intelligent Environments: Methods, Algorithms and Applications / Dorothy N. Monekosso ; Paolo Remagnino ; Yoshinori Kuno1: |
Intelligent Environments / 1.1: |
What Is An Intelligent Environment? / 1.1.1: |
How Is An Intelligent Environment Built? / 1.1.2: |
Technology for Intelligent Environments / 1.2: |
Research Projects / 1.3: |
Private Spaces / 1.3.1: |
Public Spaces / 1.3.2: |
Middleware / 1.3.3: |
Chapter Themes in This Collection / 1.4: |
Conclusion / 1.5: |
References |
A Pervasive Sensor System for Evidence-Based Nursing Care Support / Toshio Hori ; Yoshifumi Nishida ; Shin'ichi Murakami2: |
Introduction / 2.1: |
Evidence-Based Nursing Care Support / 2.2: |
Background of the Project / 2.2.1: |
Concept of Evidence-Based Nursing Care Support / 2.2.2: |
Initial Goal of the Project: Falls Prevention / 2.2.3: |
Second Goal of the Project: Obtaining ADL of Inhabitants / 2.2.4: |
Related Work / 2.3: |
Overview and Implementations of the System / 2.4: |
Overview of the Evidence-Based Nursing Care Support System / 2.4.1: |
System Implementations / 2.4.2: |
Experiments and Analyses / 2.5: |
Tracking a Wheelchair for Falls Prevention / 2.5.1: |
Activity Transition Diagram: Transition of Activities in One Day / 2.5.2: |
Quantitative Evaluation of Daily Activities / 2.5.3: |
Probability of "Toilet" Activity / 2.5.4: |
Discussion of the Experimental Results / 2.5.5: |
Prospect of the Evidence-Based Nursing Care Support System / 2.6: |
Conclusions / 2.7: |
Anomalous Behavior Detection: Supporting Independent Living / 3: |
Related work / 3.1: |
Methodology / 3.3: |
Unsupervised Classification Techniques / 3.3.1: |
Using HMM to Model Behavior / 3.3.2: |
Experimental Setup and Data Collection / 3.4: |
Noisy Data: Sources of Error / 3.4.1: |
Learning activities / 3.4.2: |
Experimental Results / 3.5: |
Instance Class Annotation / 3.5.1: |
Data Preprocessing / 3.5.2: |
Models: Unsupervised Classification: Clustering and Allocation of Activities to Clusters / 3.5.3: |
Behaviors: Discovering Patterns in Activities / 3.5.4: |
Behaviors: Discovering Anomalous Patterns of Activity / 3.5.5: |
Discussion / 3.6: |
Sequential Pattern Mining for Cooking-Support Robot / Yasushi Nakauchi3.7: |
System Design / 4.1: |
Inference from Series of Human Actions / 4.2.1: |
Time Sequence Data Mining / 4.2.2: |
Human Behavior Inference Algorithm / 4.2.3: |
Activity Support of Human / 4.2.4: |
Implementation / 4.3: |
IC Tag System / 4.3.1: |
Inference of Human's Next Action / 4.3.2: |
Cooking Support Interface / 4.3.3: |
Robotic, Sensory and Problem-Solving Ingredients for the Future Home / Amedeo Cesta ; Luca Iocchi ; G. Riccardo Leone ; Daniele Nardi ; Federico Pecora ; Riccardo Rasconi4.4: |
Components of the Multiagent System / 5.1: |
The Robotic Platform Mobility Subsystem / 5.2: |
The Interaction Manager / 5.3: |
Environmental Sensors for People Tracking and Posture Recognition / 5.4: |
Monitoring Activities of Daily Living / 5.5: |
Schedule Representation and Execution Monitoring / 5.5.1: |
Constraint Management in the RoboCare Context / 5.5.2: |
From Constraint Violations to Verbal Interaction / 5.5.3: |
Multiagent Coordination Infrastructure / 5.6: |
Casting the MAC Problem to DCOP / 5.6.1: |
Cooperatively Solving the MAC Problem / 5.6.2: |
Ubiquitous Stereo Vision for Human Sensing / Ikushi Yoda ; Katsuhiko Sakae5.7: |
Ubiquitous Stereo Vision / 6.1: |
Concept of Ubiquitous Stereo Vision / 6.2.1: |
Server-Client Model for USV / 6.2.2: |
Real Utilization Cases / 6.2.3: |
Hierarchical Utilization of 3D Data and Personal Recognition / 6.3: |
Acquisition of 3D Range Information / 6.3.1: |
Projection to Floor Plane / 6.3.2: |
Recognition of Multiple Persons and Interface / 6.4: |
Pose Recognition for Multiple People / 6.4.1: |
Personal Identification / 6.4.2: |
Interface for Space Control / 6.4.3: |
Human Monitoring in Open Space (Safety Management Application) / 6.5: |
Monitoring Railroad Crossing / 6.5.1: |
Station Platform Edge Safety Management / 6.5.2: |
Monitoring Huge Space / 6.5.3: |
Conclusion and Future Work / 6.6: |
Augmenting Professional Training, an Ambient Intelligence Approach / B. Zhan ; D.N. Monekosso ; S. Rush ; P. Remagnino ; S.A. Velastin7: |
Color Tracking of People / 7.1: |
Counting People by Spatial Relationship Analysis / 7.3: |
Simple People Counting Algorithm / 7.3.1: |
Graphs of Blobs / 7.3.2: |
Estimation of Distance Between Blobs / 7.3.3: |
Temporal Pyramid for Distance Estimation / 7.3.4: |
Probabilistic Estimation of Groupings / 7.3.5: |
Grouping Blobs / 7.3.6: |
Stereo Omnidirectional System (SOS) and Its Applications / Yutaka Satoh ; Katsuhiko Sakaue7.4: |
System Configuration / 8.1: |
Image integration / 8.3: |
Generation of Stable Images at Arbitrary Rotation / 8.4: |
An example Application: Intelligent Electric Wheelchair / 8.5: |
Overview / 8.5.1: |
Obstacle Detection / 8.5.2: |
Gesture / Posture Detection / 8.5.4: |
Video Analysis for Ambient Intelligence in Urban Environments / Andrea Prati ; Rita Cucchiara8.6: |
Visual Data for Urban AmI / 9.1: |
Video Surveillance in Urban Environment / 9.2.1: |
The LAICA Project / 9.2.2: |
Automatic Video Processing for People Tracking / 9.3: |
People Detection and Tracking from Single Static Camera / 9.3.1: |
People Detection and Tracking from Distributed Cameras / 9.3.2: |
People Detection and Tracking from Moving Cameras / 9.3.3: |
Privacy and Ethical Issues / 9.4: |
From Monomodal to Multimodal: Affect Recognition Using Visual Modalities / Hatice Gunes ; Massimo Piccardi10: |
Organization of the Chapter / 10.1: |
From Monomodal to Multimodal: Changes and Challenges / 10.3: |
Background Research / 10.3.1: |
Data Collection / 10.3.2: |
Data Annotation / 10.3.3: |
Synchrony/Asynchrony Between Modalities / 10.3.4: |
Data Integration/Fusion / 10.3.5: |
Information Complementarity/Redundancy / 10.3.6: |
Information Content of Modalities / 10.3.7: |
Monomodal Systems Recognizing Affective Face or Body Movement / 10.4: |
Multimodal Systems Recognizing Affect from Face and Body Movement / 10.5: |
Project 1: Multimodal Affect Analysis for Future Cars / 10.5.1: |
Project 2: Emotion Analysis in Man-Machine Interaction Systems / 10.5.2: |
Project 3: Multimodal Affect Recognition in Learning Environments / 10.5.3: |
Project 4: FABO-Fusing Face and Body Gestures for Bimodal Emotion Recognition / 10.5.4: |
Multimodal Affect Systems: The Future / 10.6: |
Importance of Vision in Human-Robot Communication: Understanding Speech Using Robot Vision and Demonstrating Proper Actions to Human Vision / Michie Kawashima ; Keiichi Yamazaki ; Akiko Yamazaki11: |
Understanding Simplified Utterances Using Robot Vision / 11.1: |
Inexplicit Utterances / 11.2.1: |
Information Obtained by Vision / 11.2.2: |
Language Processing / 11.2.3: |
Vision Processing / 11.2.4: |
Synchronization Between Speech and Vision / 11.2.5: |
Experiments / 11.2.6: |
Communicative Head Gestures for Museum Guide Robots / 11.3: |
Observations from Guide-Visitor Interaction / 11.3.1: |
Prototype Museum Guide Robot / 11.3.2: |
Experiments at a Museum / 11.3.3: |
Index / 11.4: |
Preface |
List of Contributors |
Intelligent Environments: Methods, Algorithms and Applications / Dorothy N. Monekosso ; Paolo Remagnino ; Yoshinori Kuno1: |