Virtual Vision:
A Simulation Framework for Camera Sensor Networks Research Demetri Terzopoulos UCLA
2007 PhD Thesis of Faisal Qureshi University of Toronto Publications: • • • • •
2008 Proceedings of the IEEE 2008 11th Communications and Networking Simulation Symposium (CNS) 2008 ACM Symposium on Virtual Reality Software and Technology (VRST) 2007 First IEEE/ACM Intl. Conf. on Distributed Smart Cameras (ICDCS) 2007 IEEE/ACM Intl. Conf. on Distributed Computing in Sensor Systems
• • • • •
2007 IEEE Conf. on Computer Vision and Pattern Recognition (CVPR) 2006 ACM Intl. Wksp. on Distributed Smart Cameras (DSC) 2006 ACM Multimedia Systems Journal 2005 ACM Wksp. on Video Surveillance and Sensor Networks (VSSN) 2005 IEEE Intl. Wksp. on Visual Surveillance (VS-PETS)
(DCOSS)
1
Camera Networks and Smart Cameras •
Visual surveillance is becoming ubiquitous – London has roughly 4,000,000 cameras
•
Effective visual coverage of large spaces require multi-camera systems
•
Operator monitoring is infeasible for large networks
•
Need networks of smart cameras capable of autonomous operation
•
Smart cameras = visual sensor nodes – Local on-board processing – Communication with neighbors
Difficulty of Doing Large-Scale Visual Sensor Networks Research • Deploying large-scale camera networks in extensive public spaces for research purposes: – Hardware related technical challenges – Privacy and legal issues – Prohibitive cost
• Infeasible for most computer vision researchers
2
Virtual Camera Sensor Networks
Virtual Vision
Visually and behaviorally realistic simulators for designing and evaluating machine vision systems
Camera Model: Virtual Camera
Environment Models: geometry, texture, illumination
Pedestrian Models:
appearance, movement, behavior Reality Emulator Virtual Penn Station
pan, tilt, zoom, camera jitter, color response, lens distortions, etc.
High-Level Control camera control, assignment, handover, etc.
Virtual Video
Machine Vision tracking, recognition, etc.
3
Virtual Vision for Camera Networks Research •
Emulates the characteristics of a physical vision system
•
Flexibility during system design and evolution
•
Readily available ground truth
•
Offline operation and testing
•
No legal impediments
•
No special hardware
•
Repeatability
•
Low cost
QuickTime™ and a decompressor are needed to see this picture.
Environmental Model of Original Penn Station in NYC [Shao & Terzopoulos, 2005]
Platforms
Concourses
Main MainWaiting WaitingRoom Room
Concourses Arcade Arcade
Platforms
4
Architecture of the Autonomous Pedestrian Model Environment & Interaction Cognition Behavior Motion Geometry
Pedestrian Simulation
5
Pedestrian Simulation
Human Activity in the Station
6
Virtual Vision • Testbed for multi-camera sensor networks
Computer Vision Emulation Using Synthetic Video • Pedestrian detection – Background subtraction using a learned background model
QuickTime™ and a decompressor are needed to see this picture.
7
Low-Level Computer Vision Emulation Using Synthetic Video •
Pedestrian tracker – Appearance based
Low-Level Computer Vision Emulation Using Synthetic Video • Appearance based pedestrian tracker – [Swain & Ballard 91]
– Zoom invariant
Video frame
Signature Backprojected image: image High intensity suggests presence of target
8
Anatomy of a Camera Node Decision Decision logic logic
Communication Communication subsystem: subsystem: message message passing passing to to neighboring neighboring nodes nodes
Tracking
Task Task “relevance” “relevance” computation computation framework framework
region of interest
failure
reacquire
Vision Vision routines: routines: pedestrian pedestrian tracking tracking
Lost
Free
state management timeout
y
Searching
timeout
left right zoom
z x
Image Image driven driven reactive reactive behaviors: behaviors: fixation fixation and and zooming zooming PD PD controllers controllers
up down
active pan-tilt-zoom camera
Virtual Vision: A Tool for Visual Sensor Network Research High-Level Camera Control
Visual Sensing
Visual Sensing
Synthetic Video
Real Video
Virtual Camera Network
Physical Camera Network
Synthetic World (Reality Emulator)
Real World
Virtual Vision
Real Vision
9
Camera Sensor Network #1 Active PTZ Camera Scheduling
Scheduling Active PTZ Cameras • Number of Pedestrians
>
Number of active cameras
• Task active cameras to observe pedestrians in the scene
Active PTZ camera
Passive wide-FOV camera
10
Scheduling Active PTZ Cameras • Scheduling problem
– Given n PTZ cameras, and m pedestrians, persistently observe every pedestrian using one PTZ camera at a time
• Goal – Observe as many pedestrians for as long as possible
Virtual Active Camera Scheduling Setup • Passive wide-FOV cameras – Calibrated – Pedestrian localization through triangulation
• Active PTZ cameras – Un-calibrated – Learn a coarse mapping between 3D locations and internal pan-tilt settings
• Reliable pedestrian identification in different cameras via appearance based signatures
Calibrated passive cameras at the four corners of the waiting room in the virtual train station
11
Distinguishing Features of Our Active Camera Scheduling Strategy •
Previous attempts –
Costello et al., 2004 Costello et al., 2006 Lim et al., 2006 Huang and Trivedi, 2003
•
Exit times are not known a priori
•
Recording duration is not known a priori
•
Allow multiple observations of the same pedestrian
•
Multiple pan/tilt/zoom cameras
•
Tracking failures due to occlusions
Active Camera Scheduling Strategy • Camera assignment via weighted round robin • First come, first served tie breaking • Multiple observation • Preemption
Close-up snapshots captured by active PTZ cameras
12
Active Camera Scheduling Results Preemption (P) No preemption (NP) Single observation (SO) Multiple observations (MO) Single-class model (SC) Multi-class model (MC)
Multiple observations, multi-class, preemption scheduler outperforms other variants
Up to 4 Cameras; 10, 20 Pedestrians
Camera Sensor Network #2 Active PTZ Camera Assignment and Grouping
13
Smart Camera Network •
Cameras self-organize to observe pedestrians during their presence in the region – Smart, PTZ active cameras – Un-calibrated – Ad hoc deployment – Camera additions & removal – Handle camera failures – Deal with imperfect communication
•
Previous work
– Simmons et al., 2006 Mallet, 2006 Park et al., 2006 Kulkarni et al., 2005 Javed et al., 2000; Devarajan and Radke, 2006
Vision and Goal • Ad hoc deployment
14
Vision and Goal • Cameras work towards common sensing goals
Task assignment
Vision and Goal • Cameras work towards common sensing goals
Group formation
15
Vision and Goal • Cameras work towards common sensing goals
Group evolution
Vision and Goal • Cameras work towards multiple common sensing goals
Group 1
Group 2
16
Vision and Goal • Cameras work towards multiple common sensing goals
Camera failure
Vision and Goal • Cameras work towards multiple common sensing goals
Task assignment
17
Vision and Goal • Cameras work towards multiple common sensing goals
Conflict resolution
Camera Grouping and Reassignment
18
Active PTZ Camera Assignment and Grouping • Camera selection, grouping, and handoff via an auction model – Announcement/Bidding/Selection – ContractNet • Smith, 1983
• Conflict resolution within a Constraint Satisfaction Problem framework – Partially distributed
• A Camera can only perform a single task at any given time
Camera Grouping: Announcement •
Start with a single camera that is tasked to observe a person
Leader
19
Camera Grouping: Announcement •
Seeks out other cameras in the vicinity to form a group to help it with the observation task
Leader
Camera Grouping: Bidding •
One or more cameras that receive the task announcement respond with their relevance values
Never Received the Message
Leader Response: high relevance
Response: low relevance
Did not respond
20
Camera Grouping: Bidding •
One or more cameras that receive the task announcement respond with their relevance values
•
Relevance encodes how successful a camera will be at an observation task Never Received the Message
Leader Response: high relevance
Response: low relevance
Did not respond
Camera Grouping: Selection •
After the leader gets relevance messages from neighboring cameras, it selects suitable cameras to join the group
Join
Group
Leader
21
Conflict Detection •
A conflict is detected when multiple tasks require the same camera node to proceed successfully
Conflict Detection •
A Red group member receives a recruit query from Green group
22
Conflict Detection •
A Red group member receives a recruit query from Green group
Conflict Detection •
A Red group member receives a recruit query from Green group
Conflict!
23
Conflict Detection •
Nodes belonging to both groups send information to one of the leaders
•
Leader selection Centralization
Conflict Detection •
The resulting node (camera) assignment is sent to the individual nodes
Solution
24
Conflict Detection •
The resulting node (camera) assignment is sent to the individual nodes
Persistent Observation: Virtual Active PTZ Camera Assignment and Grouping
25
Network Model •
It does not require camera calibration or network calibration – Can take advantage of calibration information, if available – Ad hoc deployment
•
Camera grouping is a strictly local negotiation – Typically camera groups are spatially local arrangements
•
Camera groups are dynamic arrangements
•
Camera handoffs occur naturally during negotiations
•
It can gracefully handle node and message failures
Network Model • Camera nodes can be added or removed during the lifetime of a task • Even assuming perfect sensing, the proposed model can still fail if – a significant number of messages are lost – catastrophic node failure – group evolution can’t keep up with a fast-changing observation task
• Scalability
– Small group sizes – Conflict resolution is viable as long as the number of relevant sensors for each task remains low ( < 10 ) – Optimal sensor assignment
26
Future Work • Camera sensor network
– Consistent labeling of pedestrians – Who, What, Where, When? – Cognitive modeling
• Physical sensor networks
– Environment monitoring, urban sensing, intelligent environments, …
• Bigger, better reality emulators – High-fidelity synthetic video – An entire city …
Virtual LA Urban Simulation Lab Architecture Dept UCLA
27
Virtual LA • Building level of detail
Virtual LA • Interior spaces
28
Strategy Looking Forward • Collaborative engagement of the CV and CG communities – Open source framework to develop and test state-of-the-art vision algorithms and systems – Modular www-accessible environment • e.g., SRI’s Open Agent Architecture (OAA)
– “Any-world” emulator – More realism (“The Matrix”)
Thank You !
29