Académique Documents
Professionnel Documents
Culture Documents
BIR BHANU
bhanu@cris.ucr.edu
http://vislab.ucr.edu
www.vislab.ucr.edu
SURVEILLANCE
- Dictionary Meaning -
www.vislab.ucr.edu
Video Surveillance
Sensors, Objects, Actions/Activities
- overlapped/non-overlapped
- at a distance, local, wide area
- distributed/centralized
- implementation
- real-time/forensic
- applications
Reference
B. Bhanu, C. Ravishankar, A. Roy Chowdhury, H. Aghajan and D. Terzopoulos (Eds.),
“Distributed Video Sensor Networks,” ISBN 978-0-85729-126-4, 503 pages, Springer,
2011.
www.vislab.ucr.edu
Video Surveillance
• Video cameras
• Other imaging/non-imaging sensors
• Fusion of sensors
• Algorithms
• TOPICS/Example Applications
• Humans, Re-identification
• At a distance - Gait, side face, ear
• Newborn babies pain
• Driver emotions
• Activities (sports)
• Vehicles
www.vislab.ucr.edu
Things to Come (1)
• Overview and sample applications
www.vislab.ucr.edu
VideoWeb Laboratory
• An Outdoor/Indoor Laboratory for Research & Education
• A wireless video network of 80 Pan, Tilt and Zoom (PTZ)
cameras for large scale experiments. All entry/exit
locations have cameras. (EURISP 2010)
Fixed
Camera 1
Fixed
Camera 2
Vislab
UCR
Wall of Video Streams
www.vislab.ucr.edu
Architecture and Servers
• Three levels of computation
- Low-level: Detect/track moving objects
– Mid-level: Feature extraction,
multi-camera tracking
– High-level: face/gait and activity
recognition, scene analysis,
Biometrics – soft/hard
• Requirements:
– Process over 98 MB/s of video data
in real-time
• Solution:
– Rack of 32 custom-built servers
• Intel Core 2 Quad Q6600 processors,
expandable to 8 cores/server
• 2GB of Memory, expandable to 24GB
• 250 GB of disk space, expandable to 4TB
– Interface server for user control of entire network
www.vislab.ucr.edu
Optimizing Network
• Measure the performance of a wireless camera network.
Depending on the task, there are numerous “optimal” ways to configure
– Biometrics - Maximize resolution and video quality!
– Surveillance recording - Maximize reliability!
– Tracking - Maximize frame rate and response!
• Configure the network to optimize its performance for specific
tasks. “Configuring the network” may consist of changing camera
settings, bandwidth controls, network infrastructure, etc.
• Performance metrics to optimize:
– Maximize frame rate, Minimize standard deviation of frame rate
– Minimize video compression, Maximize video resolution
– Minimize maximum lag time
• Managing the tradeoff between parameters:
– Higher resolution may lead to lower frame rates.
– Lower compression may lead to longer lag.
– Higher number of users reduces throughput.
www.vislab.ucr.edu
Pareto Set %s for All Cameras
Compression \
Resolution 100 60 30 20 0
176×120 46% 66% 46% 51% 74%
352×240 34% 46% 26% 34% 91%
704×240 51% 29% 17% 54% 97%
704×480 34% 31% 63% 94% 100%
704x480, 352x240,
20 compression 0 compression
www.vislab.ucr.edu
Data Collection: 8 Cameras
• Sample Actions:
– Running, Sit on bench, Wave off (Ignore), Observe from
afar (more than 10 feet), Sit cross legged, Walking with
crutches, Walking close (within One foot)
– CameraClient also capable of recording raw or processed
data. (Denina, et al. DVSN 2011)
–
www.vislab.ucr.edu
Intersection: 7 Camera Views
• Sample Actions:
– People: Getting out of a vehicle, Walking, Close door, Stand
by the vehicle, Carry object
– Cars: Right turn, Stop at intersection
www.vislab.ucr.edu
Camera Selection, Hand-off and Control
• How to track multiple persons using
multiple cameras?
• Which camera is the ‘best’ for each
person? Is the ‘best’ camera unique?
• How to get a stable solution?
• No detail calibration is required.
• Some Related Work
• Need topology information (Not
flexible for active cameras)
• Master-slave scheme (require
overlapping FOVs)
• No “best camera” is selected
www.vislab.ucr.edu
Re-Identification
Suspect Suspect
Camera
1 2
A Recognizing
individuals in
non-
Same
person?
overlapping
Camera cameras at
B different time
and locations
Different Changing
Occlusion
Pose Illumination
Plus low resolution, image
noise, blurriness …
Different People
Look Similar
Same People
Look Different
• Normal (Class 1)
• Break-in (Class 2)
• Stay (Class 3)
ous
suspici
suspicious
• Sudden appearance
uncertain and disappearance
Sudden appea
rance/d isappearance (Class 4)
26
www.vislab.ucr.edu
Pedestrian Tracking with Crowd Simulation
Image 1 Yes/No
•y
•x
a0
a /6
a /3
www.vislab.ucr.edu
USF HumanID Database
• Version 1.7
•Noon
•Late
Afternoon
•Night
www.vislab.ucr.edu
Gait Template Construction
•Slow
Walking
•Fast
Walking
•Run
www.vislab.ucr.edu
Performance Prediction
• Question
– How many people can be recognized by gait?
• Required information
– Feature distribution over population
– Uncertainty in feature estimation (estimation error distribution)
• Technical approach
– Define feature distribution space based on anthropometric data
– Construct feature uncertainty model in ideal situation
– Predict human recognition performance on static gait features
•Real
•Feature feature
• . •• . . distribution
Space •Feature
• .
• . uncertainty
• . • . • .
www.vislab.ucr.edu
Fusion of Color and Infrared
• Color Video • Infrared Video (Long-wave)
• Motivation
– Unreliably extracted body parts from one sensor might be reliably
extracted from the other sensor
– Moving object in a scene provides correspondence for automatic image
registration between synchronized color and infrared videos
• Contributions
– A hierarchical evolutionary computation approach for automatic image
registration
– Fusion of color and infrared videos to improve human detection
performance PR 2007
www.vislab.ucr.edu
Automatic Image Registration Result
•Original
Color Images
•Registered
Color Images
•Original
Infrared Images
www.vislab.ucr.edu
Sensor Fusion Results
• Original • Registered • Original • Detected • Detected • Sensor
Color Color Infrared Images Color Infrared Fusion
Images Images Silhouettes Silhouettes Results
www.vislab.ucr.edu
3D Gait - Body Data
• Human body measurement system captures 4 representative
poses of walking humans
angle
arbitrary pose by interpolation of time
joint angles and part lengths
Gait
Pose 1 Pose 2 Pose 3 Pose 4
Reconstruction
• Gait cycle is composed of 20 frames
• There are major differences among human gaits
• Gallery Ears
• Probe ear
www.vislab.ucr.edu
Real Videos
35x35 80x80
www.vislab.ucr.edu
Integrating Face and Gait
•t
60
Zapping Index based on Smile
Correlation of Facial Expression With Race Track
CorkScrew Turn
8-8A
Emotion:
Tense
Pain of Neonates
• It is estimated that 15 million premature babies
are born every year worldwide. 1 million of these
babies die each year due to complications in
preterm birth.
• These neonates are non verbal and cannot
communicate their feelings.
• Although a number of pain instruments have
been developed to assist health professionals,
these tools are subjective and skewed and may
under/over estimate the pain response.
• This could further lead to misdiagnosis and
under/over treatment which has a life long
impact on the society.
Vehicle Recognition (2017)
– The MIO-TCD dataset consists of 521,451 images with 11 different
classes namely:
• articulated truck (1.98%), background (30.68%), bus
(1.98%), bicycle (0.44%), car (49.96%), motorcycle
(0.38%), non-motorized vehicle (0.34%), pedestrian
(1.2%), pickup truck (9.76%), single unit truck (0.98%)
and work van (1.86%).
– We added extra images for the 8 classes that have less than 5% of
total images.
External Dataset Total number of images
added
Imagenet 2,247
MIO-TCD Localization dataset 101,234
Pedestrian Re-identification dataset (PETA) 18,000
– All the images in our dataset were resized to maintain aspect ratio
such that the shorter side has length of 256 pixels.
Vehicle Recognition
www.vislab.ucr.edu
Detection of mTBI
• Mild Traumatic Brain Injury (mTBI) occurs
in 1.3 million cases per year in US
• There is a wide range of causes: vehicle
accidents, assault, sports, falls, etc.
• The current understanding of many of the
mechanisms and direct outcomes of mTBI
are unknown. TBI caused by a Car
accident.
• This is due to the lack of quantitative
measures in the analysis of mTBI. mTBI is on the rise in the military
0.02
Blue - Normal
Red - Lesion
0.01
0.01
0.00
Low contrast nature of mTBI. Left: T2W MRI of the rat
226
251
101
126
151
176
201
276
301
326
351
376
1
26
51
76
brain 1 day after injury. Right: manually segmented T2
lesion (white) and brain (green and blue.) Lesion cannot be separated
based only on T2 value.
Papers: IEEE TMI 2014, IEEE TBME 2014, MEDIA 2014
Human Embroynic Stem Cells
• Promising in the treatment of many diseases
and for toxicological testing.
• Determining the number of various types of
cells automatically in a population of mixed
morphologies - Need good detection,
classification and tracking technique to
understand the behavior of cells over time.
• Modeling of dynamic blebbing phenomenon.
• Capacity to differentiate into diverse human
cell types (process of differentiation)
Segmentation of Stem Cells
Sample Video: General Challenges
1. Low SNR (Signal to Noise Ratio)
of phase contrast images.
2. Poor NDBSU-hESC recognition
when neighboring cells are
undergoing chemical reaction.
(NDBSU-hESC is shown in Figure (a).)
Our contribution is to reduce the
effects of the above problems for
NDBSU-hESC detection. We are
using multiple classifiers to improve
General Cell Types in the Videos: the detection accuracy.
Stem cell
colony
detection
and
analysis
IEEE/ACM TCBB 2014, Best Paper IEEE Conf Healthcare Informatics, Imaging, and Systems
Biology Conference, Sept. 27-28, 2012.
Pollen Tube Growth
For the May 2016 issue, see p. 1929 for Table of Contents.
For the June 2016 issue, see p. 2423 for Table of Contents.
Cover art from: “Segmentation of Pollen Tube Growth Videos Using Dynamic Bi-Modal Fusion and Seam Carving”
by Asongu L. Tambo and Bir Bhanu (p. 1993–2004, Fig. 1).
www.vislab.ucr.edu
Biological Significance of This Research