Benchmarking Human Ability to Recognize Faces & People

Post on 29-Jul-2022

2 views 0 download

Transcript of Benchmarking Human Ability to Recognize Faces & People

1

Benchmarking Human Ability to

Recognize Faces & People

Dr. P. Jonathon Phillips

National Institute of Standards and Technology

Who is this person?

Is this same person?

Jenkins et al. (2011)

Unfamiliar Faces: How many identities here?

Key Papers

• P. J. Phillips and A. J. O’Toole, “Comparison of Human and

Computer Performance Across Face Recognition Experiments,”

Image and Vision Computing, 32, 74-85, 2014

• A. Rice, P. J. Phillips, V. Natu, X. An, and A. J. O’Toole,

“Unaware Person Recognition from the Body when Face

Identification Fails,” Psychological Science, 24 (11), 2235-2243,

2013

5

Two Dimensions of Recognition

Human Ability

Difficulty of Images

Low aptitude Super recognizer

Super matcher

Mugshots Digital Point &

Shoot Camera

• Human subject raters respond…

– 1. sure they are the same person

– 2. think they are the same person

– 3. not sure

– 4. think they are not the same person

– 5. sure they are not the same person

Measuring Human Performance

Area Under Curve (AUC)

8

The Good, Bad, & Ugly Face Challenge

• Three performance levels

– Good

– Bad

– Ugly

• Nikon D70-6 Mpixels (SLR)

• Indoor & outdoor images

• Frontal face images

• Taken within one year

Face Pairs

Good Challenging Very Challenging

Face Pairs

Very Challenging Challenging Good

Good, Bad, Ugly Performance

Frontal Still Face Performance

0.50

0.60

0.70

0.80

0.90

1.00

FRGCEasy

FRGCDifficult

FRVT2006ND

FRVT2006

Sandia

Ugly Bad Good

Human

Algorithm

13

Are

a U

nder

the R

OC

(A

UC

)

Is this same person?

Is this same person?

Is this same person?

Human Performance on Hard Face-Pairs

17

Face masked

Face only

Original image

Algorithm

Rated Use of Internal and External Features

More Use Less Use

Example of Point & Shoot Face Images

Courtesy PittPatt

Range of Performance

20

Mugshots

Digital SLR

Web

Photos Digital Point &

Shoot Cameras

0.997

0.8

0.54

0.22

0

0.2

0.4

0.6

0.8

1

MBE

2010

GBU

2011

LFW

2012

P&SC

2012

Ver

ific

ati

on

Ra

te a

t

FA

R =

0.0

01

Glasgow Face Matching Test

Burton, White & McNeill (2010). Behavior Research Methods, 42, 286-291.

Same or different?

Glasgow Face Matching Test

Burton, White & McNeill (2010). Behavior Research Methods, 42, 286-291.

Video: Walking vs. Conversation

• Human subject raters respond…

– 1. sure they are the same person

– 2. think they are the same person

– 3. not sure

– 4. think they are not the same person

– 5. sure they are not the same person

Gait Experiments

gait video

conversation video

body only face only

CG

CC

GG

Static Face

Human and Machine Performance

• For frontal, machine and human performance

related

• Algorithms Better (Untrained Humans)

– Mugshots & Mobile Studio environments

– Digital Single Lens Reflex

• Mobile Studio and Ambient Lighting

• Humans Better

– Non-face identity cues

– Cross-pose (video—one experiment)

• Not Measured

– Point and Shot Cameras

– Change in Pose (in general)

26

Questions?

Hurdle: Measuring Success

27

Alg

orith

m A

UC

Human AUC

• Develop structure for comparing human and machine performance

• Adapting recent methods from Neuroscience.

Hurdle: Measuring Success

28

Hurdle: Measuring Success

29

The Challenge

• Problem: Robust Recognition of Unfamiliar Faces

• Goal: Human Level Performance

– Untrained Humans

– Trained Professionals

– Forensic Examiners

• Compare Machine & Human on a Face Performance

Index

• Objective: Move Machine Performance into the Goal Box

• Human subject raters respond…

– 1. sure they are the same person

– 2. think they are the same person

– 3. not sure

– 4. think they are not the same person

– 5. sure they are not the same person

Video: Walking vs. Walking

Frontal Still Face Performance

0.50

0.60

0.70

0.80

0.90

1.00

FRGCEasy

FRGCDifficult

FRVT2006ND

FRVT2006

Sandia

Ugly Bad Good

Human

Human

33

Are

a U

nder

the R

OC

(R

OC

)

Human and Machine Performance

• Mugshots & Mobile Studio environments

– FRVT 2002/2006

– MBE 2010

• Mobile Studio vs Ambient Lighting

– FRGC

– FRVT 2006

• Ambient Lighting (indoor/outdoor)

– Good, Bad, & Ugly

• Hard Still Cases (reverse ROC)

• Video

Next Directions

• In hard cases (poor viewing conditions), humans

take advantage of face, body, still, & video

• Evidence: algorithms do NOT take advantage of

face, body, still, & video

• Learn from the human visual system.

– Functional

– Perceptual

• Incorporate into algorithm design.