Top Banner
CMPSCI 670: Computer Vision Human-centric computer vision University of Massachusetts, Amherst November 24, 2014 Instructor: Subhransu Maji Many slides are based on Branson et al., ECCV 2010
61

CMPSCI 670: Computer Vision

Mar 19, 2022

Download

Documents

dariahiddleston
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: CMPSCI 670: Computer Vision

CMPSCI 670: Computer Vision!Human-centric computer vision

University of Massachusetts, Amherst November 24, 2014

Instructor: Subhransu Maji

Many slides are based on Branson et al., ECCV 2010

Page 2: CMPSCI 670: Computer Vision

• Project presentations (next Monday/Wednesday) • Presentations assignments at random • Each person (or team) will get 7 (or 10) mins to present

- Problem statement, preliminary results, data analysis, todo

• Final report due on Dec. 13 (hard deadline)

• Course evaluations • A show of hands who might be missing next class?

Administrivia

2

Page 3: CMPSCI 670: Computer Vision

Early vision:!• image formation • light and color perception • basic image processing

- edges, corners and blobs

Mid-level vision:!• texture

- synthesis and representation

• grouping - segmentation

- alignment

Overview of the course so far

3

High-level vision:!• recognition and learning • image representation

- features, etc • object detection

Misc. topics:!• deep learning • memorability [Khosla] • human-centric vision • optical flow and tracking

Page 4: CMPSCI 670: Computer Vision

• Motivation • Levels of categorization • Visual 20q game [Branson et al., ECCV 2010] • Similarity comparisons based recognition

• global similarity [Wah et al., CVPR 2014] • localized similarity [Wah et al., WACV 2015]

Overview

4

Page 5: CMPSCI 670: Computer Vision

What  type  of  bird  is  this?

5

Page 6: CMPSCI 670: Computer Vision

What  type  of  bird  is  this?

6…?

Field  Guide

Page 7: CMPSCI 670: Computer Vision

What  type  of  bird  is  this?

7

Computer  Vision

?

Page 8: CMPSCI 670: Computer Vision

What  type  of  bird  is  this?

8

Bird?

Computer  Vision

Page 9: CMPSCI 670: Computer Vision

What  type  of  bird  is  this?

9

Chair?  Bottle?

Computer  Vision

Page 10: CMPSCI 670: Computer Vision

Parakeet  Auklet

10

• Field  guides  difficult  for  average  users  

• Computer  vision  doesn’t  work  perfectly  (yet)  

• Research  mostly  on  basic-­‐level  categories

Page 11: CMPSCI 670: Computer Vision

Visual Recognition With Humans in the Loop

Parakeet  Auklet

What  kind  of  bird  is  this?

11

Page 12: CMPSCI 670: Computer Vision

Levels of Categorization

Airplane?  Chair?    Bottle?  …

Basic-­‐Level  Categories

12

[Griffin  et  al.  ‘07,  Lazebnik  et  al.  ‘06,  Grauman  et  al.  ‘06,  Everingham  et  al.  ‘06,  Felzenzwalb  et  al.  ‘08,  Viola  et  al.  ‘01,  …        ]

Page 13: CMPSCI 670: Computer Vision

Levels of Categorization

American  Goldfinch?    Indigo  Bunting?  …

Subordinate  Categories

13[Belhumeur  et  al.  ‘08  ,  Nilsback  et  al.  ’08,  …]

Page 14: CMPSCI 670: Computer Vision

Levels of Categorization

Yellow  Belly?    Blue  Belly?…

Parts  and  Attributes

14

[Farhadi  et  al.  ‘09,  Lampert  et  al.  ’09,  Kumar  et  al.  ‘09]

Page 15: CMPSCI 670: Computer Vision

Visual 20 Questions Game

Blue  Belly?  no

Cone-­‐shaped  Beak?  yes

Striped  Wing?  yes

American  Goldfinch?  yes

Hard  classification  problems  can  be  turned  into  a  sequence  of  easy  ones

15

Page 16: CMPSCI 670: Computer Vision

• Computers: reduce number of required questions • Humans: drive up accuracy of vision algorithms

Computer  Vision

Cone-­‐shaped  Beak?  yes

American  Goldfinch?  yes

Computer  Vision

Recognition With Humans in the Loop

16

Page 17: CMPSCI 670: Computer Vision

Research Agenda

Heavy  Reliance  on  Human  Assistance

More  Automated

Computer  Vision  

Improves

Blue  belly?  no  Cone-­‐shaped  beak?  yes  Striped  Wing?  yes  American  Goldfinch?  yes  

Striped  Wing?  yes  American  Goldfinch?  yes  

Fully  AutomaticAmerican  Goldfinch?  yes  

2014 2020

2025

17

Page 18: CMPSCI 670: Computer Vision

Field Guides

www.whatbird.com18

Page 19: CMPSCI 670: Computer Vision

Field Guides

www.whatbird.com19

Page 20: CMPSCI 670: Computer Vision

Example Questions

20

Page 21: CMPSCI 670: Computer Vision

Example Questions

21

Page 22: CMPSCI 670: Computer Vision

Example Questions

22

Page 23: CMPSCI 670: Computer Vision

Example Questions

23

Page 24: CMPSCI 670: Computer Vision

Example Questions

24

Page 25: CMPSCI 670: Computer Vision

Example Questions

25

Page 26: CMPSCI 670: Computer Vision

Basic Algorithm

Input  Image  (      )xQuestion  1:

Is  the  belly  black?

Question  2:Is  the  bill  hooked?

Computer  Vision

A:  NO

A:  YES

)|( xcp

),|( 1uxcp

),,|( 21 uuxcp

1u

2u

Max  Expected  Information  Gain

Max  Expected  Information  Gain

…26

Page 27: CMPSCI 670: Computer Vision

Without Computer Vision

Input  Image  (      )xQuestion  1:

Is  the  belly  black?

Question  2:Is  the  bill  hooked?

Class  Prior

A:  NO

A:  YES

)(cp

)|( 1ucp

),|( 21 uucp

1u

2u

Max  Expected  Information  Gain

Max  Expected  Information  Gain

…27

Page 28: CMPSCI 670: Computer Vision

Select the next question that maximizes expected information gain: • Easy to compute if we can to estimate probabilities of

the form:

Basic Algorithm

)...,,|( 21 tuuuxcp

Object  Class

Image Sequence  of  user  responses

28

Page 29: CMPSCI 670: Computer Vision

Basic Algorithm

Zxcpcuuup t )|()|...,( 21≈

Model  of  user  responses

Computer  vision  

estimate

)...,,|( 21 tuuuxcp

Normalization  factor

29

Page 30: CMPSCI 670: Computer Vision

Basic Algorithm

Zxcpcuuup t )|()|...,( 21≈

Model  of  user  responses

Computer  vision  

estimate

)...,,|( 21 tuuuxcp

Normalization  factor

30

Page 31: CMPSCI 670: Computer Vision

• Assume: • Estimate using Mechanical Turk

∏ =≈

ti it cupcuuup...121 )|()|...,(

Modeling User Responses

)|( cup i

What  is  the  color  of  the  belly?

Pine  Grosbeak

31

Page 32: CMPSCI 670: Computer Vision

• Use any recognition algorithm that can estimate: p(c|x)  • Two simple methods:

Incorporating Computer Vision

)}(exp{)|( xmxcp ⋅∝ γ

1-­‐vs-­‐all  SVM∏∝ i i capxcp )|()|(

Attribute-­‐based  classification

[Lampert  et  al.  ’09,  Farhadi  et  al.  ‘09]

32

Page 33: CMPSCI 670: Computer Vision

• Use combination of features to learn linear SVM classifiers

[Vedaldi  et  al.  ’08,  Vedaldi  et  al.  ’09]

Self  SimilarityColor  Histograms    Color  Layout

Bag  of  Words  Spatial  Pyramid

Geometric  Blur Color  SIFT,  SIFT

Multiple  Kernels

Incorporating Computer Vision

33

Page 34: CMPSCI 670: Computer Vision

•200 classes, 6000+ images, 288 binary attributes •Why birds?

Birds 200 Dataset

Black-­‐footed  Albatross

Groove-­‐Billed  Ani Parakeet  Auklet Field  Sparrow Vesper  Sparrow

Arctic  Tern Forster’s  Tern Common  Tern Baird’s  Sparrow Henslow’s  Sparrow34

Page 35: CMPSCI 670: Computer Vision

•200 classes, 6000+ images, 288 binary attributes •Why birds?

Birds 200 Dataset

Black-­‐footed  Albatross

Groove-­‐Billed  Ani Parakeet  Auklet Field  Sparrow Vesper  Sparrow

Arctic  Tern Forster’s  Tern Common  Tern Baird’s  Sparrow Henslow’s  Sparrow35

Page 36: CMPSCI 670: Computer Vision

•200 classes, 6000+ images, 288 binary attributes •Why birds?

Birds 200 Dataset

Black-­‐footed  Albatross

Groove-­‐Billed  Ani Parakeet  Auklet Field  Sparrow Vesper  Sparrow

Arctic  Tern Forster’s  Tern Common  Tern Baird’s  Sparrow Henslow’s  Sparrow36

Page 37: CMPSCI 670: Computer Vision

Results: Without Computer Vision

Comparing  Different  User  Models

37

Page 38: CMPSCI 670: Computer Vision

Results: Without Computer Vision

if  users  answers  agree  with  field  guides…Perfect  Users:  100%  accuracy  in  8≈log2(200)  questions

38

Page 39: CMPSCI 670: Computer Vision

Results: Without Computer Vision

Real  users  answer  questionsMTurkers  don’t  always  agree  with  field  guides…

39

Page 40: CMPSCI 670: Computer Vision

Results: Without Computer Vision

Real  users  answer  questionsMTurkers  don’t  always  agree  with  field  guides…

40

Page 41: CMPSCI 670: Computer Vision

Results: Without Computer Vision

Probabilistic  User  Model:  tolerate  imperfect  user  responses

41

Page 42: CMPSCI 670: Computer Vision

Results: With Computer Vision

42

Page 43: CMPSCI 670: Computer Vision

Results: With Computer Vision

Users  drive  performance:  19%  à68%

Just  Computer  Vision  19%

43

Page 44: CMPSCI 670: Computer Vision

Results: With Computer Vision

Computer  Vision  Reduces  Manual  Labor:  11.1à6.5  questions

44

Page 45: CMPSCI 670: Computer Vision

Examples

Without  computer  vision:        Q  #1:  Is  the  shape  perching-­‐like?  no  (Def.)  With  computer  vision:        Q  #1:  Is  the  throat  white?  yes  (Def.)

Western  Grebe

Different  Questions  Asked  w/  and  w/out  Computer  Vision  

perching-­‐like

45

Page 46: CMPSCI 670: Computer Vision

Examples

computer  vision  

Magnolia  Warbler

User  Input  Helps  Correct  Computer  Vision  

Is  the  breast  pattern  solid?  no  (definitely)

Common  Yellowthroat  

Magnolia  Warbler  

Common  Yellowthroat

46

Page 47: CMPSCI 670: Computer Vision

Recognition is Not Always Successful

Acadian  Flycatcher

Least  Flycatcher

Parakeet  Auklet

Least  Auklet

Is  the  belly  multi-­‐colored?    yes  (Def.)

Unlimited  questions

47

Page 48: CMPSCI 670: Computer Vision

Summary

11.1à6.5  questions

Computer  vision  reduces  manual  labor

Users  drive  up  performance

19%

48

Recognition  of  fine-­‐grained  categories

More  reliable  than  field  guides

Page 49: CMPSCI 670: Computer Vision

Summary

11.1à6.5  questions

Computer  vision  reduces  manual  labor

Users  drive  up  performance

19%

49

Recognition  of  fine-­‐grained  categories

More  reliable  than  field  guides

Page 50: CMPSCI 670: Computer Vision

Summary

11.1à6.5  questions

Computer  vision  reduces  manual  labor

Users  drive  up  performance

19%

50

Recognition  of  fine-­‐grained  categories

More  reliable  than  field  guides

Page 51: CMPSCI 670: Computer Vision

Summary

11.1à6.5  questions

Computer  vision  reduces  manual  labor

Users  drive  up  performance

19%

51

Recognition  of  fine-­‐grained  categories

More  reliable  than  field  guides

Page 52: CMPSCI 670: Computer Vision

• Relies on part and attribute questions • This may be:

• Hard to define for some categories, e.g. handbags, sofas, etc • Hard to annotate due to lack of domain-specific expertise or

language barriers

Drawbacks

52

Page 53: CMPSCI 670: Computer Vision

Similarity comparison based framework

53Wah et al, CVPR 14

Page 54: CMPSCI 670: Computer Vision

Interactive categorization

54Wah et al, CVPR 14

Optimal display found by maximizing information gain

Page 55: CMPSCI 670: Computer Vision

Learned embeddings

55Each image is visualized in as a point in 2d space

Page 56: CMPSCI 670: Computer Vision

Interactive categorization

56Wah et al, CVPR 14

Page 57: CMPSCI 670: Computer Vision

Localized similarity

57Wah et al, WACV 15

Page 58: CMPSCI 670: Computer Vision

Localized similarity comparisons

58

Page 59: CMPSCI 670: Computer Vision

Localized vs. Non-localized

59

11.53 g 9.85 questions

Annotations are also faster

Parts were found by clustering HOG features

Page 60: CMPSCI 670: Computer Vision

Related Work

20  Questions  Game[20q.net]

oMoby[IQEngines.com]

Many  Others:  Crowdsourcing,  Information  Theory,  Relevance  Feedback,  Active  Learning,  Expert  Systems,  …

Field  Guides[whabird.com]  

Botanist’s  Electronic  Field  Guide

[Belhumeur  et  al.  ‘08]

Oxford  Flowers[Nilsback  et  al.  ‘08]

[Lampert  et  al.  ’09]    [Farhadi  et  al.  ‘09]  [Kumar  et  al.  ‘09]

Attributes

60

Page 61: CMPSCI 670: Computer Vision

• Papers discussed today: • Visual recognition with humans in the loop, Branson et al., ECCV 2010 • Similarity comparisons for interactive fine-grained categorization, Wah et

al., CVPR 2014 • Learning localized perceptual similarity metrics for interactive

categorization, Wah et al., WACV 2015

• Minimize annotation effort: • Active learning, better user interfaces

• Learning perceptual similarity: • Stochastic Triplet Embedding, L van der Maaten, K Weinberger

• Computer vision and human computation workshop, CVPR 2014 • https://people.cs.umass.edu/~smaji/cvhc2014/index.html

Further thoughts and readings …

61