ALMA MATER STUDIORUM - UNIVERSITÀ DI …...1 ALMA MATER STUDIORUM - UNIVERSITÀ DI BOLOGNA Bologna,...
Transcript of ALMA MATER STUDIORUM - UNIVERSITÀ DI …...1 ALMA MATER STUDIORUM - UNIVERSITÀ DI BOLOGNA Bologna,...
1
ALMA MATER STUDIORUM - UNIVERSITÀ DI BOLOGNA
Bologna, September 7-10, 2010
Multimedia Databases:
Fundamentals,
Retrieval Techniques, and
Applications
A Short Course for Doctoral Students
University of Bologna
Result Accuracy, Use Cases and
Real Applications
Ilaria Bartolini - DEIS
22
Quality of the results and relevance feedback techniques
Use cases
Demos of some applications
Outline
I. Bartolini – MMDBs Course
33
Till now we focused on efficiency aspects
i.e., “How to efficiently execute a MM query?”
It is now time to consider the effectiveness of the MM data retrieval
process, which includes everything related to the user expectation!
Effectiveness in term of:
quality of result objects
availability of simple but powerful tools, able to smooth the processes of
query formulation/personalization
result interpretation
Effectiveness
I. Bartolini – MMDBs Course 44
Traditional metrics for evaluating the quality of result objects are
precision (P) and recall (R)
Quality of the results
I. Bartolini – MMDBs Course
2
55
measures the effect of false hits
measures the effect of false drops
Precision and recall
I. Bartolini – MMDBs Course
Retrieved and RelevantP
Retrieved
Retrieved and RelevantR
Relevant
How can a user effectively search?
Till now we have implicitly assumed that the user “knows” how to
formulate her queries
Although with traditional DB’s and a few attributes this might be a
reasonable assumption, when we consider many attributes/features it is
not clear how a user might guess the right combination of weights
How can you define the 64 weights of a color-based search using the
weighted Euclidean distance?
6I. Bartolini – MMDBs Course
The idea of relevance feedback
The basic idea of relevance feedback is to shift the burden of finding the
“right query formulation” from the user to the system [RHO+98]
For this being possible, the user has to provide the system with
some information about “how well” the system has performed in
answering the original query
This user feedback typically takes the form of relevance judgments
expressed over the answer set
The “feedback loop” can then be iterated multiple times, until the user
gets satisfied with the answers
Original Query
Evaluate
Query
Answers
user
New Query
FeedbackAlgorithm
User Feedback
7I. Bartolini – MMDBs Course
Relevance judgments
The most common way to evaluate the results is based on a 3-valued
assessment:
Relevant: the object is relevant to the user
Non-relevant: the object is definitely not relevant (false drop)
Don’t care: the user does not say anything about the object
Information provided by the relevant objects constitutes the so-called
“positive feedback”, whereas non-relevant objects provide the so-called
“negative feedback”
It’s common the case of systems that only allow for positive feedback
“Don’t care” is needed also to avoid the user the task of assessing the
relevance of all the results
Models that allow a finer assessment of results (e.g., relevant, very
relevant, etc.) have also been developed
8I. Bartolini – MMDBs Course
3
A practical example (1)
Euclidean distance 32-D HSV histograms
This is the initial query, for which 2 object are assessed as relevant by the user
QueryImage
Precision = 0.3 (including the query image)
9I. Bartolini – MMDBs Course
A practical example (2)
QueryImage
These are the results of the “refined” (new) query,
generated using the 1st strategy we will see
Precision = 0.6 (including the query image)
10I. Bartolini – MMDBs Course
A practical example (3)
QueryImage
These are the results of the “refined” (new) query,
generated using the 2nd strategy we will see
Precision = 0.8 (including the query image)
11I. Bartolini – MMDBs Course
A practical example (4)
QueryImage
And these are the results obtained by
combining the 2 strategies…
Precision = 0.9 (including the query image)
12I. Bartolini – MMDBs Course
4
Basic query refinement strategies
When the feature values are vectors, two basic strategies for obtaining a
refined query from the previous one and from the user feedback are:
Query point movement:
the idea is simply to move the query point so as to get closer to relevant
objects
Re-weighting:
the idea is to change the weights of the features so as to give more
importance to those features that better capture, for the given query at
hand, the notion of relevance
relevant
non-relevant
q
13I. Bartolini – MMDBs Course
New approach to interactive similarity query processing
Increases the performances of traditional relevance feedback
techniques; it complements the role of relevance feedback
engines by storing and maintaining the query parameters
determined during the feedback loop over time
Query
Default results
FeedbackBypass results
We realized two implementations of FeedbackBypass:
The first one is based on Wavelet
The second one uses Support Vector Machine (SVM)
FeedbackBypass case study [BCW00, BCW01]
14I. Bartolini – MMDBs Course
1515
Content-based MM data retrieval
Content-based MM data browsing
Automatic MM data annotation
By focusing on the effectiveness of user provided tools (i.e.,
interfaces)
Query formulation
Result interpretation
Use cases
I. Bartolini – MMDBs Course
Content-based MM data retrieval
This use case is the one we have assumed till now…
Many content-based MM data retrieval systems (both commercial and
research) have been proposed in the last ten years
especially for image and video DBs
16I. Bartolini – MMDBs Course
The growing list: ADL, AltaVista Photofinder, Amore, ASSERT, BDLP,
Blobworld, CANDID, C-bird, Chabot, CBVQ, DrawSearch, Excalibur Visual
RetrievalWare, FIDS, FIR, FOCUS, ImageFinder, ImageMiner, ImageRETRO,
ImageRover, ImageScape, Jacob, LCPD, MARS, MetaSEEk, MIR, NETRA,
Photobook, Picasso, PicHunter, PicToSeek, QBIC, Quicklook2, SIMBA, SQUID,
Surfimage, SYNAPSE, TODAI, VIR Image Engine, VisualSEEk, VP Image
Retrieval System, WebSEEk, WebSeer, Windsurf, WISE…
5
Windsurf case study: “St. Peter” query
17I. Bartolini – MMDBs Course
[ABP99, BCP00, BP00, BC03, Bar09a, BCP+09, BCP10]
Visual results: flat visualization
18I. Bartolini – MMDBs Course
Visual results: spatial visualization
19I. Bartolini – MMDBs Course
Effectiveness comparison example
“Flowers” query Winsdurf clusters Blobworld [CTB+99] clusters
Windsurf Blobworld
20I. Bartolini – MMDBs Course
6
Content-based image browsing
Till now we have implicitly assumed that the user “knows”
what she is looking for
how to formulate her queries
e.g., QBE paradigm
In some cases the user does not know at all what to look for; in these
cases a “browsing” activity should be supported
to determine a good starting point for searching
to get an overall view of the DB contents
to give the user the ability to organize her MM collections (e.g. personal photos albums) in a semi-automatic way
21I. Bartolini – MMDBs Course
Windsurf case study: flat browsing example
22I. Bartolini – MMDBs Course
[ABP99, BCP00, BP00, BC03, Bar09a, BCP+09, BCP10]
Spatial browsing example
23I. Bartolini – MMDBs Course
PIBE: Personalizable Image Browsing Engine
A novel adaptive image browsing engine
customizable hierarchical structure called Browsing Tree (BT)
graphical personalization actions to modify the BT
“local” reorganization of the DB
specific similarity criteria for each portion
(sub-tree) of the BT
user customizations persist across different sessions
p, s q r
BT example
PIBE case study [BCP06, BCP07]
24I. Bartolini – MMDBs Course
7
Visual results: vertical browsing
X
Visual results: horizontal browsing
Visual example of BT personalization
“blowfish”Initial BT
Custom BT
27I. Bartolini – MMDBs Course
The semantic gap problem
Characterizing the object content by means of low level features (e.g.,
color, texture, and shape of an image) represents a completely automatic
solution to MM data retrieval
However low level feature are not always able to properly characterize the
semantic content of objects
e.g., two images should be considered “similar” even if their semantic content is
completely different
This is due to the semantic gap existing between the user subjective
notion of similarity and the one according to which a low level features-
based retrieval system evaluate two objects to be similar
prevents to reach 100% precision results
28I. Bartolini – MMDBs Course
8
Possible solution
(Semi-)automatically provide a semantic characterization (e.g., by means
of keywords or tags) for each object able to capture its content
e.g., ([sky, cheetah] vs. [sky, eagle])
Combine visual features with tags by taking the best of the two
approaches [LSD+06, LZL+07, DJL+08]
[sky, cheetah] [sky, eagle]
29I. Bartolini – MMDBs Course
Scenique: Semantic and ContENt-based Image QUErying
Image retrieval and browsing system that profitably exploits both
low level features and manually and/or (semi-)automatically
associated textual annotations
Based on a simplified version of the multi-structural framework
[FKK+05] which allows objects, (i.e., images in our case) to be
organized into a set of orthogonal dimensions,
also called facets
Scenique case study [BC08b, Bar09]
I. Bartolini, SEBD 2009
“Photos of animals I took
during my summer vacations”
30I. Bartolini – MMDBs Course
Provides the user with two basic facilities:
1) an image annotator (…we will see it in a few minutes!!), that is able
to predict new tags for images, and
2) an integrated query facility which allows the user to search and
browse images exploiting both visual features and tags
possibly organized in visual and semantic facets
in the form of trees
default semantic facet to ensure compatibility with
systems/devices that do not consider any tag organization
(e.g., Flickr)
Principles of Scenique
31I. Bartolini – MMDBs Course
Feature-based retrieval
Images as a set of regions
Semantic-based retrieval
Multiple tags associated to images
WordNet as lexical ontology [Miller 1995]
IS-A relation
Semantic “relaxation”
Semantic similarity criterion to compare terms [Lin 1997]
Integration policies
Technical details
Flower
PetuniaOrchid
…Poppy
Flowering Plant
Seed Plant
…
…
flower petunia
“Petunia is a flower!”
“I want images of
flowers …”
32I. Bartolini – MMDBs Course
9
Query visual example
“I want images of bears from the
Arctic Ocean that look like the
provided one”
33I. Bartolini – MMDBs Course
Query visual example
34I. Bartolini – MMDBs Course
Browsing visual example
35I. Bartolini – MMDBs Course
Browsing visual example
36I. Bartolini – MMDBs Course
10
Browsing visual example
37I. Bartolini – MMDBs Course
Browsing visual example
38I. Bartolini – MMDBs Course
Automatically infer semantics to MM objects
Automatic objects annotation requires user intervention
1) Relevant feedback
Exploiting user feedback to understand which are real relevant objects
to the query
2) Learning
The system is trained by means of a set of objects that are manually
annotated by the user (training phase)
Exploiting the training set, the system is able to predict labels for
uncaptioned objects: the test object is compared to training objects;
labels associated to the “best” objects are proposed for labeling
(labeling & testing phases)
new image
user
?sky, rock, ground,
desert, Monument
Valley
DB images
39I. Bartolini – MMDBs Course
Imagination case study [BC08a]
Imagination: IMAGe (semI-)automatic anNotATION
Images as set of regions
Labels are tags which are associated at the image level
Graph-based approach (à la Page Rank)
3-level of graph objects
Images
Regions with low level features (i.e., color and texture)
Tags assigned to images
plus K-NN links computed on region similarities
“Given a new image provide tags that are affine to the image and
semantically correlated to each other”
40I. Bartolini – MMDBs Course
11
DB
images
I1 I2 I3
Intuitive example
regions
Iq
new
image
deer, grass,
bushbear, rock,
grass, water, ground
rock, bear,
grass, water, ground
tags ?, …, ?
41I. Bartolini – MMDBs Course
Imagination user interface
42I. Bartolini – MMDBs Course
Predicted tags
43I. Bartolini – MMDBs Course
Annotation visual example within Scenique“I want to annotate from
scratch the selected
image”
44I. Bartolini – MMDBs Course
12
4545
The same principles can be applied for the annotation of videos!
SHIATSU: Semantic-Based Hierarchical Automatic Tagging of Videos
by Segmentation using Cuts
Based on
Shot boundaries detection
Hierarchical annotations
SHIATSU case study [BPR10]
I. Bartolini – MMDBs Course 4646
Enjoy some
demo applications
I. Bartolini – MMDBs Course