Retrieving Geo-Location of Videos with a Divide & Conquer Hierarchical Multimodal Approach

download Retrieving Geo-Location of Videos with a  Divide & Conquer Hierarchical Multimodal Approach

of 8

Transcript of Retrieving Geo-Location of Videos with a Divide & Conquer Hierarchical Multimodal Approach

  • 8/13/2019 Retrieving Geo-Location of Videos with a Divide & Conquer Hierarchical Multimodal Approach

    1/8

    Retrieving Geo-Location of Videos with aDivide & Conquer Hierarchical Multimodal Approach

    Michele Trevisiol 1

    [email protected] Jgou 2

    [email protected] Delhumeau 2

    [email protected] Gravier 3

    [email protected] Web Research Group 1 Yahoo! Research 2 INRIA 3 CNRS/IRISA

    Universitat Pompeu Fabra Barcelona, Spain Rennes, France Rennes, FranceBarcelona, Spain

    ABSTRACTThis paper presents a strategy to identify the geographiclocation of videos. First, it relies on a multi-modal cas-cade pipeline that exploits the available sources of informa-tion, namely the users upload history, his social networkand a visual-based matching technique. Second, we presenta novel divide & conquer strategy to better exploit the tagsassociated with the input video. It pre-selects one or sev-eral geographic area of interest of higher expected relevanceand performs a deeper analysis inside the selected area(s)to return the coordinates most likely to be related to theinput tags. The experiments were conducted as part of theMediaEval 2012 Placing Task. Our approach, which differssignicantly from the other submitted techniques, achievesthe best results on this benchmark when considering thesame amount of external information, i.e. when not usingany gazetteers nor any other kind of external information.

    Categories and Subject DescriptorsH.3.1 [Information Storage and Retrieval ]: ContentAnalysis and Indexing; H.3.3 [ Information Storage andRetrieval ]: Information Search and Retrieval

    KeywordsGeotags, Location, Placing Task, Video Annotation, Flickr

    1. INTRODUCTIONGeotagging is the process of automatically adding geo-

    graphical identication metadata to media objects, in par-ticular to images and videos. This geo-information is calledgeotag(s) , and usually consists of the latitude and longitudeworld-map coordinates. Determining the place where thecontent has been captured dramatically extends the knowl-edge around the media object, especially when combined

    Work done while visiting PhD student at INRIA Rennes

    Permission to make digital or hard copies of all or part of this work forpersonal or classroom use is granted without fee provided that copies arenot made or distributed for prot or commercial advantage and that copiesbear this notice and the full citation on the rst page. To copy otherwise, torepublish, to post on servers or to redistribute to lists, requires prior specicpermission and/or a fee. ICMR13, April 1620, 2013, Dallas, Texas, USA.Copyright 2013 ACM 978-1-4503-2033-7/13/04 ...$15.00.

    Figure 1: Approach Model Sketch.

    with time information. Linking time- and geographical-related content offers a new and practical way of automati-cally searching, organizing or visualizing personal and pro-fessional media databases. It also enables the retrieval of various relevant content such as web pages, hence providingusers with a wide variety of location-specic information.

    In recent years, we have witnessed a dramatic increase inthe number of such geotagged media data. Due to the mas-sive spreading of GPS-enabled cameras and mobile phones,geographic coordinates are captured and attached to thecontent generated by these devices. However, most of themedia available remain deprived of this information. Forthis reason, the problem of automatically assigning geotagsto media content (and conversely) is a challenge that increas-ingly interests the Multimedia community, as reected bythe success of the Mediaeval benchmarks placing task [16].This paper presents an efficient and effective geotagging sys-tem to address this problem, which is illustrated in Figure 1.

    A multimodal strategy hierarchically processes the sourcesof information by decreasing order of expected informative-ness: Tags, users upload history and social graph, userspersonal information (home town) and visual content. Whenthe most reliable information is missing, the system solely re-lies on the most informative amongst the remaining sources,with the prior most likely location as a nal backup. We donot assume any prior knowledge about the city or countrythe video was taken in.

    Beyond this strategy, a key contribution is the tag analy-sis technique introduced to extract clues about the location.

    1

  • 8/13/2019 Retrieving Geo-Location of Videos with a Divide & Conquer Hierarchical Multimodal Approach

    2/8

    Our motivation is to identify the tags that are likely to con-vey some geographical information and to discard the onesthat are deemed irrelevant. Indeed, by classifying 54% of the tags from Flickr images using WordNet 1 , Sigurbj orns-son and van Zwol [19] observed that 28% of these tags wereclassied as locations, which suggests that tags have variousdegree of interest with respect to geotagging.

    After ltering out noisy tags in a pre-processing step, wepropose two different tag processing techniques, both scal-able and not limited to tags that are location names. Therst technique implements a text matching rule and servesas a (strong) baseline. The second approach is a radicallydifferent tag analysis technique. Based on a divide & con-quer strategy, the relationship between tags and coordinatesis analyzed to better reect the informativeness of tags withrespect to location.

    Our multimodal framework and this new joint tag/locationmethod are the main contributions of this paper. As sec-ondary contributions, we show the interest of different strate-gies borrowed from other contexts. In particular, we showthat the Okapi BM25 weighting scheme [ 20] is of interestin our context of video localization. Similarly, we have im-ported some techniques rst proposed for image retrieval.

    Experiments are carried out on the MediaEval 2012 plac-ing task. Our method is compared with the best partici-pating teams using the official evaluation protocol, and isshown to outperform the state of the art in the same setup,i.e. , when using the same official input provided by the or-ganizers. The specic interest of our tag-analysis techniqueis demonstrated by comparing it to a strong baseline, whichby itself already achieves state-of-the-art performance.

    The paper is organized as follows. Section 2 describes thebackground on geotagging, including related work and so-lutions adopted by Mediaeval participants that we comparewith. An overview of our multimodal processing pipelineand its components are presented in Section 3. The new joint tag/location method which is detailed separately inSection 4. The experimental setup and results are shown inSection 5. Section 6 concludes the paper.

    2. BACKGROUNDThis section makes a brief overview of key approaches and

    trends for geotagging of images and videos. As this topic hasreceived growing attention in the multimedia, computer vi-sion and social networks communities, we refer to the recentsurvey by Luo et al. [11] for a wider overview of the tech-niques. We also present the placing task of the MediaEval2012 benchmark and detail some approaches designed bysome participants. They will serve as a comparison bench-mark in the experiments.

    2.1 Related workThe location of a video or image is typically extracted

    from two main sources of information, namely the textual in-formation ( i.e. , tags, title, description) and the visual char-acteristics ( i.e. , global/local features). This specic infor-mation might be combined with some prior statistics onpossible locations. In some situations, the media data isnot associated with any textual information, therefore de-termining the location relies on the only available infor-mation, i.e. , the image or video itself. In contrast to the1 http://wordnet.princeton.edu/

    techniques exploiting metadata such as tags or EXIF infor-mation, these approaches are usually referred to as content-based approaches. In this line of research, Hays and Efros [3]proposed a purely visual approach 2 that estimates the im-age location as a probability distribution over the Earthssurface. Penatti et al. [14] proposed an approach called bag-of-scenes . First, they create a dictionary of scenes fromplaces of interest, where each of them can be represented bya certain type of low-level features. Then, each video frameis compared with the dictionary and the most similar sceneis selected. This allows associating a scene with each framein order to save more semantic information.

    However, the content-based approach alone is not reli-able enough to be considered effective. Multimodal solu-tions exploit more sources of information, in particular tex-tual information. The words extracted from the availabletext ( i.e. , title, description and tags) are often associatedto geographic coordinates to determine the most commonwords for each possible location. An example of such anapproach is the work of Serdyukov et al. [17], which con-structs a m n grid based on latitude and longitude, whereeach cell represents a location. Images whose locations areknown are associated with their corresponding grid cells.

    Finally, a language model is estimated from the tags associ-ated with a particular location, taking into account neigh-bors inuence and leveraging spatial ambiguity. Sergieh et al. [18] worked on the reciprocal problem, proposing a sta-tistical model for automatic image annotation. Given animage with coordinates, they infer some relevant tags basedon textual information from images in the database that arephysically located nearby and which have similar visual con-tent. Crandall et al. [2] used both image content and textualmetadata to predict the location of an image at two levelsof granularity: city level (about 100km), and at the indi-vidual landmark level (about 100m). However they limitedtheir experiments to a specic set of landmarks in a xed setof cities. In contrast, the problem considered in this papermakes no assumptions on the data set and on the level of granularity in the detection step. OHare and Murdock [ 13]proposed a statistical language modeling approach, also di-viding the Earth into grid cells. Their approach is based onthe Word-Document paradigm, and they investigate severalways to estimate the models, based on the term frequencyand the user frequency.

    2.2 MediaEval 2012MediaEval is an international evaluation campaign in which

    the Placing Task [ 16] is dedicated to the geo-localizationproblem addressed in this paper. The goal is to determineas accurately as possible the location, in terms of latitudeand longitude, of a set of Flickr videos. The task coversseveral cases, called runs , each of them being restricted bysome constraints on the type of information used. To ensurethat the training data is the same for all techniques so as toprovide a fair comparison, we focus on techniques that onlyused the information provided by MediaEval. Hence, extraresources, such as gazetteers ( e.g. , GeoNames, WordNet) orany kind of external information ( e.g. , Wikipedia, GoogleMaps), are excluded from all the experiments so as to focuson the data processing techniques proposed.2 Note that the dataset they consider only includes imagesassociated with a geotag such as a country, a city or as atouristic site ( e.g. , Pisa, Nikko, Orlando).

    2

    http://wordnet.princeton.edu/http://wordnet.princeton.edu/
  • 8/13/2019 Retrieving Geo-Location of Videos with a Divide & Conquer Hierarchical Multimodal Approach

    3/8

    no tags single tag sizeTrain Set 454 , 338 (14.2%) 27 , 488 (0.9%) 3 , 200 , 757

    Test Set 1 , 902 (45.5%) 139 (3.3%) 4 , 182

    Table 1: Number of media objects without tags.

    2.2.1 Dataset descriptionThe MediaEval 2012 Placing task dataset gathers content

    from Flickr in Creative Common license and is divided intoa train set with both images ( 3.2M ) and videos ( 15K )and a test set with 4 , 182 videos, from more than 71 K users.Metadata is associated with each media object and con-sists of various information such as ownership (Flickr user idand nickname), timestamps (upload and shot time), textualdata (tags/keywords, title and description), social network(owners contact user ids), comments and favorites (contentsand users that made them) and, of course, the latitude andlongitude within a certain level of accuracy. Note that inFlickr there are 16 levels of accuracy, from the most gen-eral ( i.e. , country name) to the most specic ( i.e. , streetaddress). Table 1 summarizes the number of objects withtag(s) associated for each dataset. Clearly, the test set in-cludes a large proportion of videos with no tags. Moreover,as tags are arbitrarily added by users without any constraint

    or rule, a large proportion of the tags is meaningless. Over-all, many annotated objects are not associated with a singleuseful tag. This makes this benchmark both challenging andrealistic.

    2.2.2 Evaluation protocolOur evaluation strictly follows the rules of the MediaEval

    2012 placing task. The accuracy of the estimated locationis measured by great circle distances between the predictedand the actual geo-coordinates encoded in the video. TheHaversine distance is used to measure the discrepancy be-tween the estimated location and the real one. The ground-truth is supplied by Flickr users at upload time.

    2.2.3 Description of submitted geotagging techniques

    Various approaches were taken by MediaEvals partici-pants to address the problem. This section presents somerepresentative methods, including the most successful ones,which are included in the comparison of Section 5. Choiet al. [1] gave priority to the textual information,using title and tags/keywords, but discarding the descrip-tion. They computed a geographic spread for each word (intags and title), similar to what we do. In addition, theyexploited the GeoNames database to have a toponym res-olution in order to lter out irrelevant words. They alsoincluded part-of-speech retrieved to perform more preciseltering using Augmented-WordNet 3 . In case of no can-didate coordinates, they used the users home location, oras last resort, the prior location ( i.e. , xed location com-puted a priori ). Li et al. [10 ] extended the successfulbag-of-scenes technique [14], including the histogram of motion patterns. They aggregated with a fusion moduleboth a textual (based on tags, title and description) and avisual approach. Interesting results are presented for thecontent-based (visual) task, but they are not the main focusof this paper. Popescu and Ballas [15 ] tackled the prob-lem by splitting the Earth surface in small cells of size 0 .01of latitude and longitude degree, characterized by a set of tags and their probability of occurrence in that cell. They3 http://ai.stanford.edu/~rion/swn

    selected only pairs of tags with a high probability of occur-rence within a smaller radius in order to extract a set of un-ambiguous pairs of potential toponyms. Then they matchedthe tags for each test video with the cells of the unambiguouspair (if it is found), or with the whole set of cells, consid-ering as top ranked the selected cells and their neighbors.Van Laere et al. [9] applied a divide & conquer approachsplitting the problem in two phases. Given the test video,in the rst step they nd the most likely cluster to containthe location with a Naive Bayes classier. Then with a sim-ilarity search, they nd the training items whose tags arethe closest to the ones of the test video. If the test videohas no tags, they use users hometown, title and descriptionas if they were regular tags. As a last resort, they also useda prior static location. Kelm et al. [8] presented a hi-erarchical framework that combines textual and visualfeatures for different granularity. First they divide up theEarth in regions using meridians and parallels, then theygenerated textual and visual prototypes for each of them.For the textual part, they translated in English tags, titleand description, then they extracted words using a NLP ap-proach, and nally they applied a stemmer and a stop-wordelimination. Given the test video, they select the region

    and the images/videos with highest probability to containthe extracted words (using a bag-of-words approach). Then,given a list of ranked candidates, with a visual search theyselect the most similar.

    3. MULTI-MODAL CASCADEThis section describes our multimodal and hierarchical

    processing pipeline. It starts with a tag comparison tech-nique based on frequency matching, followed by a descrip-tion of how the remaining sources of information are pro-cessed. As shown in Table 1, many videos in the test setare not described by tags. To handle these cases, we exploitadditional information in a pipeline: If one source of infor-mation is absent or fail to provide a reliable prediction, thenext is considered. The pipeline operates in the followingorder which was chosen according to the amount of infor-mation conveyed by each source, as discussed later in theexperimental section: a ) tags b) users upload history, socialinformation, c ) users home town, d ) content-based match-ing, e ) prior-location. This process is illustrated in Fig. 1.

    3.1 Tag processing: IR-frequencyThe frequency tag processing technique proposed here-

    after is the rst way we propose to exploit the tags. Thistechnique, which is referred to as IR-frequency in the follow-ing, mainly serves as a baseline. A better novel techniquewill be presented in the dedicated Section 4.

    3.1.1 Pre-processing

    Flickr normalizes the set of raw tags by lower-casing them,removing white space and stop-words, and replacing commaswith white space. For example, the set of tags Trip 2010,Sagrada Familia, Barcelona becomes trip2010 sagradafa-milia barcelona. Remember that tags are arbitrarily cho-sen by a user to describe the image. Hence, they might beinconsistent with the image content or location. We furthernormalized tags so as to dened a set of clean tags, T ctrainderived from T train , the entire set of tags in the trainingdata. We removed the accents, discarded numeric tags (al-most never relevant for the location), and removed numeric

    3

    http://ai.stanford.edu/~rion/swnhttp://ai.stanford.edu/~rion/swn
  • 8/13/2019 Retrieving Geo-Location of Videos with a Divide & Conquer Hierarchical Multimodal Approach

    4/8

    characters from the alphanumeric tags. A stop-list contain-ing common words ( e.g. , travel, birthday, cat, geotag, cam-era) and product or brand names ( e.g. , iPhone, Canon) wasused to lter out non informative tags. So called machinetags 4 (or mtags ), i.e. , one or more tags that Flickr recog-nized as a location (usually a country name or sometimes acity name), are kept unchanged and will be processed inde-pendently of the other tags as they are highly accurate andrelevant. Note that after pre-processing, only 39.9 % of thevideos contain tags.

    3.1.2 Geo-relevance lteringFor a baseline method based on direct tag matching, se-

    lecting tags relevant to the geo-location is a crucial step.Apart from machine tags which are deemed relevant, we im-plemented a geo-relevance ltering based on the geographicspread of a tag in the training data. Figure 2 illustrate thisidea by showing how some tags are spread across the globe:Tags specic to a location (bottom row) are mostly concen-trated in a small area while others (top row) exhibit a highdispersion.

    To select relevant tags in T c train , we compute for each tagt i its frequency of occurence f t i in the training data and

    the average Haversine distance dt i between the coordinatesof the data which contain t i . Tags that do not match thefollowing condition

    t i T train , t i T c train f t i 50,d t i 200.

    are removed from T ctrain where the thresholds were experi-mentally dened.

    3.1.3 Frequency matchingGiven the set of tags retained, one can group coordinates

    associated to the same set of tags. The idea is rst, to pre-select some set of tags that have at least one mtag in common(if available otherwise a normal tag), and nally to rank each

    of them by the occurrences of the common (m)tags.We consider each training document, image or video, as ageo-annotated document described by a set of tags. For eachset of tags, including machine tags, we collect all the coor-dinates from documents described by the same set of tags,along with the number of such documents. For example,for the set of tags france, pompidou and paris, we col-lect the following coordinates (48.8611, 2.3521):12, (48.6172,2.213):3.

    Given a test video, if it contains mtags we retrieve all thedocuments where there is at least one common mtag, other-wise we do the same with tags. Those documents are furtherranked according to the number of tags they share with thetest video. The top ranked document (or documents in caseof equality) is selected and the medoid of all the locationsattached to the corresponding set of tags, weighted by thenumber of occurences of the coordinates, is taken as the testmaterials geo-coordinate.

    3.2 User data processingWhen no tags are left after ltering or if no documents in

    the training data is found with at least one tag in commonwith the test video, we rely on the user data provided topredict a location.4

    http://www.flickr.com/groups/api/discuss/72157594497877875/

    3.2.1 User upload historyFor each user with images or videos in the training set, we

    picked a pre-computed user location based on the most fre-quent location for his content. We found that 35 .6% of theusers in the test set appear in the training data. Assumingthat users tend to visit the same places more than once, weseek to exploit the documents previously uploaded. For eachuser in the training set, we compute the medoid of the geo-coordinates of all its training data. The obtained locationis used as geo-coordinate when tag-based geotagging fails.We observed that using the user prior location signicantlyimproved the results.

    3.2.2 Social network extensionFor users not present in the training data, we make use

    of their social connections to infer a potential location. Theidea is to nd the user locations of all the contacts and usethe medoid as the most likely location for the test video.This general idea is rened based on the groups which areused in Flickr user connections, namely family , friends , andcontacts . We assume that family is closer to friends whichin turn is closer to contacts and process the groups in thatspecic order. If the user has enough connections in one

    group, then the video location is obtained from the contactsin the group. Else, we move on to the next group.Using both user upload history and social network exten-

    sion, 79% of the test videos are covered.

    3.2.3 User hometownIn case neither upload history nor social connections are

    available, the hometown of the user, as given by its Flickrprole is used. When available, the hometown is given as aplace name, e.g., San Francisco, California, United States,rather than as coordinates. We process the hometown infor-mation as if they were tags describing the test video, apply-ing the same process as described in Section 3.1.3 to deter-mine geo-coordinates. Note however that the user hometownis not always well specied ( i.e. , only the state or the coun-try is specied) and is not always precise ( e.g. , with verylarge cities like New York, the estimated coordinates can bevery far from the real ones).

    3.3 Content-based processingContent-based geo-tagging exploiting image matching is

    nally used. However, the input video set is not large enoughwith respect to the total number of locations, and includemany indoor scenes. Therefore the visual approach, whichrequires the same views of a given location, is less importantthan other sources of information. Anyway, for this purposeeach keyframe or image is described based on SIFT localdescriptors computed over a dense grid. A power law of 0.5is applied before L2 normalization [ 7]. PCA and whitening

    are applied before aggregating vectors into a global high-dimensionality VLAD descriptor [7] which is reduced to di-mension 1,024 by PCA, whitened and normalized. An indexis built from those descriptors using product quantization [6]which enables fast approximate nearest neighbor search onall of the test keyframes. For each query, we get the co-ordinates of the best candidate keyframes and return theirmedoid.

    3.4 When all elses fail...As a last chance, if all elses fail, we assign a default prior

    4

    http://www.flickr.com/groups/api/discuss/72157594497877875/http://www.flickr.com/groups/api/discuss/72157594497877875/
  • 8/13/2019 Retrieving Geo-Location of Videos with a Divide & Conquer Hierarchical Multimodal Approach

    5/8

    (a) beach (b) nature (c) iphone

    (d) italy (e) california (f) paris

    Figure 2: Coordinates of six tags plotted on the world map. The rst row shows the spreading of three tags that are notlocations. The second one shows respectively a country, a state, and a city.

    location regardless of the content. We experimented twostrategies for the default location: The medoid of all loca-tions in the training datawhich falls near Tokyoor themedoid of all locations from the training data with no tagswhich falls in London.

    4. HIERARCHICAL METHODThe tag processing described in the previous section re-

    mains limited and we seek to develop a new approach betterexploiting the information conveyed by the tags. In the fol-lowing, we propose a hierarchical approach based on thevector-space model, using the divide & conquer paradigm,to infer the relationship between tags and locations. This

    method will be referred to as IR-matrix in the following, byanalogy to the word-document matrix analysis performed insome information retrieval techniques.

    The tags of a given test video are considered as a query vector . The idea is to rst determine the approximate ge-ographic area in which the video is likely to belong, and tond in turn the most probable coordinates from the knownlocations in that area. The geographic areas are arbitrar-ily dened by quantifying the coordinates on a grid, whereeach cell of the grid is described by a vector of tag weightsaccording to the tag relevances to the area considered. Eachgeographic area is further dened by a set of geo-coordinatesalso described with a specic tag vector. The test documentis represented as a weighted vector of tags from which themost likely areas and the most likely coordinates are deter-

    mined.The steps described in this section replaces the tag lter-ing and frequency matching in the pipeline described in theprevious section. The other steps are left unchanged.

    4.1 Tag weightingDescribing a geographic area or a specic coordinate as

    a set of tags with weights require a weighting scheme thatreects the relationship between tags and coordinates. Sim-ilar quantities as those used for tag ltering in section 3.1.2,namely frequency and average distance, are used to mea-

    sure the relevance of a tag. Rather than eliminating nonrelevant tags as before, a low weight is given. The followingheuristics was used to identifying how geo-descriptive a tagis: t i T c train

    w t i =

    1 if f t i > 100K or d t i < 0.210 if f t i 200 and 10 d t i 505 if f t i 150 and d t i 701 otherwise

    This weighting was designed to assign higher weights totags representing geographic information, i.e. , not only placesbut also references to locations such as monuments. Figure 3shows an example of tags with the highest weight ( w t i = 10)sorted by f t i as opposed to tags sorted by frequency only.All tags with high weights clearly designate locations. Fig-ure 6 shows some examples of weighted query tags.

    4.2 Finding the areasGiven a set of tags from a test video, we rst want to

    identify the most likely geographic area(s).Areas were dened by quantifying the coordinates on a

    cell grid of 0.1 , i.e. a coordinates with latitude 41 .12 andlongitude 1.23 belongs to the area identied as (41 .1, 1.2).Though not the most compact representation, quantizationon a cells grid is computationally not expensive. Each cell j in the grid is described by a vector where each bin corre-sponds to a tag with the corresponding weight dened as:

    w t i ,j = f t i ,j w t i ,j (1)

    where f t i ,j is the number of occurences of t i in the area j ,and w t i ,j is dened from f t i ,j as in section 4.1.

    The set of areas is thus represented by a matrix whoserows correspond to tags and whose columns correspond tothe geographic areas. The Okapi weighting scheme is ap-plied to all entries in the area matrixsee section 4.4 fordetailsbefore smoothing using signed SQRT and L2 nor-malization, generalizing to text features results from imageprocessing [5]. The area that best ts a test image repre-sented as a vector of tag weights is obtained by multiplying

    5

  • 8/13/2019 Retrieving Geo-Location of Videos with a Divide & Conquer Hierarchical Multimodal Approach

    6/8

    Figure 3: On the left side of the line, there are listed tagsbefore the weighting scheme is applied. On the right sideinstead, there are shown the tags with highest score ( w t i =10). Both of lists are sorted by term frequency ( tf t i ).

    the query vector by the area matrix, thus providing a rankedlist of areas. The area with the highest matching score isselected, several areas being selected in case of equality.

    4.3 Finding the coordinatesGiven a selected area, we proceed to nd the most likely

    coordinates for the tags of the test video, following the sameprinciple as before. Similarly to what is done for areas,

    a tag/coordinate matrix is used to represent coordinateswithin each area, where each row corresponds to a tag andeach column corresponds to a coordinate in the area cell.The weights in the matrix are obtained following the sameprocedure as for the area matrix, with tag frequencies com-puted for each coordinate. Okapi weighting, smoothing andL2 normalization are also applied. Given a test query ob-tained from the tags of the test videos, a ranked list of co-ordinates is obtained within each of the areas selected inthe previous step. The best ranked coordinates are selectedfrom each of the ranked lists and the medoid is used as thegeo-coordinates for the test video.

    4.4 Tuning Okapi BM25While tf-idf is commonly employed as a weighting scheme

    for text representation in the vector-space model, the OkapiBM25 weighting scheme was experimentally found to per-form better in our case, conrming previous results [20].The Okapi weighs are dened as

    W BM ( j, t i ) = IDF (t i ) w t i ,j (k + 1)

    w t i ,j + k (1 b + b | D |avg dl

    )

    where w t i ,j is dened by Eq. 1, avg dl is the average numberof tags per training sample, k and b are free parametersusually chosen as k [1.2, 2.0] and b = 0 .75 [12]. The I DF

    radius (km) 0.001 0.01 0.1 1.2 10 201 756 749 752 720 714 713

    10 1626 1641 1627 1601 1587 1582100 2071 2086 2095 2085 2071 2068

    1000 2737 2739 2751 2760 2763 276010000 3885 3884 3890 3892 3889 3891

    Table 2: Estimating values of k 1 for the step of selection of the area, comparing different values of k1 from 0.001 to 20.For each radius (in km) the correctly detected coordinates

    for the test videos are counted.radius (km)

    k 1 , 1 =0 . 001k 1 , 2 =0 . 001

    k 1 , 1 =0 . 001k 1 , 2 =1 . 2

    k 1 , 1 =0 . 1k 1 , 2 =0 . 1

    k 1 , 1 =1 . 2k 1 , 2 =1 . 2

    1 786 756 752 72010 1635 1626 1628 1601

    100 2071 2065 2091 20791000 2759 2753 2769 2774

    10000 3962 3959 3964 3964

    Table 3: Estimating values of k1 for rst ( k1 , 1 , selection of the area) and for second step ( k1 , 2 , selection of coordinates).Where for each radius (in km) the correctly detected coor-dinates for the test videos are counted.

    part instead is given by

    IDF (q i ) = log N N t i + 0 .5N t i + 0 .5

    where N is the total number of training samples, and N t i isthe number of samples containing tag t i .

    We experimented different values of k , both for area selec-tion and coordinates selection. Contrary to the conclusionsof Whissell et al. [20], where large values of k (k 20) im-prove the results, we found that small values of k performedbetter in our case. Table 2 shows some results in terms of co-ordinates correctly identied for various values of k. Whilelarge values decrease performance, small values of k tend toincrease the accuracy at a small radius ( i.e. , 1 km). Variouscombination of k , for the coarse grain area selection and thene grain coordinate selection where tested, results being

    reported in table 3. Combining small values of k improvesfor both the 1 km and 10 km radii.

    5. EXPERIMENTSWe evaluate our approaches with the dataset from Me-

    diaEvals 2012 Placing Task described in Section 2.2. Oursystem was trained using the 3 .2M geotagged images andvideos released by the organizers. As shown as an outcomeof the 2011s campaign, the tag information is the most reli-able one. However, in many situations, a large proportion of the videos have no tag after our ltering steps, for instanceabout 60% on the Mediaeval benchmark. That is why thissection rst discusses the respective interests of the othersources of information, which led us to determine the order

    of priority in our cascade multi-modal approach. We thenpresent how our system performs on the Placing Task of Me-diaEval 2012, and shows the large improvement brought theIR-Matrix method of Section 4 compared to the baseline tagmethod and to the submitted techniques.

    5.1 Sources of informationSection 3 introduced the secondary sources of information

    that we exploit when the test video is not associated withany tag after the ltering step. In our cascade architecture,an important choice is the order in which the corresponding

    6

  • 8/13/2019 Retrieving Geo-Location of Videos with a Divide & Conquer Hierarchical Multimodal Approach

    7/8

    !

    #!!

    $!!!

    $#!!

    %!!!

    %#!!

    &!!!

    !!

    '!!!

    $() $!() $!!() $!!!() $!!!!()

    *+,-./0+,1 2*/3

    */4567809

    :6),;6-86- 2?6=16=3

    >-86- 2@6(A63

    B6=;,=;./0+,1

    Figure 4: Cumulative values of correctly detected locationsfor pipeline methods: number of video founds (y-axis) in aradius of x km (x-axis).

    methods appear in the pipeline, as this order impacts thenal quality of the system. For this purpose, we have eval-uated the respective geo-tagging accuracy provided by eachcomponent of our system 5 . Figure 4 shows the number of correctly identied locations for varying radius and for thedifferent sources, except for the visual search, which providesinferior results. For example, in the 1 km radius, the user-based estimation identies more than 600 video locations,while the estimation with Hometown nds less than 500 of them. However, for radius higher than 1 km, the interest of the Hometown improves and gives the best results amongthe secondary sources of information. Another observationis the social connections are useful and signicantly improvean estimation based only on the other metadata related tothe user. The content-based approach is performing poorlydue to the types of test videos that contain mainly indoorscenes. However for the smaller radii (1km and 10km) it isslightly better than the prior location, and for this reasonin our pipeline approach it is used before. The prior loca-

    tion does not use any information about the query and, as tobe expected, leads to a very imprecise estimation which onlyimpacts the 1000 and 10000 km precision measures. Interest-ingly, London and Tokyo give very different performances.However, in our opinion, this prior information is not reallyinteresting for a real application, as it is not related to a par-ticular video. The fact that the Hometown gives the bestresults for radius higher than 1 km suggests that it shouldbe used as the primary alternative to the tag-based method.However, when combining the different sources in cascade,our preliminary experiments showed that it is worth exploit-ing user-based and social information rst.

    To conclude this discussion, our nal framework uses theUB+Social as a primary alternative to the tag-based method(See Section 3). If this fails to output a location, we use theHometown estimation instead if provided, else the visualsearch engine. The prior location is used as a nal backup.

    5.2 IR-MATRIX EvaluationThis Section compares the results of our approach to the

    ones shown by the participants at MediaEval 2012 PlacingTask. In addition, we separately show the interest of the5 Note that this evaluation of the respective interest of in-formation sources was rst done on the 2011s Mediaevalcampaign, without knowing the 2012s test data.

    !"

    $!"

    %!"

    &!"

    '!"

    (!"

    )!"

    *!"

    +!"

    ,!"

    $!!"

    $-. $!-. $!!-. $!!!-. $!!!!-.

    /01234 5/46

    74389:; 57;6

    /0123?@ 5826

    AB04@>891C3DD 5A86

    31E:FGHIJ

    31EDHKLMKNOP

    Figure 5: Cumulative correctly detected locations: rate of video founds (y-axis) in a radius of x km (x-axis).

    IR-Matrix technique introduced in Section 4 over the IR-frequency baseline (See Section 3). These two tag-based ap-proaches are evaluated using the same pipeline, i.e. , in thesame multi-modal cascade. Therefore, the results of these

    methods are directly comparable6

    . Figure 5 shows that,overall, our IR-Matrix technique performs the best exceptin two cases: The last radius, 10000km, mainly depends onthe Prior location and has arguably no practical interest.As mentioned in footnote, the 1 km measurement is not reli-able because it is impacted by the artifacts of the train/testduplicates.

    Discussion. Among the various methods submitted to Me-diaeval2012, different textual lters have been used. BY [1]computes a geographic spreading based on the spatial vari-ance distribution ltering tags with high variance. BN [8]performs a more complex procedure, translating everything(also title and description) in English, applying a stemmingand stop-word ltering, and nally extracting words with

    a NLP approach. CL [15] works only with tags that havebeen used by at least two users, and considers only pairs of unambiguous pre-computed toponyms.

    Our IR-Matrix technique is less restrictive than these tech-nique because it does not discard any tag, except the onesltered by a common stop-word list. Instead, it automati-cally assigns different weights to each of them, which is lessradical than the techniques mentioned above and leads toexploit more tags, thereby reducing the information loss.In addition, by considering each cell of the Earth grid asa separate document, our Word-Document matrix-basedapproach better identies the relationship between tags andlocalization, which in turn provides a useful measure of geo-informativeness to tags.

    Concerning the secondary methods, only CL exploited theusers previous uploads in the case where no tags is associ-ated to the test video, and nobody used the social informa-tion in order to expand this knowledge. This gives a slightimprovement which is exploited in both our IR-matrix andIR-frequency methods.6 There is a bias for the 1 km radius measure, as some testimages were also included in the training set. This basicallyfavors the baseline approach for this measure (and other sys-tems) because one could match some test videos perfectlybased on irrelevant tags. We have not exploited this knowl-edge in our system.

    7

  • 8/13/2019 Retrieving Geo-Location of Videos with a Divide & Conquer Hierarchical Multimodal Approach

    8/8

    piazzabra verona italy venetonorthernitaly worldheritagesite unesco

    unescoworldheritagesite fountainwaterfeature video videoclip

    provence southoffrance francebouchesdurhoneandnimes bouchesdurhone

    provencealpescotedazuraixenprovence coursmirabeau fountain

    video videoclip lafontainedeauchaudethehotwaterfountain mossyfountain hotspring

    moussue bagniers aixcentre a ix

    norfolk england unitedkingdomgreatbritain westnorfolk hunstanton

    beach sea thewash northsea northpromenadecliff cliffs video videoclip chalk redchalk

    carstone whitechalk hunstantonformationnewhunstanton hunstantonstedmunds eastanglia

    Figure 6: Query examples with tags. Lightness and size indicate initial weights ( w t i = 10 , 5 or 1), ignored tags are striked-through ( stop words or not in the database ).

    6. CONCLUSIONSThis paper introduces a novel system for geo-tagging videos

    which signicantly outperforms techniques of the state of theart, as demonstrated by our experiments performed on thelast Mediaeval benchmark.

    A key contribution is the novel IR-Matrix location/tagtechnique based on the Divide & Conquer paradigm, whichis simply and efficiently implemented by (query)vector-mat-rices multiplications. It rst provides an estimation of thearea of interest, which is then used to determine more precisecoordinates that best match the input set of tags. It signif-icantly outperforms a more conventional tag-vector match-ing technique, such as our IR-Frequency baseline which rstdetects all the images and videos that contain the specictag(s), and then selects the one with the highest numberof matches. As a complementary technique, we show theinterest of the Okapi weighting scheme in this context.

    When no reliable tag is available, our processing cascadeallows our system to make a prediction based on other sourcesof information, such as user-related metadata or visual con-tent. To our knowledge, our system is also the rst to exploitthe social connections for this geo-tagging task.

    Although we only considered the meta-data provided inthe Mediaeval benchmark, i.e. , the Flickr data associatedwith the videos, we believe that integrating external sourcesof information, such as a gazetteer, should further improvethe overall localization performance, as demonstrated byother works in the eld.

    7. ACKNOWLEDGMENTSThis work was partially funded by OSEO, French state

    agency for innovation, in the framework of the Quaero projectand by Grant TIN2009-14560-C03-01 of the Ministry of Sci-ence and Innovation of Spain. Furthermore, we would liketo thank Vincent Claveau for his helpful suggestions.

    8. REFERENCES[1] J. Choi, G. Friedland, V. Ekambaram, and

    K. Ramchandran. The 2012 ICSI/Berkeley Video LocationEstimation System. In MediaEval , 2012.

    [2] D. Crandall, L. Backstrom, D. Huttenlocher, andJ. Kleinberg. Mapping the Worlds Photos. In WWW , 2009.

    [3] J. Hays and A. A. Efros. IM 2 GPS : estimating geographicinformation from a single image. In CVPR , 2008.

    [4] M. Jain, R. Benmokhtar, P. Gros, and H. Jegou. HammingEmbedding Similarity-based Image Classication. InICMR , Jun. 2012.

    [5] H. Jegou and O. Chum. Negative evidences andco-occurrences in image retrieval: the benet of PCA andwhitening. In ECCV , Oct. 2012.

    [6] H. Jegou, M. Douze, and C. Schmid. Product quantizationfor nearest neighbor search. PAMI , 33(1), Jan. 2011.

    [7] H. Jegou, F. Perronnin, M. Douze, J. S anchez, P. Perez,and C. Schmid. Aggregating local image descriptors intocompact codes. PAMI , Sep. 2012.

    [8] P. Kelm, S. Schmiedeke, and T. Sikora. How SpatialSegmentation improves the Multimodal. In MediaEval ,2012.

    [9] O. V. Laere, S. Schockaert, and J. Quinn. Ghent andCardiff University at the 2012 Placing Task. In MediaEval ,2012.

    [10] L. Li, J. Almeida, and D. Pedronette. A MultimodalApproach for Video Geocoding. In MediaEval , 2012.

    [11] J. Luo, D. Joshi, J. Yu, and A. Gallagher. Geotagging inmultimedia and computer visiona survey. Multimedia Tools Appl. , 51(1), Jan. 2011.

    [12] C. D. Manning, P. Raghavan, and H. Sch utze. Introduction to Information Retrieval . Cambridge University Press,2008.

    [13] N. OHare and V. Murdock. Modeling locations with socialmedia. Information Retrieval , Apr. 2012.

    [14] O. A. B. Penatti, L. T. Li, J. Almeida, and R. da S. Torres.A Visual Approach for Video Geocoding usingBag-of-Scenes. In ICMR , 2012.

    [15] A. Popescu and N. Ballas. CEA LISTs Participation atMediaEval 2012 Placing Task. In MediaEval , 2012.

    [16] A. Rae and P. Kelm. Working Notes for the Placing Taskat MediaEval 2012. In MediaEval , 2012.

    [17] P. Serdyukov, V. Murdock, and R. van Zwol. Placing ickrphotos on a map. In SIGIR , May 2009.[18] H. M. Sergieh, G. Gianini, M. D oller, H. Kosch,

    E. Egyed-Zsigmond, and J.-M. Pinon. Geo-basedAutomatic Image Annotation. In ICMR , 2012.

    [19] B. Sigurbj ornsson and R. van Zwol. Flickr tagrecommendation based on collective knowledge. In WWW ,2008.

    [20] J. Whissell and C. Clarke. Improving document clusteringusing Okapi BM25 feature weighting. Information Retrieval , 14, 2011.

    8