CAMEL: Concept Annotated iMagE Libraries - KU ITTCjsv/Papers/NCS01.camel.pdf · CAMEL: Concept...

12
CAMEL: Concept Annotated iMagE Libraries Apostol (Paul) Natsev a Atul Chadha b Basuki Soetarman c Jeffrey Scott Vitter a a Department of Computer Science, Duke University, P. O. Box 90129, Durham, NC 27708. Email: natsev, jsv @cs.duke.edu. b eJadoo, Inc., 2548 Minuet Drive, San Jose, CA 95131. Email: [email protected] c Enterprise Information Portal, IBM Silicon Valley Laboratory, San Jose, CA 95141. Email: [email protected]. ABSTRACT The problem of content-based image searching has received considerable attention in the last few years. Thousands of images are now available on the internet, and many important applications require searching of images in domains such as E-commerce, medical imaging, weather prediction, satellite imagery, and so on. Yet, content-based image querying is still largely unestab- lished as a mainstream field, nor is it widely used by search engines. We believe that two of the major hurdles for this poor acceptance are poor retrieval quality and usability. In this paper, we introduce the CAMEL system—an acronym for Concept Annotated iMagE Libraries—as an effort to address both of the above problems. The CAMEL system provides and easy-to-use, and yet powerful, text-only query interface, which allows users to search for images based on visual concepts, identified by specifying relevant keywords. Conceptually, CAMEL annotates images with the visual concepts that are relevant to them. In practice, CAMEL defines visual concepts by looking at sample images off-line and extracting their relevant visual features. Once defined, such visual concepts can be used to search for relevant images on the fly, using content-based search methods. The visual concepts are stored in a Concept Library and are represented by an associated set of wavelet features, which in our implementation were extracted by the WALRUS image querying system. Even though the CAMEL framework applies independently of the underlying query engine, for our prototype we have chosen WALRUS as a back-end, due to its ability to extract and query with image region features. CAMEL improves retrieval quality because it allows experts to build very accurate representations of visual concepts that can be used even by novice users. At the same time, CAMEL improves usability by supporting the familiar text-only interface currently used by most search engines on the web. Both improvements represent a departure from traditional approaches to improving image query systems—instead of focusing on query execution, we emphasize query specification by allowing simpler and yet more precise query specification. Keywords: CAMEL, WALRUS, concepts, content-based query, images, multimedia 1. INTRODUCTION The proliferation of digital images on the internet, and in domain-specific applications, has made the problem of searching among these images increasingly important in recent years. Image querying by content—or searching for visually similar images/regions to a query image/region—has important applications for internet search engines, E-commerce, medical diagno- sis, weather prediction, agriculture and environmental change tracking, insurance, entertainment, and the petroleum industry, among others. These applications require content-based querying and retrieval of images in diverse image domains such as internet, medical, satellite, seismic, etc. Despite the need for content-based image search and the abundance of image search application domains, however, the field is still not well established or widely popular. We believe that two major hurdles for making the image search field more mature and well accepted are its inconsistent retrievalquality and poor usability. With respect to quality, most image search engines return just a few relevant images with a significant amount of noise. This is largely due to the fact that content-based image searching is difficult to begin with, and correspondingly, most research up to date has been focused on that problem. The usability problem, however, has hardly CONTACT AUTHOR. The work was done while the author was at the Database Technology Institute, IBM Silicon Valley Laboratory.

Transcript of CAMEL: Concept Annotated iMagE Libraries - KU ITTCjsv/Papers/NCS01.camel.pdf · CAMEL: Concept...

CAMEL: ConceptAnnotated iMagE Libraries

Apostol(Paul)Natseva ��� Atul Chadhab�

BasukiSoetarmanc

Jeffrey ScottVittera

a Departmentof ComputerScience,DukeUniversity, P. O. Box 90129,Durham,NC 27708.Email: � natsev, jsv� @cs.duke.edu.

b eJadoo,Inc.,2548MinuetDrive,SanJose,CA 95131.Email: [email protected]

c EnterpriseInformationPortal,IBM SiliconValley Laboratory, SanJose,CA 95141.Email: [email protected].

ABSTRACT

Theproblemof content-basedimagesearchinghasreceivedconsiderableattentionin thelastfew years.Thousandsof imagesarenow availableontheinternet,andmany importantapplicationsrequiresearchingof imagesin domainssuchasE-commerce,medicalimaging,weatherprediction,satelliteimagery, andsoon. Yet, content-basedimagequeryingis still largely unestab-lishedasa mainstreamfield, nor is it widely usedby searchengines.We believe that two of themajorhurdlesfor this pooracceptancearepoorretrieval qualityandusability.

In this paper, we introducethe CAMEL system—anacronym for ConceptAnnotatediMagE Libraries—asan effort toaddressbothof theaboveproblems.TheCAMEL systemprovidesandeasy-to-use,andyetpowerful, text-only queryinterface,which allows usersto searchfor imagesbasedon visual concepts, identifiedby specifyingrelevantkeywords. Conceptually,CAMEL annotatesimageswith thevisualconceptsthatarerelevantto them. In practice,CAMEL definesvisualconceptsbylookingatsampleimagesoff-line andextractingtheirrelevantvisualfeatures.Oncedefined,suchvisualconceptscanbeusedtosearchfor relevantimagesonthefly, usingcontent-basedsearchmethods.Thevisualconceptsarestoredin aConceptLibraryandarerepresentedby an associatedsetof wavelet features,which in our implementationwereextractedby the WALRUSimagequeryingsystem.EventhoughtheCAMEL framework appliesindependentlyof theunderlyingqueryengine,for ourprototypewehavechosenWALRUSasa back-end,dueto its ability to extractandquerywith imageregion features.

CAMEL improvesretrieval quality becauseit allows expertsto build very accuraterepresentationsof visualconceptsthatcanbeusedevenby noviceusers.At thesametime,CAMEL improvesusabilityby supportingthefamiliar text-only interfacecurrentlyusedby mostsearchengineson the web. Both improvementsrepresenta departurefrom traditionalapproachestoimproving imagequerysystems—insteadof focusingon query execution, we emphasizequery specificationby allowingsimplerandyet moreprecisequeryspecification.

Keywords: CAMEL, WALRUS,concepts,content-basedquery, images,multimedia

1. INTRODUCTION

The proliferationof digital imageson the internet,and in domain-specificapplications,hasmadethe problemof searchingamongtheseimagesincreasinglyimportantin recentyears. Imagequeryingby content—orsearchingfor visually similarimages/regionsto aqueryimage/region—hasimportantapplicationsfor internetsearchengines,E-commerce,medicaldiagno-sis,weatherprediction,agricultureandenvironmentalchangetracking,insurance,entertainment,andthepetroleumindustry,amongothers. Theseapplicationsrequirecontent-basedqueryingandretrieval of imagesin diverseimagedomainssuchasinternet,medical,satellite,seismic,etc. Despitetheneedfor content-basedimagesearchandtheabundanceof imagesearchapplicationdomains,however, thefield is still notwell establishedor widely popular.

We believe that two major hurdlesfor makingthe imagesearchfield morematureandwell acceptedareits inconsistentretrieval quality andpoorusability. With respectto quality, mostimagesearchenginesreturnjust a few relevant imageswitha significantamountof noise. This is largely due to the fact that content-basedimagesearchingis difficult to begin with,andcorrespondingly, mostresearchup to datehasbeenfocusedon thatproblem.Theusabilityproblem,however, hashardly�

CONTACT AUTHOR.�Thework wasdonewhile theauthorwasat theDatabaseTechnologyInstitute,IBM SiliconValley Laboratory.

beenaddressedso far, even thoughfor someapplicationsit is the singlemostdecidingfactorfor acceptanceor rejectionofthe technology. For example,for internetsearchengines,it is very hardto find andspecifyanappropriatequeryimage.Theproblemof findingrelevantimagesis compoundedby theproblemof findinganappropriatequeryimage.Thetypicalsolutionsareto providesomerandomimagesuntil somethingremotelysimilar is found,andthento refinethesearch.An alternative isto manuallycategorizesomeimagesto beusedasqueryimages.Bothapproaches,though,areinefficient,andinconvenienttotheuserandthedatabaseadministrator.

In thispaper, weproposetheCAMEL systemthattriesto addressbothof thehurdlesdescribedabove,with emphasisbeingplacedonusability, whichhasmostlybeenignoredto thisdatedespiteof its importance.CAMEL standsfor ConceptAnnotatediMagE Librariesandprovidesa naturalapproachto theusabilityproblem.Theideais to allow usersto searchfor imagesbysimplyspecifyingthekeywordtermsthey areinterestedin, without requiringmanualannotationof all imagesbeforehand,andwithout losing the power of content-basedimagesearch.For example,givena largecollectionof patients’X-ray images,adoctormay want to searchfor imagesthat contain“lung cancer”in order to compareprevious casesto the currentpatient.Thefamiliar text-only queryinterfaceis clearlya significantimprovementin termsof usability. Thechallengeis how how topreserve theability to efficiently searchfor relevantimagesbasedon their visualcontent,andhow to avoid thesubjective andlabor-intensive processof manualannotationof all images.CAMEL addressesthatchallenge,aswell astheretrieval qualityproblem,by providing semi-automatic,highqualityobjectiveannotationof imageswith visualconcepts.

Theremainderof thepaperis organizedasfollows. Section1.1providessomebackgroundoncontent-basedimagesearch-ing anddescribesexisting approachesto that problem. Section1.2 givesa generaloverview of our proposedapproachandpointsout someof its advantages.The featuresof the back-endimagequerysystemchosenfor our prototypearedescribedin Section2. In Section3, we discussthe similarity modeladoptedby our framework andin Section4, we describethe ar-chitectureof the CAMEL systemin moredetail, alongwith a high-level descriptionof the WALRUS components.Finally,Section5 illustratesour implementationandsomesampleresults,andSection6 concludeswith asummaryof ourcontributionsandfuturework.

1.1. Previouswork

While work on improving the usability of imagesearchapplicationsis fairly scarce,the problemof improving the retrievalqualityof imagesearchsystemshasreceivedconsiderableattentionin theliterature.Thewidely acceptedapproachto solvingthisproblemis typically to extracta featurevector(or asignature) from every image,andto mapeachimageessentiallyinto a�-dimensionalpoint � , wherethecoordinatesof � correspondto thefeaturevectoror a low-dimensionalapproximationof it.

All of theimagesin thedatabasearethereforemappedto pointsin some�-dimensionalspace,andthepointsareindexedusing

a multi-dimensionalindex, suchasanR-Tree. During querying,thepoint correspondingto thequeryimageis usedto searchtheindex for neighboringpointsthatarewithin a smalldistanceaway from thequerypoint,with respectto somemetricsuchasEuclideandistance,for example.Theretrievedpointscorrespondto thesimilar images,whichmaybefilteredout furtherbyperformingfinersimilarity computation.Theapproachesin theliteraturediffer mainly in thetypesof featuresbeingextracted,themappingof imagesto ametricspace,thedistancefunctionusedfor similarity computation,andthetypeof indexingmethodbeingused.

Theearlyimagequerysystems,suchasIBM’ s QBIC,� � theViragesystem� by VirageInc., andthePhotobooksystem��� �from theMIT MediaLab, usedseparateindexesfor color histograms,shape,andtexture features.The taskof appropriatelyweighing the featuresrelative to eachotherwas not easy, however, and was left for the user, thushinderingusability sig-nificantly. Also, the fact that for eachfeaturetype, only a single featurevectorwasextractedfor the entire imagelead toinconsistentretrieval quality sinceoftenonesignatureper imageis not sufficient. Jacobset al. � werethefirst to usewaveletfeaturesthat capturedcolor, texture andshapesimultaneously. Their systemimprovedusability by eliminatingthe needtoweighthedifferentfeatures,andprovidedgoodperformancedueto theexcellentapproximationanddimensionalityreductionpropertiesof wavelets. TheWBIIS systemdevelopedby Wanget al. � improvedon the wavelet methodof Jacobset al. � byusingDaubechies’wavelets,abetterdistancemetric,anda three-stepsearchprocess.Still, though,all of thesemethodsextractonly onesignatureperimageandperformsimilarity comparisonsat thegranularityof theentireimage.

JohnSmith consideredimagequerysystemsthat integratespatialandfeatureinformation,both at the imageandregionlevels���� . Region extractionwasachievedby reverse-mappingregion featuresfrom a staticlibrary of imagepatterns.Hisapproachallowed the userto specifythe spatiallocationof regions,both in absolutetermsaswell asrelative to eachother,but thepipelinealsoallowedfor falsedismissals.Still, his Ph.D.thesiswasa very completetreatmentof region-basedimagequerying,andaddressedissuessuchasscale-,rotation-,andtranslation-invarianceof imageregions.

TheWALRUSsystem���� �� usedanalternativeapproachfor region-basedqueryingandplacedemphasisonachieving scale-and translation-invarianceof imageobjects. The systemcomputedwavelet signaturesfor hundreds,or even thousands,ofsliding windows per image. Thesignaturesweresubsequentlyclusteredin orderto identify a setof regionsfor eachimage.Thus,thesystemextracteda variablenumberof waveletregion featuresperimage,dependingon thecomplexity of theimage.By consideringwindowsof differentsizesandwith differentpositionsin eachimage,thesystemwasableto achieveexcellentretrieval quality. It alsoemployed a moreintuitive similarity model independentof scaleor locationof imageregions. Forfurtherdiscussionof theWALRUSsystem,seeSection2.

1.2. Proposedapproach

The CAMEL approachfor improving both retrieval quality and usability is a departurefrom the traditional approachestocontent-basedimagesearch.Insteadof concentratingon theactualsearchitself, andtrying to improve thequery execution,CAMEL focuseson thequery specification. Usability is improvedthroughasimplifiedqueryinterface,while retrieval perfor-manceis improveddueto moreexpressive representationof thequeryterm. Thekey ideais to split thequeryprocessin twostages:conceptcatalogingphase(doneoff-line by anexpert)andactualsearching(doneon-lineby user).Thetwo phasesareillustratedin Figure1. Notethatwe do not describea separateindexing phasebecauseit is donesolelyfor speeduppurposesandis necessaryfrom a purefunctionalityperspective.

����� �"!$#&%')("*,+.-�/$0 1&2,3�4

5"6�7 8 9;: <=?> @BADC;E;FHGJI$K

LNMPO Q"R$S&TUWV&X�Y.Z\[^]._H`

acb\dce\fcg\h\icjck\l\mcnco\pcqrtsvuxwzy|{|}�~�\�c�c�\��\�c�c�\��c�\�c�\�c�\�\�c�c�\�\�c�c�\�c�

���&���¡ &¢¤£¥§¦c¨ª©¬« ­¯®

°B±³²ª´¶µ$·¹¸ º¼»P½B¾À¿ªÁ

ÂÄÃÆÅÈÇÊÉ&˹Ì&ͼÎPÏÆЪѪÒ

Figure1. CAMEL systemoverview

Theconceptcataloging,or learning,phaseis usedto definevisualconceptsandbuild aconceptlibrary. Conceptsaredefinedby specifyingoneor moresampleimagesthatareusedby thesystemto automaticallyextract relevantsemanticinformationaboutthe conceptandto associateit with the givenconcept.The associationis performedby the ConceptCataloger andisstoredin a ConceptLibrary for future use. The ConceptLibrary is a modulefor persistentstorageof conceptsand it can

be maintainedandusedby many people. Note that even thoughthe ConceptLibrary shouldbe built by an expert in orderto maximizeperformance,it can later be modifiedby users,eitherby insertingnew conceptsor by refining existing ones.Also, suchConceptLibrariescancomeasoptionaldomain-specificplug-inmodules,similarly, for example,to IBM’ sviaVoicespeechrecognitionproductthatcanbeshippedwith specificadd-onmodulesfor recognizingdomain-specificterminology.

Thequeryingphasesimply takesa keyword phrase,andmapsit to a visualconcept(our implementationusestheidentitymappingbut moresophisticatedapproaches,includingtheuseof multiplekeywords,arediscussedin thefuturework section).TheConceptInterpreterthenlooksuptherepresentationof thevisualconceptfrom theConceptLibraryandusesit to querythedatabasefor imagescontainingtherelevantconcept.Eventhoughconceptuallythisapproachannotatesimageswith concepts,in reality, it is significantlydifferentfrom manualannotationof images.First of all, oncea conceptis defined,it canbeusedto searchfor relevant imageswithout having annotatedeachandevery oneof them. Second,the conceptdefinition is moreobjective than a person’s annotationbecausea conceptis representedwith visual features,as opposedto limited keyworddescriptions.Therefore,the imagesearchin CAMEL is basedon imagecontent(asopposedto a simpletext searchamongkeyword-annotatedimages),andthefinal similarity scorefor eachimagematchis moreintuitive. Third, theextractionof theappropriatesemanticinformationfrom eachimageis automatic,andtheconceptannotationis implicit. In contrast,themanualannotationapproachrequiresthateachimagebeexplicitly annotatedbeforeit canbeconsideredin akeywordquery.

2. WALRUSFEATURES

Thehigh-levelapproachthatwedescribedin theprevioussectionrequiresaback-endimagequerysystemto performtheactualimagesearchandpossiblyto constructthevisual conceptrepresentationfrom oneor moresampleimages.Even thoughtheapproachis generalenough,andis independentof theunderlyingqueryengine,therearesomeconsiderationsthatmakecertainqueryenginesmoresuitablethanothers.In thefollowing,weconsiderthemajorfactorsfor selectinga back-endquerysystemandwemotivateourchoice.

Themostfundamentalquestionrelatingto ourapproachis thenatureof theconcepts’representationin theConceptLibrary.Thesimplestandthemostsimilar to themanualannotationapproachis to usetext representation.However, in thatcase,wewould losetheability to queryby content,andin additionto themanuallaborrequired,theapproachwouldsuffer from lackofexpressivity for theconceptrepresentation.In thespirit of theold sayingthatapictureequalsathousandwords,webelievethatsearchbasedon imagecontentis a muchbetterandmoreobjectiveapproach.Thefirst suchapproachthatcomesto mind is tosimplystoretheimageIDs of somerepresentativeimagesfor eachvisualconcept.However, althoughthisapproachwouldhelpwith theusabilityproblem,it would be inefficient in termsof performance,bothwith respectto retrieval quality andrunningtime. Sinceultimatelythesearchis basedon imagefeaturesratherthantheimagesthemselves,storingimageIDs wouldmeanthat for eachquery, several imageswould have to be fetchedfrom disk andprocessedfor featureextraction. This would beredundantcomputationandcanbesavedsimply by doingit off-line andstoringtherelevant imagefeaturesfor eachconcept,ratherthanimageIDs. For retrieval quality considerations,storingonly the relevant imagefeaturesfor eachconceptwill bea moreaccuraterepresentationof theconcept,comparedto theentireimage.This is dueto thefact thateachimagecontainssomenoisein additionto therelevantcontent,andwithout pre-processingtheimageto extractonly therelevant information,this noisewould deterioratethequalityof theconceptrepresentations,andthereforetheoverall retrieval quality. We thereforeoptedto representconceptsby relevant imagefeaturesonly, andwe try to maximizetheamountof computationdonein theoff-line phasesothatthereis minimaloverheadin theonlinephase.ThatwayweachievebothCPUtimesavingsandincreasedretrieval qualitydueto higherprecisionof theconceptrepresentations.

Thenext questionthatwe needto answeris what featuresexactly we shouldstorein theconceptrepresentations.Oneofthemostcommonlyusedimagefeaturesis thecolor histogram,which encodesstatisticalinformationaboutthedistributionofcolorsin thepicture.Colorhistograms,however, donotcaptureany shapeor textureinformation,soby themselvesthey arenotsufficient to representconcepts.Combiningthemwith otherindependentfeaturesfor shapeandtextureis noteasyandusuallyplacestheburdenontheuserto specifyweightsfor thedifferentfeatures.Sincethisgoesagainstoureffort to improveusability,weusewaveletfeaturesinstead.Waveletsprovideverycompactrepresentations,offer excellentapproximationproperties,andcapturecolor, shape,andtexturein a singleunifiedframework.��� ��� ���� �� Ideally, we would alsowantto beableto extractsuchfeaturesfor imageregions,ratherthantheentireimages,andwe would like to generatea variablenumberof regionsbasedonthecomplexity of theimagecontent.Theregion featureextractionandmatchingat thesub-imagelevel areimportantbecausein orderto achieve the mostaccuratepossibleconceptrepresentations,we needto have the highestpossiblegranularityforimagefeatures.By beingableto extractimageregionsandtheir features,wecandescribeconceptsmorepreciselyandwecaneliminatenoisefrom sampleimagesmoreeffectively. SincetheWALRUS system�� meetsall of theabove requirements,andwasavailableto us,wechoseit asaback-endto theCAMEL system.

In order to extract regions for an image,WALRUS considerssliding windows of varying sizesand then clustersthembasedon theproximity of their waveletsignatures.Thus,thesystemgeneratesa variablenumberof featurevectorsfor eachimage—oneper imageregion,wherethenumberof imageregionsdependson thecomplexity of the image.Furthermore,byperformingsimilarity matchingat thesub-imagegranularity, WALRUS’s similarity modelcandetectsimilarity in caseswheresomeregionsin animagearescaledor translatedversionsof regionsfrom anotherimage.Theperformanceof theWALRUSsystemis very competitive in practiceandachievesvery goodretrieval quality. All of thesepropertiesarenaturallyinheritedby theCAMEL system.

3. SIMILARITY MODEL

Theimagesimilarity modelthatweusein theCAMEL framework is roughlythesameastheWALRUS modelbut with slightmodificationto accommodatethe notion of visual concepts.The WALRUS similarity model is illustratedin Figure2. Theformaldefinitionsappearin Natsev etal. ¬� Informally, thesimilarity betweentwo imagesis definedasthefractionof matchedimageareavs. the total imageareain both images.For simplicity, we assumeherethat thematchedareais simply theunionof all matchedregionsin the images.1 Imageregionsaresaidto be similar if their wavelet signaturesarewithin a certain Ódistancefrom eachother. Scaleandpositionof regionsareignored,asshown in Figure2, wheretranslatedandscaledversionsof queryregionsarematchedin thetargetimage.Accordingto theabovesimilarity model,thesimilarity scorebetweenthetwoimagesshouldbearound0.5 to 0.6 becausetheareaof matchedregionscomprisesabout50%to 60%of the total areain thetwo images.

Query Image Target image

Figure2. Imagesimilarity model

For the CAMEL system,we needto considerthe fact that the queryobject is not an imagebut rathera visual concept,consistingof regionsignaturesextractedfrom oneor moreimages.Sincethefeaturevectorsstoredin aconcept’srepresentationcorrespondto actualregionsfrom somesampleimages,theregionmatchingdefinitionstill appliesandwecouldidentify pairsof matchingregionsfrom thequeryconceptandthe target image. However, sincethe queryregionsno longercomefrom asingleimage,andwe do not storepositioninformationwith eachregion,we cannotsimply computetheunionof all matchedqueryregionsin orderto calculatethe total matchedqueryarea. Instead,we definethe similarity betweena queryconceptandtargetimageto betheweightedaveragebetweena querymatchscoreanda targetmatchscore.Thetargetmatchscoreisdefinedasbeforeto bethefractionof matchedareavs total imageareain thetargetimage.Thequerymatchscore,however, is

1An alternative definition is discussedin Natsev et al., Ô¹Õ wherethesystemenforcestherestrictionthateachimageregion of thequeryimagecanmatchnomorethanonetargetimageregionandviceversa.

definedasthefractionof matchedquerysliding windowsvs thetotal numberof querysliding windows. Recallthattheimageregionsareextractedby clusteringthewaveletsignaturesof sliding windows of the imagesoeachimageregion representsanumberof slidingwindows. By addingup thenumbersof slidingwindowsfor all matchedregions,anddividing it by thetotalnumberof sliding windows,we arrive at a sensibleway of computingthequerymatchscore,eventhoughthequeryconceptsarerepresentedby regionsextractedfrom multiple images.We formally definethesimilarity modelbetweena conceptandacandidateimageby modifying thefollowing definitionsfrom theoriginalWALRUSmodel�� :DEFINITION 3.1. (Similar region pair set) For a queryconceptÖ and a target image × , the setof ordered region pairsØÚÙ Ö ÜÛ × �Ý�ÛÊÞ�Þ�ÞÊÛ Ù Ö$ß Û ×àß Ý,á is referred to as a similar region pair set for Ö and × if ÖHâ is similar to ×àâ and for every ãåäæèç ,ÖHâ"äæ Ö^é and ×êâ"äæ ×zé .DEFINITION 3.2. (Imagesimilarity) A queryconceptÖ is saidto besimilar to a target image × if thereexistsa similar regionpair setfor Ö and × ØvÙ Ö Û × Ý , Þ�ÞÊÞ , Ù Ö ß Û × ß Ý,á , such that:

ëíìzî ïÚð � Ù�ñ ß âcò Ù ÖHâ ÝóÝî ïÚð � Ù Ö Ýô Ù�õÆö ë Ý ìzïÚð¶÷Üï

Ù�ñ ß âcò Ù ×àâ ݬÝïÚð¶÷øï Ù × Ýùûú

In theabove definition, î ïÚð � Ù¹ñ ß âJò Ù Ö.â Ý¬Ý is thenumberof sliding windows in Ö representedby regions Ö ÜÛ�ÞÊÞ�Þ Û Ö$ß . Theterm ïÚð¶÷øï Ù�ñ ß âcò

Ù × â Ý¬Ý representsthenumberof pixels in × coveredby regions × Û�Þ�ÞÊÞ�Û × ß consideredtogether. Theweightë

canbestaticor canbedynamicallymanipulatedby theuseraccordingto theapplicationneeds.Alternativedefinitionsarealsopossible,similarly to thevariationssuggestedby theWALRUSpaper.¬�

4. SYSTEM ARCHITECTURE

Our CAMEL prototypeis built on top of the WALRUS systemfor imageindexing and retrieval. Figure 3 illustratesthearchitectureof the CAMEL systemalongwith that of the WALRUS system. The WALRUS systemis comprisedof threemajorcomponents:region featureextraction,regionmatching,andimagematching.In addition,theCAMEL systemaddstheConceptCatalogerandtheConceptInterpretercomponents,aswell astheConceptLibrary storagemodule.

Thebasicoperationspreviously supportedby theWALRUS systemwereindexing animageandqueryingwith animage.Theimageindexing phaseinvolvesextractingregion signaturesandinsertingtheminto an üHý -treeindex, while thequeryingoperationinvolvesregionfeatureextraction,probingof the ü ý -treeindex to find matchingregionsto theextractedqueryimageregions,andfinally usingthematchedregionpairstocomputeafinal similarityscorefor rankingof eachcandidateimagematch.Theregion featureextractionprocedureconsistsof decomposingthe imageinto sliding windows, computinga Haarwaveletsignaturefor eachsliding window, andthenclusteringall window signaturesin orderto identify regionswith homogeneouswaveletsignatures.For a detaileddescriptionof theWALRUSsystemandits componentssee.��

Thenew functionalityaddedby the CAMEL systemincludesthe conceptcatalogingphaseandthe queryingby conceptphase. In the conceptcatalogingphase,the systemtakesas an input a conceptspecification(e.g., concept“apple”) andasampleimagecontainingthespecifiedconcept(i.e.,animageof anapple).ThesystemthenusestheWALRUS region featureextractionmoduleto get all regionsfrom the sampleimage. The regionsareclassifiedaseither important,unimportant, orneutral, accordingto a significancescoreevaluatedfor eachregion. Thesignificancescoreusesa heuristicformulathattakesinto accounttheimageareacoveredby theregion(i.e., thespanof theregion)andthenumberof slidingwindowsrepresentedby theregion (i.e., thecardinalityof theregion). Thefinal significancescorefor eachregion is a fractionalnumberbetween0and1, with 1 representingabsolutelycritical regions,and0 representingnoiseandregion outliers. Regionswith significancescorehigherthana certainthresholdareconsideredimportant, andarealwaysinsertedinto theConceptLibrary aspartof therepresentationof thespecifiedconcept.If thereis anexistingregionfeaturein theconcept’srepresentationthatis closeenoughto a new importantregion, thenthe two region clustersaremerged,andtheadjustedcentroid(or boundingbox) is usedasarepresentativefeaturefor thenew cluster. If thenew regionsignatureis notcloseto any of theexistingrepresentativesignatures,thenit is insertedasa new representative signaturefor thegivenconcept.If on theotherhand,thesignificancescorefor theregion is below the thresholdbut thereis an existing signaturesimilar enoughto the new signature,thenthe new signatureis deemedneutral andis mergedwith the existing region in orderto refineits signature.All othersignaturesareconsideredunimportantandarediscarded.The above mechanismallows for a moreselective procedurefor defininga visual concept’srepresentationandit triesto filter outnoisesothattheresultingconceptrepresentationis pure.Thisprocedureandthefactthatregion signaturesfrom multiple imagescanbeusedto refinetheconcept,improvesthequality of theconceptrepresentationandleadto betterretrieval quality. Improvementsin theheuristicscoringformulaandthethresholdestimationfunctionalityarestill openproblems.

þêÿ������������ ���������� ���

�����! �"�#$�%'&�(*)�+ ,.-�/10

2'3�4!5�6�7�8:9<;>=@?1A@B�C@D�EGF�H

IKJ�LNMPOQPRS>T U!VXW.Y[Z

\�]�^!_a`�bdcegf�h�i�j�kgl!m

n�o�p!qar�sdtu�v.w�xXy�zg{!|

}�~���G���G����>�����

�����.������

���������G���������!�

 �¡£¢�¤�¥¦@§¨G©.ª�«g¬!­

®°¯�± ²´³�µ·¶¸º¹�»!¼�½>¾¿�À�Á�Â�ÃGÄ�ÅGÆ1Ç

ÈÊÉ�Ë�Ì�Í Î�ÏÐ.Ñ Ò�Ó.ÔÕ*Ö�×@ØÙ�Ú Û.Ü·ÝGÞ�ßáàgâ�ã

äæå�ç.è�éëê ìí°î�ï�ð

ñóò�ô�õ�ö ÷�ø

ù1ú!û�ü�ý�þ ÿ ��������� ��������������������� ��! #"%$

& ')(+*-,). /1032547698

:<;>= ?+@ A1B3C5D�E9F

Figure3. Architectureof theCAMEL system

Thelastpieceof new functionality includestheConceptInterpreterandthequeryingby conceptphase.During anactualqueryby concept,the systemtakesasinput only the concept,andusesthe ConceptInterpreterto look up its representationfrom theConceptLibrary. Thereturnedsetof conceptfeaturesarethenusedto probethe GIH -treeindex to find matchingimageregions,andthe modifiedsimilarity scoringis usedto rank potentialimagematchesrelevant to the specifiedconcept. Thepseudo-codefor theabovefour basicoperationsis listedin Figures4 and5.

5. IMPLEMENT ATION

In this sectionwe show somesamplequeriesof the CAMEL system,andcompareits retrieval quality with the WALRUSsystem.Theresultsin thissectionarenotmeantto beanextensivestudyonperformancebut ratherto appearsimplyasaproofof concept,andto demonstratethatusabilitycanbeimprovedwithoutnecessarilysacrificingperformance.

In termsof retrievalquality, weshow thatarefinedconceptrepresentationcanachievebetterperformancethantheWALRUSsystem,which by itself hasvery goodretrieval quality. Dueto spaceandtime constraintswe do not reportcomparisonswithimagequerysystemsotherthanWALRUS, althoughCAMEL did performfavorablywith respectto QBIC, anda comparisonof WALRUS with an alternative wavelet approachcan be found in the WALRUS paper.JLK The running time and spaceconsiderationsof CAMEL arecomparableto that of the underlyingimagequeryengine,or WALRUS in our case. We dohave someextra spacerequirementsto storetheConceptLibrary but they aregenerallyoffsetby the fact that thedatastoredthereconsistsof precomputedfeaturevectorsandthereforesavesCPUtime. Also, it couldbearguedthatwithout theConceptLibrary, asearchenginewouldneedto storeactualimagesontheclientsideor transmitthemthroughthenetwork beforeeveryquery, which is somewhatequivalentto keepingthat informationin a ConceptLibrary. Also, our experiencehasbeenthattheconceptrepresentationsin theConceptLibrary areverycompactdueto theregionfiltering andnoiseeliminationperformedin

Index imageoperation:

1. Extractregionfeaturesfrom input image:

(a) Generateslidingwindowsof differentsizes(b) DynamicallycomputeHaarwaveletfor all slidingwindows

(c) Clusterwaveletsignaturesfor all slidingwindows(d) Useeachcluster’s centroid(or boundingbox)asa regionfeature

2. For eachregionsignature,do:

(a) Inserttheregionsignatureinto GIH -treeindex

Query by imageoperation:

1. Extractregionfeaturesfrom input image M :

(a) Generateslidingwindowsof differentsizes(b) DynamicallycomputeHaarwaveletfor all slidingwindows(c) Clusterwaveletsignaturesfor all slidingwindows

(d) Useeachcluster’s centroid(or boundingbox)asa regionfeature

2. For eachqueryregionsignatureGON , do:

(a) Probethe GIH -treeindex to find similarsignatures

3. For eachreturnedregionmatchGQP from targetimageR , do:

(a) Add R to thesetof candidatematchesS , if it’snotalreadypresent(b) Add GOP to theunion TQRVUWRQX of matchedtargetregionsin imageR(c) Add GON to theunion TIM3UWROX of matchedqueryregionsby imageR

4. For eachtargetimageR from S , do:

(a) Use TQRVUYROX and TIM3UWROX to computesimilarity scoreZ[UYROX for targetimageR5. Sortcandidateimagesby similarity score

6. Piperankedlist of imagematchesto theoutput

Figure4. Pseudo-codefor WALRUSfunctionality

theconceptcatalogingphase.We canoftenstoremoreexpressive conceptrepresentationsderivedfrom several imageswithroughlythesamenumberof region featuresasgeneratedfrom asingleimagein theWALRUSsystem.

Our implementationof the CAMEL prototypesystemis built on top of a re-implementedlibrary of WALRUS back-endfunctionality. As in theoriginal WALRUS implementation,we usedtheBIRCH pre-clusteringphasefor clusteringof signa-tures.J�\ Wehavealsousedthe GIH -treeinstantiationof theGiSTpackage1 asadisk-basedindex for thefeaturevectors,andtheImageMagicklibrary2 for readingvariousimageformatsandconvertingbetweendifferentcolor spaces.TheConceptLibrarywasimplementedasthreedatabasetablespopulatedandqueriedthroughtheCall Level Interfaceof IBM’ sDB2 v. 6.1database.Thedatasetweusedfor queryingis theoneusedin theWALRUSpaper,JLK andconsistsof approximately10000JPEGimageswith sizes]_^a`cbedf] , bgdh]3`i]_^ , or j_ka`lbgdh] . Thequeryresponsetime wasin theorderof 10 seconds.The GIH -treeindex wasslightly biggerthanthetotal sizeof theimagessince,on theaverage,we storeabout25 featurevectorsper image.Theimagedatabasewe usedcontainedrelatively small images(seeabove) andwasthereforevery compact.We don’t expectthesideoftheindex to benecessarilylargerthantheoriginaldatabasein otherapplicationscenarios,wheretheimagesarelarge.

Figure6 shows a samplequeryresultof theWALRUS system.Thequeryimageis in thetop left cornerandhasanID of885. Theclusteringandqueryingparametersusedin theWALRUS systemwereasspecifiedin theoriginal paper,JmK andthe

1Availableat http://epoch.cs.berkeley.edu:8000/gist/libgistv1.2Availableat http://www.wizards.dupont.com/cristy/ImageMagick.html

Catalogconceptoperation:

1. Extractregionfeaturesfrom inputsampleimage:

(a) Generateslidingwindowsof differentsizes(b) DynamicallycomputeHaarwaveletfor all slidingwindows

(c) Clusterwaveletsignaturesfor all slidingwindows(d) Useeachcluster’s centroid(or boundingbox)asa regionfeature

2. For eachregionsignature,G , do:

(a) npo�qhrYsutwv9xysgz{o}|~|�o{� = False

(b) Computesignificancescorefor thatregion, Z�Z[U�G�X(c) For eachexisting regionfeature� in givenconcept’s representation,do:

i. If �ur�|���UY���yG�X9��� , thenA. Merge G into �B. Usenew cluster’scentroid(orboundingbox)asafeaturerepresentativefor new region �C. npo�qhrYsutwv9xysgz{o}|~|�o{� = TrueD. Break

(d) If nOo�qfrYsft�v9xysgz~o�|~|�o~� = False&& Z�Z[UYG�X9��� , then

i. Insert G asnew regionfor givenconcept(e) Elsediscardregion G

Query by conceptoperation:

1. Lookupconcept’s regionfeaturesfrom theConceptLibrary

2. For eachregionsignatureGQN , do:

(a) Probethe GIH -treeindex to find similarsignatures

3. For eachreturnedregionmatchG P from imageR , do:

(a) Add R to thesetof candidatematchesS , if it’snotalreadypresent(b) Add GOP to theunion TQRVUWRQX of matchedtargetregionsin imageR(c) Add cardinalityof G N to thecardinality ��M3UWROX of matchedqueryconceptfeaturesby imageR

4. For eachtargetimageR from S , do:

(a) Use TQRVUYROX and ��M�UWROX to computesimilarity scoreZ[UYROX for targetimageR5. Sortcandidateimagesby similarity score

6. Piperankedlist of imagematchesto theoutput

Figure5. Pseudo-codefor CAMEL functionality

wavelettransformwasperformedin theYCC color space.Thetop 10 images(9 if we don’t countthequeryimage)arelistedfrom left to right, top to bottom,in orderof decreasingsimilarity. With theexceptionof image5712,all of thereturnedimagesareindeedvisually similar to the query. We shouldnotethat for all threesamplequeriespresentedin this paper, the top 20imagesappearingimmediatelyaftertheonesreportedhereweretypically visually similar to thequery, with someexceptions.For technicalandprintingconsiderationsweonly show thetop10 resultsfrom eachquery.

Figure7 illustratesa samplequeryresultin theCAMEL system.Theconceptusedfor queryingwasdefinedusingonly asingleimage(with ID 885)asanexample.We adjustedthenoisefilter sothatregionswith spanbelow 50%of theimageareaarediscardedasnoise. Dueto noisefiltering effects,theCAMEL systemstoresonly 9 out of the11 regionsgeneratedfromimage885.jpg.Fromtheresults,wecanobservethatthequality is slightly worsethanthecorrespondingWALRUSquery—thiscanbeattributedto themodifiedsimilarity modelwhich canonly considercardinalityratio asa matchingscorefor thequery

(a)Query (b) 861.jpg (c) 891.jpg (d) 866.jpg (e)821.jpg

(f) 883.jpg (g) 5712.jpg (h) 879.jpg (i) 890.jpg (j) 847.jpg

Figure6. Imagesretrievedby WALRUS.ThequeryimagehasID 866andgenerates11 regions.

(a)885.jpg (b) 866.jpg (c) 821.jpg (d) 845.jpg (e)879.jpg

(f) 8408.jpg (g) 891.jpg (h) 861.jpg (i) 890.jpg (j) 8146.jpg

Figure7. Imagesretrievedby CAMEL for a flowerconceptqueryrepresentedby 9 regionsextractedfrom image885.jpg.

concept,asopposedto themoreexpressive matchedareafractionusedfor images.However, thequality is still excellentandconsideringthefactthattheconceptwasdefinedusinga singleimage,thequeryresultsarestill promising.

The potentialfor improved retrieval quality is moreclearly establishedin Figure8, wherethe flower conceptis refinedby considering4 samplepictures,ratherthanjust one. In otherwords,we took four examplesof flower pictures,extractedtheir regions,mergedthesimilar ones,anddiscardedtheoutliers. This resultsin a slightly biggerconceptrepresentationbutproducesthebestresultsout of thethreeby returningrelevant imagesin all of thetop spots.In addition,the imagesreturnedin thethird queryaremorediversethantheprevioustwo queryresultsin thatthey includepicturesof singleflowers(e.g.,812,821,883,861,885,and847),aswell asgroupsof asmallnumberof flowers(e.g.,865and892),andalsoflowersappearingasa bunch(e.g.,866and845). Thateffect is not clearlyexpressedin theWALRUS queryresultsor in the initial conceptquery,which shows that theretrievedanswerscanbemanipulatedby carefulrefinementof thevisualconceptso that theresultsaretailoredtowardsa specificapplication.Thiscapabilityaloneis animportantadvantageof theCAMEL system.

(a)866.jpg (b) 845.jpg (c) 812.jpg (d) 821.jpg (e)883.jpg

(f) 865.jpg (g) 861.jpg (h) 892.jpg (i) 885.jpg (j) 847.jpg

Figure 8. Imagesretrievedby CAMEL usinga refinedflower representationconsistingof 25 regionsfrom images885.jpg,865.jpg,866.jpg,and892.jpg.

6. CONCLUSION AND FUTURE WORK

In this paper, we have presenteda new directionfor improving currentimagequerysystems,namelyfocusingon the queryspecificationpartasa gateway to betterperformanceandusability. TheCAMEL systemthatwe have proposed,coupledwiththe WALRUS systemasa back-end,performsimagematchingat the sub-imagelevel andusesan intuitive similarity modelthat is robustwith respectto scaleandposition.Thewaveletregion featuresextractedby theWALRUS systemcapturecolor,textureandshape,andenablehigh quality conceptrepresentationsto beautomaticallyconstructedfrom sampleimages.Theseparationof queryingfrom featureselectionprovidesa way for expertsto build domain-specificconceptlibrariesthatuserscanrefineevenfurther. As a resultof combiningsemanticinformationfrom multiple imagesinto a singleconcept,thequeriesbecomemoreexpressive andmorerepresentative of what usersarereally looking for, andthereforeusersobserve increasedretrieval quality. The flexible query specificationmechanismof CAMEL allows novice usersto query througha familiartext-only interface,without theneedfor intensive manualannotationlaborandwhile allowing expertsto utilize thepowerfulsearchfeaturesof WALRUS.Thus,CAMEL successfullycombinesthepowerof content-basedqueryingwith thesimplicity ofkeywordquerying.

As futurework wewould like to considersemantichierarchiesof visualconcepts,sothatlower level concepts(e.g.,“pine”and“oak” aregroupedunderhigher level categories(e.g., “tree”). This extensionis motivatedby the fact that the current“flat” architecturemakesit hardto definemoregeneralconcepts,suchas“animals”, for example,sincetheimagesthatfall insuchcategoriesmaybe very differentfrom eachother. We would alsolike to investigatewaysfor automatictaxonomyandclassificationof images,basedonsuchconcepthierarchies.

Anotherimprovementthatwe arecurrentlyconsideringis the introductionof spatialconstraintsin thequeryengine.Forexample,we maywant to preserve thescaleandpositionof two objectsrelative to eachother, eventhoughtheabsolutescaleandpositiondo not matter. As anexample,if we arelooking for picturesof a motherwith a youngdaughter, we might wantto imposethe restrictionthat the motheris roughly twice astall asthedaughter, andalsothat they areneareachother. Theproximity constrainthasanequivalentanalogyin text searching,wherethelocationsof thekeywordsin thedocumentaretakeninto accountwhenrankingthedocuments.Thus,if thekeywordsappearin thesamesentenceor paragraph,thecorrespondingdocumentis rankedhigher.

Finally, anotherwayof improvingusabilityevenmorewouldbeto investigatemoresophisticatedmappingsfrom keywordsto visualconcepts,aswell asBooleancombinationsof keywords. In thecurrentimplementation,this mappingis the identitymappingandeachkeyword representsa uniqueconcept.An alternative would be to usea thesaurusandcomputesynonymequivalenceclassesthat representvisual concepts(using for exampleWordNetJm� ). During a query, the systemwould thentransformeachkeyword into the canonicalrepresentationof an appropriateequivalenceclass(perhapswith the help of theuser),sothattheusercanfind relevantmatchesevenif they didn’t phrasethequeryprecisely. Thishasalreadybeeninvestigatedin thecontext of text searchandis currentlyavailable,for example,in IBM’ sDB2 Text Extender.

REFERENCES

1. C. Faloutsoset al., “Efficient andeffective queryingby imagecontent,” Journal of Intelligent InformationSystems3,pp.231–262,1994.

2. W. Niblack et al., “The QBIC project:Queryimageby contentusingcolor, textureandshape,” in Storage andRetrievalfor ImageandVideoDatabases, pp.173–187,SPIE,(SanJose),1993.

3. M. Flickner, H. Sawhney, W. Niblack,J.Ashley, B. Dom,Q.Huang,M. Gorkani,J.Hafner, D. Lee,D. Petkovic, D. Steele,andP. Yanker, “Queryby imageandvideocontent:theQBIC system,” IEEEComputer28(9), pp.23–32,1995.

4. A. GuptaandR. Jain,“Visualinformationretrieval,” Communicationsof theACM 40(5), pp.69–79,1997.5. R. W. PicardandT. Kabir, “Finding similar patternsin large imagedatabases,” in IEEE ICASSP, vol. V, pp. 161–164,

(Minneapolis),1993.6. A. Pentland,R.W. Picard,andS.Sclaroff, “Photobook:Content-basedmanipulationof imagedatabases,” in SPIEStorage

andRetrieval ImageandVideoDatabasesII , (SanJose),1995.7. C. E. Jacobs, A. Finkelstein, and D. H. Salesin, “Fast multiresolution image querying,” in

Proc. of SIGGRAPH 95, Annual Conference Series, pp. 277–286, August 1995. Available athttp://www.cs.washington.edu/research/projects/grail2/www/pub/abstracts.html.

8. J. Z. Wang, G. Wiederhold, O. Firschein, and S. X. Wei, “Content-basedimage indexing and searchingusingDaubechies’wavelets,” Intl. Journal of Digital Libraries (IJODL) 1(4), pp. 311–328,1998. Availableat http://www-db.stanford.edu/� zwang/project/imsearch/IJODL97/.

9. J.R. SmithandS.-F. Chang,“Integratedspatialandfeatureimagequery,” MultimediaSystems7(2), pp.129–140,1999.10. J. R. Smith andS.-F. Chang,“Queryingby color regionsusingthe VisualSEEkcontent-basedvisual querysystem,” in

IntelligentMultimediaInformationRetrieval, T. M. Maybury, ed.,IJCAI, 1997.11. J. R. Smith, Integrated Spatial and Feature Image Systems: Retrieval, Compression and Analysis.

PhD thesis, Graduate School of Arts and Sciences, Columbia University, Feb. 1997. Available athttp://www.ctr.columbia.edu/� jrsmith/publications.html.

12. A. Natsev, R. Rastogi,andK. Shim, “WALRUS: A similarity retrieval algorithmfor imagedatabases,” in Proc. 1999ACM SIGMODInternationalConferenceonManagementof Data, pp.395–406,(Philadelphia,PA), May 1999.

13. A. Natsev, R. Rastogi,andK. Shim,“WALRUS: A similarity matchingalgorithmfor imagedatabases,” tech.rep.,BellLaboratories,MurrayHill, 1998.

14. T. Zhang,R. Ramakrishnan,andM. Livny, “BIRCH: An efficient dataclusteringmethodfor very large databases,” inProceedingsof theACM SIGMODConferenceonManagementof Data, pp.103–114,(Montreal,Canada),June1996.

15. C. Fellbaum,ed.,WordNet: An Electronic Lexical Database, no. ISBN 0-262-06197-X,MIT Press,1998. Softwarecanbedownloadedfrom http://www.cogsci.princeton.edu/ wn/.