Andrew Rosenberg- Lecture 20: Model Adaptation

download Andrew Rosenberg- Lecture 20: Model Adaptation

of 35

Transcript of Andrew Rosenberg- Lecture 20: Model Adaptation

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    1/35

    Lecture20:ModelAdaptaon

    MachineLearning

    April15,2010

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    2/35

    Today

    AdaptaonofGaussianMixtureModelsMaximumAPosteriori(MAP)MaximumLikelihoodLinearRegression(MLLR)

    Applicaon:SpeakerRecognionUBM-MAP+SVM

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    3/35

    TheProblem

    IhavealiOlebitoflabeleddata,andalotofunlabeleddata.

    Icanmodelthetrainingdatafairlywell.

    ButwealwaysfittrainingdatabeOerthantesngdata.

    CanweusethewealthofunlabeleddatatodobeOer?

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    4/35

    LetsuseaGMM

    GMMstomodellabeleddata. Insimplestform,onemixturecomponentperclass.

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    5/35

    LabeledtrainingofGMM

    MLEesmatorsofparameters

    rthesecanbeusedtoseedEM.

    i =

    tp(i|xt

    )xt

    tp(i|xt)

    =

    xt

    x

    nkt =

    tp(i|xt)N

    = niN

    i =

    xt

    (xt )(xt )T

    nk

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    6/35

    Adapngthemixturestonewdata

    Essenally,letEMstartwithMLEparametersasseeds. ExpandtheavailabledataforEM,proceedunlconvergence

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    7/35

    Adapngthemixturestonewdata

    Essenally,letEMstartwithMLEparametersasseeds. ExpandtheavailabledataforEM,proceedunlconvergence

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    8/35

    ProblemwithEMadaptaon

    TheiniallabeledseedscouldcontributeveryliOletothefinalmodel

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    9/35

    neProblemwithEMadaptaon

    TheiniallabeledseedscouldcontributeveryliOletothefinalmodel

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    10/35

    MAPAdaptaon

    Constrainthecontribuonofunlabeleddata.

    Letthealphatermsdictatehowmuchweighttogivetothenew,unlabeleddatacomparedtotheexingesmates.

    i = i

    u p(i|xu)xuu p(i|xu)

    + (1 i )i

    i =

    i

    up(i|xu)

    U+ (1

    i)i

    i =

    i

    up(i|xu)(xu i)(xu i)

    T

    U

    + (1 i

    )i

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    11/35

    MAPadaptaon

    Themovementoftheparametersisconstrained.

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    12/35

    MLLRadaptaon

    Anotheridea MaximumLikelihoodLinearRegression. Applyanaffinetransformaontothemeans. Dontchangethecovariancematrices

    =W

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    13/35

    MLLRadaptaon

    Anotherviewonadaptaon. Applyanaffinetransformaontothemeans. Dontchangethecovariancematrices

    =W

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    14/35

    MLLRadaptaon

    ThenewmeansaretheMLEofthemeanswiththenewdata.

    i = Wii =

    x

    p(i|x,i, i,i

    )xi

    xp(i|x,i, i,i)

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    15/35

    MLLRadaptaon

    ThenewmeansaretheMLEofthemeanswiththenewdata.

    i = Wii =

    x

    p(i|x,i, i,i

    )xi

    xp(i|x,i, i,i)

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    16/35

    MLLRadaptaon

    ThenewmeansaretheMLEofthemeanswiththenewdata.

    i = Wii =

    xp(i|x,i, i,i)xi

    x p(i|x,

    i

    , i

    ,i

    )Wi =

    xp(i|x,i, i,i)xixp(i|x,i, i,i)

    (1)T

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    17/35

    WhyMLLR?

    Wecanethetransformaonmatricesofmixturecomponents.

    Forexample: Youknowthattheredandgreenclassesaresimilar Assumpon:Theirtransformaonsshouldbesimilar

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    18/35

    WhyMLLR?

    Wecanethetransformaonmatricesofmixturecomponents.

    Forexample: Youknowthattheredandgreenclassesaresimilar Assumpon:Theirtransformaonsshouldbesimilar

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    19/35

    ApplicaonofModelAdaptaon

    SpeakerRecognion. Task:Givenspeechfromaknownsetofspeakers,idenfythespeaker.

    Assumethereistrainingdatafromeachspeaker. Approach:

    Modelagenericspeaker. Idenfyaspeakerbyitsdifferencefromthegenericspeaker

    Measurethisdifferencebyadaptaonparameters

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    20/35

    SpeechRepresentaon

    Extractafeaturerepresentaonofspeech. Samplesevery10ms.

    MFCC16dims

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    21/35

    Similarityofsounds

    MFCC1

    MFCC2 /s/

    /b/

    /o//u/

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    22/35

    UniversalBackgroundModel

    Ifwehadlabeledphoneinformaonthatwouldbegreat.

    Butitsexpensive,andmeconsuming. SojustfitaGMMtotheMFCCrepresentaonofallofthespeechyouhave.

    Generallyallbutoneexample,butwellcomebacktothis.

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    23/35

    MFCCScaOer

    MFCC1

    MFCC2 /s/

    /b/

    /o//u/

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    24/35

    UBMfing

    MFCC1

    MFCC2 /s/

    /b/

    /o//u/

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    25/35

    MAPadaptaon

    Whenwehaveasegmentofspeechtoevaluate,

    GenerateMFCCfeatures.

    UseMAPadaptaonontheUBMGaussianMixtureModel.

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    26/35

    MAPAdaptaon

    MFCC1

    MFCC2 /s/

    /b/

    /o//u/

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    27/35

    MAPAdaptaon

    MFCC1

    MFCC2 /s/

    /b/

    /o//u/

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    28/35

    UBM-MAP

    Claim:Thedifferencesbetweenspeakerscanberepresentedbythemovementofthemixture

    componentsoftheUBM.

    Howdowetrainthismodel?

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    29/35

    UBM-MAPtraining

    Training

    Data

    Heldout

    SpeakerN

    UBM

    Training

    MAP

    Supervector

    Supervector Avectorofadaptedmeansofthegaussianmixturecomponents

    xi =

    0 1 . . . kT

    ti = Speaker ID

    Trainasupervisedmodelwiththese

    labeledvectors.

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    30/35

    UBM-MAPtraining

    Training

    Data

    Heldout

    SpeakerN

    UBM

    Training

    MAP

    Supervector

    xi =

    0 1 . . . kT

    ti = Speaker ID

    Repeatforalltrainingdata

    Mulclass

    SVM

    Training

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    31/35

    UBM-MAPEvaluaon

    TestData

    UBM

    MAP

    Supervector Mulclass

    SVM

    Predicon

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    32/35

    AlternateView

    Doweneedallthis? WhatifwejusttrainanSVMonlabeledMFCCdata?

    TestData

    Mulclass

    SVM

    Predicon

    Labeled

    Training

    Data

    Mulclass

    SVM

    Training

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    33/35

    Results

    UBM-MAP(withsomevariants)isthestate-of-the-artinSpeakerRecognion.

    Currentstateoftheartperformanceisabout97%accuracy(~2.5%EER)withafewminutesof

    speech.

    DirectMFCCmodelingperformsabouthalfaswell~5%EER.

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    34/35

    ModelAdaptaon

    AdaptaonallowsGMMstobeseededwithlabeleddata.

    Incorporaonofunlabeleddatagivesamorerobustmodel.

    Adaptaonprocesscanbeusedtodifferenatemembersofthepopulaon

    UBM-MAP

  • 8/3/2019 Andrew Rosenberg- Lecture 20: Model Adaptation

    35/35

    NextTime

    SpectralClustering