遷移學(xué)習(xí)算法研究課件_第1頁
遷移學(xué)習(xí)算法研究課件_第2頁
遷移學(xué)習(xí)算法研究課件_第3頁
遷移學(xué)習(xí)算法研究課件_第4頁
遷移學(xué)習(xí)算法研究課件_第5頁
已閱讀5頁,還剩82頁未讀 繼續(xù)免費(fèi)閱讀

下載本文檔

版權(quán)說明:本文檔由用戶提供并上傳,收益歸屬內(nèi)容提供方,若內(nèi)容存在侵權(quán),請進(jìn)行舉報或認(rèn)領(lǐng)

文檔簡介

【5A文】遷移學(xué)習(xí)算法研究TrainingDataClassifierUnseenData(…,long,T)good!Whatif…2023/6/102傳統(tǒng)監(jiān)督機(jī)器學(xué)習(xí)(1/2)[fromProf.QiangYang]傳統(tǒng)監(jiān)督機(jī)器學(xué)習(xí)(2/2)2023/6/103傳統(tǒng)監(jiān)督學(xué)習(xí)同源、獨(dú)立同分布兩個基本假設(shè)標(biāo)注足夠多的訓(xùn)練樣本在實際應(yīng)用中通常不能滿足!訓(xùn)練集測試集分類器訓(xùn)練集測試集分類器遷移學(xué)習(xí)2023/6/104實際應(yīng)用學(xué)習(xí)場景HP新聞Lenovo新聞不同源、分布不一致人工標(biāo)記訓(xùn)練樣本,費(fèi)時耗力遷移學(xué)習(xí)運(yùn)用已有的知識對不同但相關(guān)領(lǐng)域問題進(jìn)行求解的一種新的機(jī)器學(xué)習(xí)方法放寬了傳統(tǒng)機(jī)器學(xué)習(xí)的兩個基本假設(shè)遷移學(xué)習(xí)場景(1/4)2023/6/105遷移學(xué)習(xí)場景無處不在遷移知識遷移知識圖像分類HP新聞Lenovo新聞新聞網(wǎng)頁分類遷移學(xué)習(xí)場景(2/4)異構(gòu)特征空間2023/6/106Theappleisthepomaceousfruitoftheappletree,speciesMalusdomesticaintherosefamilyRosaceae...BananaisthecommonnameforatypeoffruitandalsotheherbaceousplantsofthegenusMusawhichproducethiscommonlyeatenfruit...Training:TextFuture:ImagesApplesBananas[fromProf.QiangYang]XinJin,FuzhenZhuang,SinnoJialinPan,ChangyingDu,PingLuo,QingHe:HeterogeneousMulti-taskSemanticFeatureLearningforClassification.CIKM2015:1847-1850.TestTestTrainingTrainingClassifierClassifier72.65%DVDElectronicsElectronics84.60%ElectronicsDrop!遷移學(xué)習(xí)場景(3/4)2023/6/107[fromProf.QiangYang]遷移學(xué)習(xí)場景(4/4)2023/6/108DVDElectronicsBookKitchenClothesVideogameFruitHotelTeaImpractical![fromProf.QiangYang]OutlineConceptLearningforTransferLearningConceptLearningbasedonNon-negativeMatrixTri-factorizationforTransferLearningConceptLearningbasedonProbabilisticLatentSemanticAnalysisforTransferLearningTransferLearningusingAuto-encodersTransferLearningfromMultipleSourceswithAutoencoderRegularizationSupervisedRepresentationLearning:TransferLearningwithDeepAuto-encoders2023/6/109ConceptLearningbasedonNon-negativeMatrixTri-factorizationforTransferLearning2023/6/10ConceptLearningforTransferLearning10Introduction2023/6/10ConceptLearningforTransferLearning11Manytraditionallearningtechniquesworkwellonlyundertheassumption:Trainingandtestdatafollowthesamedistribution

Training(labeled)ClassifierTest(unlabeled)FromdifferentcompaniesEnterpriseNewsClassification:includingtheclasses“ProductAnnouncement”,“Businessscandal”,“Acquisition”,……Productannouncement:HP'sjust-releasedLaserJetProP1100printerandtheLaserJetProM1130andM1210multifunctionprinters,price…performance

...AnnouncementforLenovoThinkPad

ThinkCentre–price$150offLenovoK300desktopusingcouponcode...LenovoThinkPad

ThinkCentre–price$200offLenovoIdeaPadU450plaptopusing....theirperformanceHPnewsLenovonewsDifferentdistributionFail!Motivation(1/3)2023/6/10ConceptLearningforTransferLearning12ExampleAnalysis

Productannouncement:HP'sjust-releasedLaserJetProP1100printerandtheLaserJetProM1130andM1210multifunctionprinters,price…performance

...AnnouncementforLenovoThinkPad

ThinkCentre–price$150offLenovoK300desktopusingcouponcode...LenovoThinkPad

ThinkCentre–price$200offLenovoIdeaPadU450plaptopusing....theirperformanceHPnewsLenovonewsProductwordconceptLaserJet,printer,price,performanceThinkPad,ThinkCentre,price,performanceRelatedProductannouncementdocumentclass:Sharesomecommonwords:announcement,price,performance…indicateMotivation(2/3)2023/6/10ConceptLearningforTransferLearning13ExampleAnalysis:

HPLaserJet,printer,price,performanceetal.LenovoThinkpad,Thinkcentre,price,performanceetal.Thewordsexpressingthesamewordconceptaredomain-dependent

ProductProductannouncementwordconceptindicatesTheassociationbetweenwordconceptsanddocumentclassesisdomain-independent

Motivation(3/3)2023/6/10ConceptLearningforTransferLearning14Furtherobservations:Differentdomainsmayusesamekeywordstoexpressthesameconcept(denotedasidenticalconcept)Differentdomainsmayalsousedifferentkeywordstoexpressthesameconcept(denotedasalikeconcept)Differentdomainsmayalsohavetheirowndistinctconcepts(denotedasdistinctconcept)TheidenticalandalikeconceptsareusedasthesharedconceptsforknowledgetransferWetrytomodelthesethreekindsofconceptssimultaneouslyfortransferlearningtextclassificationPreliminaryKnowledge2023/6/10ConceptLearningforTransferLearning15Basicformulaofmatrixtri-factorization:wheretheinputXistheword-documentco-occurrencematrix

denotesconceptinformation,mayvaryindifferentdomainsFdenotesthedocumentclassificationinformation

indeedistheassociationbetweenwordconceptsanddocumentclasses,mayretainstablecrossdomainsGSPreviousmethod-MTrickinSDM2010(1/2)2023/6/10ConceptLearningforTransferLearning16SketchmapofMTrick

SourcedomainXs

FsGsFtGtTargetdomainXtSKnowledgeTransferConsideringthealikeconcepts MTrick(2/2)OptimizationproblemforMTrick2023/6/10ConceptLearningforTransferLearning17G0isthesupervisioninformationtheassociationSissharedasbridgetotransferknowledgeDualTransferLearning(Longetal.,SDM2012),consideringidenticalandalikeconceptsTriplexTransferLearning(TriTL)(1/5)2023/6/10ConceptLearningforTransferLearning18Furtherdividethewordconceptsintothreekinds:

F1,identicalconcepts;F2,alikeconcepts;F3,distinctconceptsInput:ssourcedomainXr(1≤r≤s)withlabelinformation,ttargetdomainXr(s+1≤r≤s+t)WeproposeTriplexTransferLearningframeworkbasedonmatrixtri-factorization(TriTLforshort)

F1,S1andS2

aresharedasthebridgeforknowledgetransferacrossdomainsThesupervisioninformationisintegratedbyGr(1≤r≤s)insourcedomainsTriTL(2/5)OptimizationProblem

2023/6/10ConceptLearningforTransferLearning19TriTL(3/5)Wedevelopanalternativelyiterativealgorithmtoderivethesolutionandtheoreticallyanalyzeitsconvergence 2023/6/10ConceptLearningforTransferLearning20TriTL(4/5)Classificationontargetdomains When1≤r≤s,Grcontainsthelabelinformation,soweremainitunchangedduringtheiterations

whenxibelongstoclassj,thenGr(i,j)=1,elseGr(i,j)=0Aftertheiteration,weobtaintheoutputGr(s+1≤r≤s+t),thenwecanperformclassificationaccordingtoGr2023/6/10ConceptLearningforTransferLearning21TriTL(5/5)AnalysisofAlgorithmConvergence Accordingtothemethodologyofconvergenceanalysisinthetwoworks[Leeetal.,NIPS’01]and[Dingetal.,KDD’06],thefollowingtheoremholds.Theorem(Convergence):Aftereachroundofcalculatingtheiterativeformulas,theobjectivefunctionintheoptimizationproblemwillconvergemonotonically.2023/6/10ConceptLearningforTransferLearning222023/6/10ConceptLearningforTransferLearning23rec.autosrec.motorcyclesrec.baseballrec.hockeysci.cryptsic.electronicssci.medsci.spacecomp.graphicscomp.sys.ibm.pc.hardwarecomp.sys.mac.hardwarecomp.windows.xtalk.politics.misctalk.politics.gunstalk.politics.mideasttalk.religion.miscrecscicomptalkDataPreparation(1/3)20Newsgroups Fourtopcategories,eachtopcategorycontainsfoursub-categories SentimentClassification,fourdomains:books,dvd,electronics,kitchenRandomlyselecttwodomainsassources,andtherestastargets,then6problemscanbeconstructed

2023/6/10ConceptLearningforTransferLearning24rec.autosrec.motorcyclesrec.baseballrec.hockeysci.cryptsic.electronicssci.medsci.spacerec+sci-baseballcrypySourcedomainautosspaceTargetdomainFortheclassificationproblemwithonesourcedomainandonetargetdomain,wecanconstruct144()

problemsDataPreparation(2/3)Constructclassificationtasks(TraditionalTL)2023/6/10ConceptLearningforTransferLearning25Constructnewtransferlearningproblemsrec.autosrec.motorcyclesrec.baseballrec.hockeysci.cryptsic.electronicssci.medsci.spacerec+sci-baseballcrypyautosspacecomp.graphicscomp.sys.ibm.pc.hardwarecomp.sys.mac.hardwarecomp.windows.xtalk.politics.misctalk.politics.gunstalk.politics.mideasttalk.religion.misccomptalkautosgraphicsMoredistinctconceptsmayexist!DataPreparation(3/3)SourcedomainTargetdomain2023/6/10ConceptLearningforTransferLearning26ComparedAlgorithmsTraditionallearningAlgorithmsSupervisedLearning:LogisticRegression(LR)[Davidetal.,00]SupportVectorMachine(SVM)[Joachims,ICML’99]Semi-supervisedLearning:TSVM[Joachims,ICML’99]TransferlearningMethods:CoCC[Daietal.,KDD’07],DTL[Longetal.,SDM’12]Classificationaccuracyisusedastheevaluationmeasure2023/6/10ConceptLearningforTransferLearning27ExperimentalResults(1/3)SorttheproblemswiththeaccuracyofLRDegreeoftransferdifficultyeasierGenerally,thelowerofaccuracyofLRcanindicatethehardertotransfer,whilethehigheronesindicatetheeasiertotransferharder2023/6/10ConceptLearningforTransferLearning28ExperimentalResults(2/3)ComparisonsamongTriTL,DTL,MTrick,CoCC,TSVM,SVMandLRondatasetrecvs.sci(144problems)TriTLcanperformwelleventheaccuracyofLRislowerthan65%2023/6/10ConceptLearningforTransferLearning29ExperimentalResults(3/3)Resultsonnewtransferlearningproblems,weonlyselecttheproblems,whoseaccuraciesofLRarebetween(50%,55%](Onlyslightlybetterthanrandomclassification,thustheymightbemuchmoredifficult).Weobtain65problemsTriTLalsooutperformsallthebaselinesConclusions2023/6/10ConceptLearningforTransferLearning30Explicitlydefinethreekindsofwordconcepts,i.e.,identicalconcept,alikeconceptanddistinctconceptProposeageneraltransferlearningframeworkbasedonnonnegativematrixtri-factorization,whichsimultaneouslymodelthethreekindsofconcepts(TriTL)Extensiveexperimentsshowtheeffectivenessoftheproposedapproach,especiallywhenthedistinctconceptsmayexistConceptLearningbasedonProbabilisticLatentSemanticAnalysisforTransferLearning2023/6/10ConceptLearningforTransferLearning312023/6/10ConceptLearningforTransferLearning32MotivationProductannouncement:HP'sjust-releasedLaserJetProP1100printerandtheLaserJetProM1130andM1210multifunctionprinters,price…performance

...AnnouncementforLenovoThinkPad

ThinkCentre–price$150offLenovoK300desktopusingcouponcode...LenovoThinkPad

ThinkCentre–price$200offLenovoIdeaPadU450plaptopusing....theirperformanceHPnewsLenovonewsProductwordconceptLaserJet,printer,price,performanceThinkPad,ThinkCentre,price,performanceRelatedProductannouncementdocumentclass:Sharesomecommonwords:announcement,price,performance…indicateRetrospecttheexample

2023/6/10ConceptLearningforTransferLearning33SomenotationsddocumentydocumentclasszwordconceptSomedefinitionse.g.,p(price|Product),p(LaserJet|Product,)wwordrdomaine.g,p(Product|Productannouncement)PreliminaryKnowledge(1/3)2023/6/10ConceptLearningforTransferLearning34PreliminaryKnowledge(2/3)ProductLaserJet,printer,announcement,price,ThinkPad,ThinkCentre,announcement,priceProductannouncementp(w|z,r1)p(w|z,r2)p(z|y)p(w|z,r1)≠p(w|z,r2)E.g.,p(LaserJet|Product,HP)≠p(LaserJet|Product,Lenovo)p(z|y,r1)=p(z|y,r2)E.g.,p(Product|Productannoucement,HP)=p(Product|Productannoucement,Lenovo)Alikeconcept2023/6/10ConceptLearningforTransferLearning35DualPLSA

(D-PLSA)Jointprobabilityoverallvariablesp(w,d)=p(w|z)p(z|y)p(d|y)p(y)GivendatadomainX,theproblemofmaximumloglikelihoodislogp(X;θ)=logΣz

p(Z,X;θ)

θ

includesalltheparametersp(w|z),p(z|y),p(d|y),p(y).Z

denotesallthelatentvariablesPreliminaryKnowledge(3/3)TheproposedtransferlearningalgorithmbasedonD-PLSA,denotedasHIDC2023/6/10ConceptLearningforTransferLearning36Identicalconceptp(w|za)p(za|y)AlikeconceptTheextensionandintensionaredomainindependentp(w|zb,r)p(zb|y)HIDC(1/3)Theextensionisdomaindependent,whiletheintensionisdomainindependent2023/6/10ConceptLearningforTransferLearning37Distinctconceptp(w|zc,r)p(zc|y,r)ThejointprobabilitiesofthesethreegraphicalmodelsHIDC(2/3)Theextensionandintensionarebothdomaindependent2023/6/10ConceptLearningforTransferLearning38Givens+t

datadomainsX={X1,…,Xs,Xs+1,…,Xs+t},withoutlossofgenerality,thefirstsdomainsaresourcedomains,andthelefttdomainsaretargetdomainsConsiderthethreekindsofconcepts:TheLog

likelihoodfunctionislogp(X;θ)=logΣz

p(Z,X;θ)

θ

includesallparametersp(w|za),p(w|zb,r),p(w|zc,r),p(za|y),p(zb|y),p(zc|y,r),p(d|y,r),p(y|r),p(r).HIDC(3/3)2023/6/10ConceptLearningforTransferLearning39UsetheEMalgorithmtoderivethesolutionsEStep:ModelSolution(1/4)2023/6/10ConceptLearningforTransferLearning40M

Step:ModelSolution(2/4)2023/6/10ConceptLearningforTransferLearning41Semi-supervisedEMalgorithm:whenrisfromsourcedomains,thelabeledinformationp(d|y,r)isknownandp(y|r)

canbeinferedp(d|y,r)=1/ny,r,ifdbelongsyindomainr,ny,risthenumberofdocumentsinclassyindomainr,else

p(d|y,c)=0p(y|r)=ny,r/nr

,nr

isthenumberofdocumentsindomainr

whenrisfromsourcedomains,p(d|y,r)andp(y|r)keepunchangedduringtheiterations,whichsupervisetheoptimizingprocessModelSolution(3/4)2023/6/10ConceptLearningforTransferLearning42ClassificationfortargetdomainsAfterweobtainthefinalsolutionsofp(w|za),p(w|zb,r),p(w|zc,r),p(za|y),p(zb|y),p(zc|y,r),p(d|y,r),p(y|r),p(r)Wecancomputetheconditionalprobabilities:

ThenthefinalpredictionisDuringtheiterations,alldomainssharep(w|za),p(za|y),p(zb|y),

whichactasthebridgeforknowledgetransferModelSolution(4/4)2023/6/10ConceptLearningforTransferLearning43BaselinesComparedAlgorithmsSupervisedLearning:LogisticRegression(LG)[Davidetal.,00]SupportVectorMachine(SVM)[Joachims,ICML’99]Semi-supervisedLearning:TSVM[Joachims,ICML’99]TransferLearning:CoCC[Daietal.,KDD’07]CD-PLSA[Zhuangetal.,CIKM’10]DTL[Longetal.,SDM’12]OurMethodsHIDCMeasure:classificationaccuracy2023/6/10ConceptLearningforTransferLearning44Resultsonnewtransferlearningproblems,weselecttheproblems,whoseaccuraciesofLRarehigherthan50%,then334problemsareobtainedExperimentalResults(1/5)2023/6/10ConceptLearningforTransferLearning45Resultsonnewtransferlearningproblems,weselecttheproblems,whoseaccuraciesofLRarehigherthan50%,then334problemsareobtainedExperimentalResults(2/5)2023/6/10ConceptLearningforTransferLearning46ExperimentalResults(3/5)2023/6/10ConceptLearningforTransferLearning47Sourcedomain:S

(rec.autos,

sci.space),Targetdomain:T(rec.sport.hockey,talk.politics.mideast)STSTDistinctconceptSTAlikeconceptExperimentalResults(4/5)2023/6/10ConceptLearningforTransferLearning48ExperimentalResults(5/5)Indeed,theproposedprobabilisticmethodHIDCisalsobetterthanTriTLThismayduetothereasonthatthereismoreclearerprobabilisticexplanationofHIDCp1(z,y)=p2(z,y)orp1(z|y)=p2(z|y)whichisbetter?p(z|y)p(y)2023/6/10ConceptLearningforTransferLearning49[1]FuzhenZhuang,PingLuo,HuiXiong,QingHe,YuhongXiong,ZhongzhiShi:ExploitingAssociationsbetweenWordClustersandDocumentClassesforCross-DomainTextCategorization.SDM2010,pp.13-24.[2]FuzhenZhuang,PingLuo,ZhiyongShen,QingHe,YuhongXiong,ZhongzhiShi,HuiXiong:CollaborativeDual-PLSA:miningdistinctionandcommonalityacrossmultipledomainsfortextclassification.CIKM2010,pp.359-368.[3]FuzhenZhuang,PingLuo,ZhiyongShen,QingHe,YuhongXiong,ZhongzhiShi,HuiXiong:MiningDistinctionandCommonalityacrossMultipleDomainsUsingGenerativeModelforTextClassification.IEEETrans.Knowl.DataEng.24(11):2025-2039(2012).[3]FuzhenZhuang,PingLuo,ChangyingDu,QingHe,ZhongzhiShi:Triplextransferlearning:exploitingbothsharedanddistinctconceptsfortextclassification.WSDM2013,pp.425-434.[4]FuzhenZhuang,PingLuo,PeifengYin,QingHe,ZhongzhiShi.:ConceptLearningforCross-domainTextClassification:aGeneralProbabilisticFramework.IJCAI2013,pp.1960-1966.ReferencesOutlineConceptLearningforTransferLearningConceptLearningbasedonNon-negativeMatrixTri-factorizationforTransferLearningConceptLearningbasedonProbabilisticLatentSemanticAnalysisforTransferLearningTransferLearningusingAuto-encodersTransferLearningfromMultipleSourceswithAutoencoderRegularizationSupervisedRepresentationLearning:TransferLearningwithDeepAuto-encoders2023/6/1050TransferLearningfromMultipleSourceswithAutoencoderRegularization2023/6/10TransferLearningUsingAuto-encoders512023/6/1052Motivation(1/2)TransferlearningbasedonoriginalfeaturespacemayfailtoachievehighperformanceonTargetdomaindataWeconsidertheautoencodertechniquetocollaborativelyfindanewrepresentationofbothsourceandtargetdomaindataElectronicsVideoGames

Compact;easytooperate;verygoodpicture,excited

aboutthequality;lookssharp!Averygood

game!Itisactionpacked

andfullofexcitement.Iamverymuchhooked

onthisgame.52TransferLearningUsingAuto-encodersPreviousmethodsoftentransferfromonesourcedomaintoonetargetdomainWeconsidertheconsensusregularizedframeworkforlearningfrommultiplesourcedomainsDVDBookKitchenElectronicsWeproposeatransferlearningframeworkofconsensusregularizationautoencoderstolearnfrommultiplesourcesMotivation(2/2)2023/6/10TransferLearningUsingAuto-encoders53AutoencoderNeuralNetworkMinimizingthereconstructionerrortoderivethesolution:whereh,garenonlinearactivationfunction,e.g.,Sigmoidfunction,forencodinganddecoding2023/6/10TransferLearningUsingAuto-encoders54ConsensusMeasure-(1/3)Example:three-classclassificationproblem,threeclassifierspredictinstancesf1f2f3f1f2f3x1111x2333x3222x4231x5313x61232023/6/10TransferLearningUsingAuto-encoders55ConstraintSource1:D1Source2:D2Source3:D3ConsensusMeasure-(2/3)Example:three-classclassificationproblem,predictiononinstancex2023/6/10TransferLearningUsingAuto-encoders56Minimalentropy,MaximalConsensusMaximalentropy,MinimalConsensusEntropybasedConsensusMeasure(Luoetal.,CIKM’08)θiistheparametervectorofclassifieri,CistheclasslabelsetConsensusMeasure-(3/3)Forsimplicity,theconsensusmeasureforbinaryclassificationcanberewrittenasInthiswork,weimposetheconsensusregularizationtoautoencoders,andtrytoimprovethelearningperformancefrommultiplesourcedomainssincetheireffectsonmakingthepredictionconsensusaresimilar.2023/6/10TransferLearningUsingAuto-encoders57SomeNotations

SourcedomainsGivenrsourcedomains:,i.e.,

,.

ThefirstcorrespondingdatamatrixisTargetdomainThecorrespondingdatamatrixis

Thegoalistotrainaclassifier

ftomakeprecisepredictionson.2023/6/10TransferLearningUsingAuto-encoders58FrameworkofCRAThedatafromallsourceandtargetdomainssharethesameencodinganddecodingweightsTheclassifierstrainedfromthenewrepresentationareregularizedtopredictthesameresultsontargetdomaindata2023/6/10TransferLearningUsingAuto-encoders59OptimizationProblemofCRATheoptimizationproblem:ReconstructionError2023/6/10TransferLearningUsingAuto-encoders60OptimizationProblemofCRATheoptimizationproblem:ConsensusRegularization2023/6/10TransferLearningUsingAuto-encoders61OptimizationProblemofCRATheoptimizationproblem:ThetotallossofsourceclassifiersoverthecorrespondingsourcedomaindatawiththehiddenrepresentationWeighdecayterm2023/6/10TransferLearningUsingAuto-encoders62TheSolutionofCRAWeusethegradientdescentmethodtoderivethesolutionofallparameters?isthelearningrate.ThetimecomplexityisO(rnmk)Theoutput:theencodinganddecodingparameters,andsourceclassifierswithlatentrepresentation.2023/6/10TransferLearningUsingAuto-encoders63TargetClassifierConstructionTwoScheme:Trainthesourceclassifiersbasedonandcombinethemas,whereCombineallthesourcedomaindataasZSandtrainaunifiedclassifierusinganysupervisedlearningalgorithms,e.g.,SVM,LogisticRegression(LR).ThetwoaccuraciesaredenotedasCRAvandCRAu,respectively2023/6/10TransferLearningUsingAuto-encoders64DataSets-(1/2)ImageData(fromLuoetal.,CIKM08)(Someexamples)AB

A1A2A3A4B1B2B3B4Threesources:A1B1A2B2A3B3Targetdomain:A4B4Totally,96()3-sourcevs1-targetdomain(3vs1)probleminstancescanbeconstructedfortheexperimentalevaluation2023/6/10TransferLearningUsingAuto-encoders65DataSets-(2/2)SentimentClassification(fromBlitzeretal.,ACL07)Four3-sourcevs1-targetdomainclassificationproblemsareconstructedDVDBookKitchenElectronicsTheaccuracyontargetdomaindataisusedastheevaluationmeasureBothSVMandLRareusedtotrainclassifiersonthenewrepresentation2023/6/10TransferLearningUsingAuto-encoders66AllComparedAlgorithmsBaselinesSupervisedlearningonoriginalfeatures:SVM

[Joachims,ICML’99],LogisticRegression(LR)[Davidetal.,00]Embeddingmethodbasedonautoencoders(EAER)[Yuetal.,ECML’13]MarginalizedStackedDenoisingAutoencoders

(mSDA)[Chenetal.,ICML’12]TransferComponentAnalysis(TCA)[Panetal.,TNN’11]Transferlearningfrommultiplesources(CCR3)(Luoetal.,CIKM’08)Ourmethod:CRAvandCRAuForthemethodswhichcannothandlemultiplesources,wetraintheclassifiersfromeachsourcedomainandmergeddataofallsources(r+1accuracies).Finally,maximal,meanandminimalvaluesarereported.2023/6/10TransferLearningUsingAuto-encoders67TransferLearningwithMultipleSourcesviaConsensusRegularizationAutoencodersFuzhenZhuang,XiaohuCheng,SinnoJialinPan,WenchaoYu,QingHe,andZhongzhiShi68ExperimentalResults-(1/2)Resultson96imageclassificationproblemsTransferLearningwithMultipleSourcesviaConsensusRegularizationAutoencodersFuzhenZhuang,XiaohuCheng,SinnoJialinPan,WenchaoYu,QingHe,andZhongzhiShi69ExperimentalResults-(2/2)Resultson4sentimentclassificationproblemsConclusionsThewellknownrepresentationlearningtechniqueautoencoderisconsidered,andweformalizetheautoencodersandconsensusregularizationintoaunifiedoptimizationframeworkExtensivecomparisonexperimentsonimageandsentimentdataareconductedtoshowtheeffectivenessoftheproposealgorithm2023/6/10TransferLearningUsingAuto-encoders70SupervisedRepresentationLearning:TransferLearningwithDeepAuto-encoders2023/6/10TransferLearningUsingAuto-encoders71Autoencoderisanunsupervisedfeaturelearningalgorithm,whichcannoteffectivelymakeuseofthelabelinformationLimitationofBasicAutoencoderContributionofThisWorkWeextendAutoencodertomulti-layerstructure,andincorporatethelabelasonelayerMotivation2023/6/10TransferLearningUsingAuto-encoders72源領(lǐng)域和目標(biāo)領(lǐng)域共享編碼和解碼權(quán)重利用KL距離對隱層空間進(jìn)行約束利用多類回歸模型對類標(biāo)層進(jìn)行約束FrameworkofTLDA(1/5)2023/6/10TransferLearningUsingAuto-encoders73目標(biāo)是最小化重構(gòu)誤差:DeepAutoencoderFrameworkofTLDA(2/5)2023/6/10TransferLearningUsingAuto-encoders74KL距離KL距離衡量的是兩個概率分布的差異情況,計算公式如下:以上KL距離并不滿足傳統(tǒng)距離的對稱性

溫馨提示

  • 1. 本站所有資源如無特殊說明,都需要本地電腦安裝OFFICE2007和PDF閱讀器。圖紙軟件為CAD,CAXA,PROE,UG,SolidWorks等.壓縮文件請下載最新的WinRAR軟件解壓。
  • 2. 本站的文檔不包含任何第三方提供的附件圖紙等,如果需要附件,請聯(lián)系上傳者。文件的所有權(quán)益歸上傳用戶所有。
  • 3. 本站RAR壓縮包中若帶圖紙,網(wǎng)頁內(nèi)容里面會有圖紙預(yù)覽,若沒有圖紙預(yù)覽就沒有圖紙。
  • 4. 未經(jīng)權(quán)益所有人同意不得將文件中的內(nèi)容挪作商業(yè)或盈利用途。
  • 5. 人人文庫網(wǎng)僅提供信息存儲空間,僅對用戶上傳內(nèi)容的表現(xiàn)方式做保護(hù)處理,對用戶上傳分享的文檔內(nèi)容本身不做任何修改或編輯,并不能對任何下載內(nèi)容負(fù)責(zé)。
  • 6. 下載文件中如有侵權(quán)或不適當(dāng)內(nèi)容,請與我們聯(lián)系,我們立即糾正。
  • 7. 本站不保證下載資源的準(zhǔn)確性、安全性和完整性, 同時也不承擔(dān)用戶因使用這些下載資源對自己和他人造成任何形式的傷害或損失。

最新文檔

評論

0/150

提交評論