Tensor Contraction for Parsimonious Deep Netstensorlab.cms.caltech.edu/users/...presentation.pdf ·...

12
Tensor Contraction for Parsimonious Deep Nets Jean Kossaifi, Aran Khanna, Zachary C. Lipton, Tommaso Furlanello and Anima Anandkumar AMAZON AI

Transcript of Tensor Contraction for Parsimonious Deep Netstensorlab.cms.caltech.edu/users/...presentation.pdf ·...

Page 1: Tensor Contraction for Parsimonious Deep Netstensorlab.cms.caltech.edu/users/...presentation.pdf · •Potential space savings •Performance improvement. Tensor Contraction •Tensor

Tensor Contraction for Parsimonious Deep Nets

Jean Kossaifi, Aran Khanna, Zachary C. Lipton, TommasoFurlanello and Anima Anandkumar

AMAZON AI

Page 2: Tensor Contraction for Parsimonious Deep Netstensorlab.cms.caltech.edu/users/...presentation.pdf · •Potential space savings •Performance improvement. Tensor Contraction •Tensor

Traditionalapproaches

• DATA=>CONV=>RELU=>POOL=>Activationtensor• Flatteningloosesinformation• Canweleveragedirectlytheactivationtensorbeforetheflattening?• Potentialspacesavings• Performanceimprovement

Page 3: Tensor Contraction for Parsimonious Deep Netstensorlab.cms.caltech.edu/users/...presentation.pdf · •Potential space savings •Performance improvement. Tensor Contraction •Tensor

TensorContraction

• Tensorcontraction:contractalongeachmodetoobtainalowrank,compacttensor

Page 4: Tensor Contraction for Parsimonious Deep Netstensorlab.cms.caltech.edu/users/...presentation.pdf · •Potential space savings •Performance improvement. Tensor Contraction •Tensor

TensorContractionLayers(TCL)

• Takeactivationtensorasinput

• FeeditthroughaTCL

• Outputalow-rankactivationtensor

Page 5: Tensor Contraction for Parsimonious Deep Netstensorlab.cms.caltech.edu/users/...presentation.pdf · •Potential space savings •Performance improvement. Tensor Contraction •Tensor

TensorContractionLayers(TCL)

• Compactrepresentationlessparameters

• Measuredintermsofpercentageofspacesavingsinthefullyconnectedlayers:

1–𝑛𝑇𝐶𝐿

𝑛𝑜𝑟𝑖𝑔𝑖𝑛𝑎𝑙

• Similarandsometimesbetterperformance

Page 6: Tensor Contraction for Parsimonious Deep Netstensorlab.cms.caltech.edu/users/...presentation.pdf · •Potential space savings •Performance improvement. Tensor Contraction •Tensor

PerformanceoftheTCL

• Trainedend-to-end

• OnImageNetwithVGG:• 65.9%spacesavings• performancedropof0.6%only

• OnImageNetwithAlexNet:• 56.6%spacesavings• Performanceimprovementof0.5%

Page 7: Tensor Contraction for Parsimonious Deep Netstensorlab.cms.caltech.edu/users/...presentation.pdf · •Potential space savings •Performance improvement. Tensor Contraction •Tensor

Low-ranktensorregression

TensorRegressionNetworks,J.Kossaifi,Z.C.Lipton,A.Khanna,T.Furlanello andA.Anandkumar,ArXiv pre-publication

Page 8: Tensor Contraction for Parsimonious Deep Netstensorlab.cms.caltech.edu/users/...presentation.pdf · •Potential space savings •Performance improvement. Tensor Contraction •Tensor

Performanceandrank

Page 9: Tensor Contraction for Parsimonious Deep Netstensorlab.cms.caltech.edu/users/...presentation.pdf · •Potential space savings •Performance improvement. Tensor Contraction •Tensor

PerformanceoftheTRL

TRLrank Performance(in%)Top-1 Top-1 Spacesavings

baseline 77.1 93.4 0

(200,1,1,200) 77.1 93.2 68.2

(150,1,1,150) 76 92.9 76.6

(100,1,100) 74.6 91.7 84.6

(50,1,1,50) 73.6 91 92.4

ResultsonImageNetwithaResNet-101

• 92.4%spacesavings,4%decreaseinTop-1accuracy• 68.2%spacesavings,nodecreaseinTop-1accuracy

Page 10: Tensor Contraction for Parsimonious Deep Netstensorlab.cms.caltech.edu/users/...presentation.pdf · •Potential space savings •Performance improvement. Tensor Contraction •Tensor

Implementation

• MXNet asaDeepLearningframeworkhttp://mxnet.io/

• TensorLy fortensoroperationshttps://tensorly.github.io

• Comingsoon:mxnet backendforTensorLytensoroperationonGPUandCPU

Page 11: Tensor Contraction for Parsimonious Deep Netstensorlab.cms.caltech.edu/users/...presentation.pdf · •Potential space savings •Performance improvement. Tensor Contraction •Tensor

Conclusionandfuturework• TensorcontractionandtensorregressionforDeepNeuralNets• Addasanadditionallayerorreplaceone• Lessparameters• Similarorevenbetterperformance

Futurework• Fastertensoroperations• Exploremoretensoroperations/networksarchitectures

Page 12: Tensor Contraction for Parsimonious Deep Netstensorlab.cms.caltech.edu/users/...presentation.pdf · •Potential space savings •Performance improvement. Tensor Contraction •Tensor

Mahalo!

Anyquestion?

[email protected]@JeanKossaifi