tensor contraction for parsimonious deep netstensorlab.cms.caltech.edu/users/...presentation.pdf ·...

12
Tensor Contraction for Parsimonious Deep Nets Jean Kossaifi, Aran Khanna, Zachary C. Lipton, Tommaso Furlanello and Anima Anandkumar AMAZON AI

Upload: others

Post on 11-Oct-2020

6 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Tensor Contraction for Parsimonious Deep Netstensorlab.cms.caltech.edu/users/...presentation.pdf · •Potential space savings •Performance improvement. Tensor Contraction •Tensor

Tensor Contraction for Parsimonious Deep Nets

Jean Kossaifi, Aran Khanna, Zachary C. Lipton, TommasoFurlanello and Anima Anandkumar

AMAZON AI

Page 2: Tensor Contraction for Parsimonious Deep Netstensorlab.cms.caltech.edu/users/...presentation.pdf · •Potential space savings •Performance improvement. Tensor Contraction •Tensor

Traditionalapproaches

• DATA=>CONV=>RELU=>POOL=>Activationtensor• Flatteningloosesinformation• Canweleveragedirectlytheactivationtensorbeforetheflattening?• Potentialspacesavings• Performanceimprovement

Page 3: Tensor Contraction for Parsimonious Deep Netstensorlab.cms.caltech.edu/users/...presentation.pdf · •Potential space savings •Performance improvement. Tensor Contraction •Tensor

TensorContraction

• Tensorcontraction:contractalongeachmodetoobtainalowrank,compacttensor

Page 4: Tensor Contraction for Parsimonious Deep Netstensorlab.cms.caltech.edu/users/...presentation.pdf · •Potential space savings •Performance improvement. Tensor Contraction •Tensor

TensorContractionLayers(TCL)

• Takeactivationtensorasinput

• FeeditthroughaTCL

• Outputalow-rankactivationtensor

Page 5: Tensor Contraction for Parsimonious Deep Netstensorlab.cms.caltech.edu/users/...presentation.pdf · •Potential space savings •Performance improvement. Tensor Contraction •Tensor

TensorContractionLayers(TCL)

• Compactrepresentationlessparameters

• Measuredintermsofpercentageofspacesavingsinthefullyconnectedlayers:

1–𝑛𝑇𝐶𝐿

𝑛𝑜𝑟𝑖𝑔𝑖𝑛𝑎𝑙

• Similarandsometimesbetterperformance

Page 6: Tensor Contraction for Parsimonious Deep Netstensorlab.cms.caltech.edu/users/...presentation.pdf · •Potential space savings •Performance improvement. Tensor Contraction •Tensor

PerformanceoftheTCL

• Trainedend-to-end

• OnImageNetwithVGG:• 65.9%spacesavings• performancedropof0.6%only

• OnImageNetwithAlexNet:• 56.6%spacesavings• Performanceimprovementof0.5%

Page 7: Tensor Contraction for Parsimonious Deep Netstensorlab.cms.caltech.edu/users/...presentation.pdf · •Potential space savings •Performance improvement. Tensor Contraction •Tensor

Low-ranktensorregression

TensorRegressionNetworks,J.Kossaifi,Z.C.Lipton,A.Khanna,T.Furlanello andA.Anandkumar,ArXiv pre-publication

Page 8: Tensor Contraction for Parsimonious Deep Netstensorlab.cms.caltech.edu/users/...presentation.pdf · •Potential space savings •Performance improvement. Tensor Contraction •Tensor

Performanceandrank

Page 9: Tensor Contraction for Parsimonious Deep Netstensorlab.cms.caltech.edu/users/...presentation.pdf · •Potential space savings •Performance improvement. Tensor Contraction •Tensor

PerformanceoftheTRL

TRLrank Performance(in%)Top-1 Top-1 Spacesavings

baseline 77.1 93.4 0

(200,1,1,200) 77.1 93.2 68.2

(150,1,1,150) 76 92.9 76.6

(100,1,100) 74.6 91.7 84.6

(50,1,1,50) 73.6 91 92.4

ResultsonImageNetwithaResNet-101

• 92.4%spacesavings,4%decreaseinTop-1accuracy• 68.2%spacesavings,nodecreaseinTop-1accuracy

Page 10: Tensor Contraction for Parsimonious Deep Netstensorlab.cms.caltech.edu/users/...presentation.pdf · •Potential space savings •Performance improvement. Tensor Contraction •Tensor

Implementation

• MXNet asaDeepLearningframeworkhttp://mxnet.io/

• TensorLy fortensoroperationshttps://tensorly.github.io

• Comingsoon:mxnet backendforTensorLytensoroperationonGPUandCPU

Page 11: Tensor Contraction for Parsimonious Deep Netstensorlab.cms.caltech.edu/users/...presentation.pdf · •Potential space savings •Performance improvement. Tensor Contraction •Tensor

Conclusionandfuturework• TensorcontractionandtensorregressionforDeepNeuralNets• Addasanadditionallayerorreplaceone• Lessparameters• Similarorevenbetterperformance

Futurework• Fastertensoroperations• Exploremoretensoroperations/networksarchitectures

Page 12: Tensor Contraction for Parsimonious Deep Netstensorlab.cms.caltech.edu/users/...presentation.pdf · •Potential space savings •Performance improvement. Tensor Contraction •Tensor

Mahalo!

Anyquestion?

[email protected]@JeanKossaifi