Deep Neural Networks Pruning via the Structured Perspective Regularization
From MaRDI portal
Publication:6148352
Abstract: In Machine Learning, Artificial Neural Networks (ANNs) are a very powerful tool, broadly used in many applications. Often, the selected (deep) architectures include many layers, and therefore a large amount of parameters, which makes training, storage and inference expensive. This motivated a stream of research about compressing the original networks into smaller ones without excessively sacrificing performances. Among the many proposed compression approaches, one of the most popular is emph{pruning}, whereby entire elements of the ANN (links, nodes, channels, ldots) and the corresponding weights are deleted. Since the nature of the problem is inherently combinatorial (what elements to prune and what not), we propose a new pruning method based on Operational Research tools. We start from a natural Mixed-Integer-Programming model for the problem, and we use the Perspective Reformulation technique to strengthen its continuous relaxation. Projecting away the indicator variables from this reformulation yields a new regularization term, which we call the Structured Perspective Regularization, that leads to structured pruning of the initial architecture. We test our method on some ResNet architectures applied to CIFAR-10, CIFAR-100 and ImageNet datasets, obtaining competitive performances w.r.t.~the state of the art for structured pruning.
Recommendations
Cites work
- scientific article; zbMATH DE number 5251637 (Why is no real title available?)
- Approximated perspective relaxations: a project and lift approach
- Deep neural networks and mixed integer linear optimization
- Fast convex pruning of deep neural networks
- Lossless compression of deep neural networks
- OAMIP: optimizing ANN architectures using mixed-integer programming
- On mathematical programming with indicator constraints
- Perspective cuts for a class of convex 0-1 mixed integer programs
- Projected perspective reformulations with applications in design problems
- Regularization and Variable Selection Via the Elastic Net
- Robust Estimation of a Location Parameter
- Sensitivity-informed provable pruning of neural networks
- The adaptive BerHu penalty in robust regression
This page was built for publication: Deep Neural Networks Pruning via the Structured Perspective Regularization
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6148352)