Home > Research > Publications & Outputs > Progressive Channel-Shrinking Network

Electronic data

  • 2304.00280

    Accepted author manuscript, 3.06 MB, PDF document

    Available under license: CC BY: Creative Commons Attribution 4.0 International License

Links

Text available via DOI:

View graph of relations

Progressive Channel-Shrinking Network

Research output: Contribution to Journal/MagazineJournal articlepeer-review

Published

Standard

Progressive Channel-Shrinking Network. / Pan, Jianhong; Yang, Siyuan; Foo, Lin Geng et al.
In: IEEE Transactions on Multimedia, Vol. 26, 01.02.2024, p. 2016-2026.

Research output: Contribution to Journal/MagazineJournal articlepeer-review

Harvard

Pan, J, Yang, S, Foo, LG, Ke, Q, Rahmani, H, Fan, Z & Liu, J 2024, 'Progressive Channel-Shrinking Network', IEEE Transactions on Multimedia, vol. 26, pp. 2016-2026. https://doi.org/10.1109/TMM.2023.3291197

APA

Pan, J., Yang, S., Foo, L. G., Ke, Q., Rahmani, H., Fan, Z., & Liu, J. (2024). Progressive Channel-Shrinking Network. IEEE Transactions on Multimedia, 26, 2016-2026. https://doi.org/10.1109/TMM.2023.3291197

Vancouver

Pan J, Yang S, Foo LG, Ke Q, Rahmani H, Fan Z et al. Progressive Channel-Shrinking Network. IEEE Transactions on Multimedia. 2024 Feb 1;26:2016-2026. Epub 2023 Jun 30. doi: 10.1109/TMM.2023.3291197

Author

Pan, Jianhong ; Yang, Siyuan ; Foo, Lin Geng et al. / Progressive Channel-Shrinking Network. In: IEEE Transactions on Multimedia. 2024 ; Vol. 26. pp. 2016-2026.

Bibtex

@article{27931779427f48e58ee841ca07cb22a3,
title = "Progressive Channel-Shrinking Network",
abstract = "Currently, salience-based channel pruning makes continuous breakthroughs in network compression. In the realization, the salience mechanism is used as a metric of channel salience to guide pruning. Therefore, salience-based channel pruning can dynamically adjust the channel width at run-time, which provides a flexible pruning scheme. However, there are two problems emerging: a gating function is often needed to truncate the specific salience entries to zero, which destabilizes the forward propagation; dynamic architecture brings more cost for indexing in inference which bottlenecks the inference speed. In this paper, we propose a Progressive Channel-Shrinking (PCS) method to compress the selected salience entries at run-time instead of roughly approximating them to zero. We also propose a Running Shrinking Policy to provide a testing-static pruning scheme that can reduce the memory access cost for filter indexing. We evaluate our method on ImageNet and CIFAR10 datasets over two prevalent networks: ResNet and VGG, and demonstrate that our PCS outperforms all baselines and achieves state-of-the-art in terms of compression-performance tradeoff. Moreover, we observe a significant and practical acceleration of inference. The code will be released upon acceptance.",
keywords = "Convolution, Costs, Feature extraction, Generators, Indexing, Network Shrinking, Progressive, Testing, Training",
author = "Jianhong Pan and Siyuan Yang and Foo, {Lin Geng} and Qiuhong Ke and Hossein Rahmani and Zhipeng Fan and Jun Liu",
year = "2024",
month = feb,
day = "1",
doi = "10.1109/TMM.2023.3291197",
language = "English",
volume = "26",
pages = "2016--2026",
journal = "IEEE Transactions on Multimedia",
issn = "1520-9210",
publisher = "Institute of Electrical and Electronics Engineers Inc.",

}

RIS

TY - JOUR

T1 - Progressive Channel-Shrinking Network

AU - Pan, Jianhong

AU - Yang, Siyuan

AU - Foo, Lin Geng

AU - Ke, Qiuhong

AU - Rahmani, Hossein

AU - Fan, Zhipeng

AU - Liu, Jun

PY - 2024/2/1

Y1 - 2024/2/1

N2 - Currently, salience-based channel pruning makes continuous breakthroughs in network compression. In the realization, the salience mechanism is used as a metric of channel salience to guide pruning. Therefore, salience-based channel pruning can dynamically adjust the channel width at run-time, which provides a flexible pruning scheme. However, there are two problems emerging: a gating function is often needed to truncate the specific salience entries to zero, which destabilizes the forward propagation; dynamic architecture brings more cost for indexing in inference which bottlenecks the inference speed. In this paper, we propose a Progressive Channel-Shrinking (PCS) method to compress the selected salience entries at run-time instead of roughly approximating them to zero. We also propose a Running Shrinking Policy to provide a testing-static pruning scheme that can reduce the memory access cost for filter indexing. We evaluate our method on ImageNet and CIFAR10 datasets over two prevalent networks: ResNet and VGG, and demonstrate that our PCS outperforms all baselines and achieves state-of-the-art in terms of compression-performance tradeoff. Moreover, we observe a significant and practical acceleration of inference. The code will be released upon acceptance.

AB - Currently, salience-based channel pruning makes continuous breakthroughs in network compression. In the realization, the salience mechanism is used as a metric of channel salience to guide pruning. Therefore, salience-based channel pruning can dynamically adjust the channel width at run-time, which provides a flexible pruning scheme. However, there are two problems emerging: a gating function is often needed to truncate the specific salience entries to zero, which destabilizes the forward propagation; dynamic architecture brings more cost for indexing in inference which bottlenecks the inference speed. In this paper, we propose a Progressive Channel-Shrinking (PCS) method to compress the selected salience entries at run-time instead of roughly approximating them to zero. We also propose a Running Shrinking Policy to provide a testing-static pruning scheme that can reduce the memory access cost for filter indexing. We evaluate our method on ImageNet and CIFAR10 datasets over two prevalent networks: ResNet and VGG, and demonstrate that our PCS outperforms all baselines and achieves state-of-the-art in terms of compression-performance tradeoff. Moreover, we observe a significant and practical acceleration of inference. The code will be released upon acceptance.

KW - Convolution

KW - Costs

KW - Feature extraction

KW - Generators

KW - Indexing

KW - Network Shrinking

KW - Progressive

KW - Testing

KW - Training

U2 - 10.1109/TMM.2023.3291197

DO - 10.1109/TMM.2023.3291197

M3 - Journal article

VL - 26

SP - 2016

EP - 2026

JO - IEEE Transactions on Multimedia

JF - IEEE Transactions on Multimedia

SN - 1520-9210

ER -