Home > Research > Publications & Outputs > SiGBDT: Large-Scale Gradient Boosting Decision ...

Links

Text available via DOI:

View graph of relations

SiGBDT: Large-Scale Gradient Boosting Decision Tree Training via Function Secret Sharing

Research output: Contribution in Book/Report/Proceedings - With ISBN/ISSNConference contribution/Paperpeer-review

Published

Standard

SiGBDT: Large-Scale Gradient Boosting Decision Tree Training via Function Secret Sharing. / Jiang, Yufan; Mei, Fei; Dai, Tianxiang et al.
ASIA CCS '24: Proceedings of the 19th ACM Asia Conference on Computer and Communications Security. New York: ACM, 2024. p. 274-288.

Research output: Contribution in Book/Report/Proceedings - With ISBN/ISSNConference contribution/Paperpeer-review

Harvard

Jiang, Y, Mei, F, Dai, T & Li, Y 2024, SiGBDT: Large-Scale Gradient Boosting Decision Tree Training via Function Secret Sharing. in ASIA CCS '24: Proceedings of the 19th ACM Asia Conference on Computer and Communications Security. ACM, New York, pp. 274-288. https://doi.org/10.1145/3634737.3657024

APA

Jiang, Y., Mei, F., Dai, T., & Li, Y. (2024). SiGBDT: Large-Scale Gradient Boosting Decision Tree Training via Function Secret Sharing. In ASIA CCS '24: Proceedings of the 19th ACM Asia Conference on Computer and Communications Security (pp. 274-288). ACM. https://doi.org/10.1145/3634737.3657024

Vancouver

Jiang Y, Mei F, Dai T, Li Y. SiGBDT: Large-Scale Gradient Boosting Decision Tree Training via Function Secret Sharing. In ASIA CCS '24: Proceedings of the 19th ACM Asia Conference on Computer and Communications Security. New York: ACM. 2024. p. 274-288 doi: 10.1145/3634737.3657024

Author

Jiang, Yufan ; Mei, Fei ; Dai, Tianxiang et al. / SiGBDT: Large-Scale Gradient Boosting Decision Tree Training via Function Secret Sharing. ASIA CCS '24: Proceedings of the 19th ACM Asia Conference on Computer and Communications Security. New York : ACM, 2024. pp. 274-288

Bibtex

@inproceedings{dfbd5b28dc394c7c8ee359ff5b3e7269,
title = "SiGBDT: Large-Scale Gradient Boosting Decision Tree Training via Function Secret Sharing",
abstract = "As a well known machine learning model, Gradient Boosting Decision Tree (GBDT) is widely used in many real-world scenes such as online marketing, risk management, fraud detection and recommendation systems. Due to limited data resources, two data owners may collaborate with each other to jointly train a high-quality model. As privacy regulations such as HIPPA and GDPR come into force, Privacy-Preserving Machine Learning (PPML) has drawn increasingly higher attention. Recently, a line of works [3--6] studies function secret sharing (FSS) schemes in the preprocessing model, where the online stage of secure two-party computation (2PC) is significantly improved. While recent privacy-preserving GDBT frameworks mainly focus on improving the performance of a singular module (e.g. secure bucket aggregation), we propose SiGBDT, a globally silent two-party GBDT framework via function secret sharing on a vertically partitioned dataset. During the training process, we apply FSS schemes to construct efficient modular protocols, such as secure bucket aggregation, argmax computation and a node split approach. We run in-depth experiments and discover that SiGBDT completely outperforms state-of-the-art frameworks. The experiment results show that SiGBDT is at least 3.32 X faster in LAN and at least 6.4 X faster in WAN.",
author = "Yufan Jiang and Fei Mei and Tianxiang Dai and Yong Li",
year = "2024",
month = jul,
day = "1",
doi = "10.1145/3634737.3657024",
language = "English",
pages = "274--288",
booktitle = "ASIA CCS '24",
publisher = "ACM",

}

RIS

TY - GEN

T1 - SiGBDT: Large-Scale Gradient Boosting Decision Tree Training via Function Secret Sharing

AU - Jiang, Yufan

AU - Mei, Fei

AU - Dai, Tianxiang

AU - Li, Yong

PY - 2024/7/1

Y1 - 2024/7/1

N2 - As a well known machine learning model, Gradient Boosting Decision Tree (GBDT) is widely used in many real-world scenes such as online marketing, risk management, fraud detection and recommendation systems. Due to limited data resources, two data owners may collaborate with each other to jointly train a high-quality model. As privacy regulations such as HIPPA and GDPR come into force, Privacy-Preserving Machine Learning (PPML) has drawn increasingly higher attention. Recently, a line of works [3--6] studies function secret sharing (FSS) schemes in the preprocessing model, where the online stage of secure two-party computation (2PC) is significantly improved. While recent privacy-preserving GDBT frameworks mainly focus on improving the performance of a singular module (e.g. secure bucket aggregation), we propose SiGBDT, a globally silent two-party GBDT framework via function secret sharing on a vertically partitioned dataset. During the training process, we apply FSS schemes to construct efficient modular protocols, such as secure bucket aggregation, argmax computation and a node split approach. We run in-depth experiments and discover that SiGBDT completely outperforms state-of-the-art frameworks. The experiment results show that SiGBDT is at least 3.32 X faster in LAN and at least 6.4 X faster in WAN.

AB - As a well known machine learning model, Gradient Boosting Decision Tree (GBDT) is widely used in many real-world scenes such as online marketing, risk management, fraud detection and recommendation systems. Due to limited data resources, two data owners may collaborate with each other to jointly train a high-quality model. As privacy regulations such as HIPPA and GDPR come into force, Privacy-Preserving Machine Learning (PPML) has drawn increasingly higher attention. Recently, a line of works [3--6] studies function secret sharing (FSS) schemes in the preprocessing model, where the online stage of secure two-party computation (2PC) is significantly improved. While recent privacy-preserving GDBT frameworks mainly focus on improving the performance of a singular module (e.g. secure bucket aggregation), we propose SiGBDT, a globally silent two-party GBDT framework via function secret sharing on a vertically partitioned dataset. During the training process, we apply FSS schemes to construct efficient modular protocols, such as secure bucket aggregation, argmax computation and a node split approach. We run in-depth experiments and discover that SiGBDT completely outperforms state-of-the-art frameworks. The experiment results show that SiGBDT is at least 3.32 X faster in LAN and at least 6.4 X faster in WAN.

U2 - 10.1145/3634737.3657024

DO - 10.1145/3634737.3657024

M3 - Conference contribution/Paper

SP - 274

EP - 288

BT - ASIA CCS '24

PB - ACM

CY - New York

ER -