Home > Research > Publications & Outputs > Optimization for Interval Type-2 Polynomial Fuz...

Links

Text available via DOI:

View graph of relations

Optimization for Interval Type-2 Polynomial Fuzzy Systems: A Deep Reinforcement Learning Approach

Research output: Contribution to Journal/MagazineJournal articlepeer-review

E-pub ahead of print

Standard

Optimization for Interval Type-2 Polynomial Fuzzy Systems: A Deep Reinforcement Learning Approach. / Xiao, Bo; Lam, Hak-Keung; Xuan, Chengbin et al.
In: IEEE Transactions on Artificial Intelligence, 07.07.2022.

Research output: Contribution to Journal/MagazineJournal articlepeer-review

Harvard

APA

Xiao, B., Lam, H-K., Xuan, C., Wang, Z., & Yeatman, E. (2022). Optimization for Interval Type-2 Polynomial Fuzzy Systems: A Deep Reinforcement Learning Approach. IEEE Transactions on Artificial Intelligence. Advance online publication. https://doi.org/10.1109/TAI.2022.3187951

Vancouver

Xiao B, Lam H-K, Xuan C, Wang Z, Yeatman E. Optimization for Interval Type-2 Polynomial Fuzzy Systems: A Deep Reinforcement Learning Approach. IEEE Transactions on Artificial Intelligence. 2022 Jul 7. Epub 2022 Jul 7. doi: 10.1109/TAI.2022.3187951

Author

Xiao, Bo ; Lam, Hak-Keung ; Xuan, Chengbin et al. / Optimization for Interval Type-2 Polynomial Fuzzy Systems : A Deep Reinforcement Learning Approach. In: IEEE Transactions on Artificial Intelligence. 2022.

Bibtex

@article{ba736f9757a44711918ec2817c3c2dfe,
title = "Optimization for Interval Type-2 Polynomial Fuzzy Systems: A Deep Reinforcement Learning Approach",
abstract = "It is known that the interval type-2 (IT2) fuzzy controllers are superior compared to their type-1 counterparts in terms of robustness, flexibility, etc. However, how to conduct the type reduction optimally with the consideration of system stability under the fuzzy-model-based (FMB) control framework is still an open problem. To address this issue, we present a new approach through the membership-function-dependent (MFD) and deep reinforcement learning (DRL) approaches. In the proposed approach, the reduction of IT2 membership functions of the fuzzy controller is completing during optimizing the control performance. Another fundamental issue is that the stability conditions must hold subject to different type-reduction methods. It is tedious and impractical to resolve the stability conditions according to different type-reduction methods, which could lead to infinite possibility. It is more practical to guarantee the holding of stability conditions during type-reduction rather than resolving the stability conditions, the MFD approach is proposed with the imperfect premise matching (IPM) concept. Thanks to the unique merit of the MFD approach, the stability conditions according to all the different embedded type-1 membership functions within the footprint of uncertainty (FOU) are guaranteed to be valid. During the control processes, the state transitions associated with properly engineered cost/reward function can be used to approximately calculate the deterministic policy gradient to optimize the acting policy and then to improve the control performance through determining the grade of IT2 membership functions of the fuzzy controller. The detailed simulation example is provided to verify the merits of the proposed approach.",
author = "Bo Xiao and Hak-Keung Lam and Chengbin Xuan and Ziwei Wang and Eric Yeatman",
year = "2022",
month = jul,
day = "7",
doi = "10.1109/TAI.2022.3187951",
language = "English",
journal = "IEEE Transactions on Artificial Intelligence",
issn = "2691-4581",
publisher = "IEEE",

}

RIS

TY - JOUR

T1 - Optimization for Interval Type-2 Polynomial Fuzzy Systems

T2 - A Deep Reinforcement Learning Approach

AU - Xiao, Bo

AU - Lam, Hak-Keung

AU - Xuan, Chengbin

AU - Wang, Ziwei

AU - Yeatman, Eric

PY - 2022/7/7

Y1 - 2022/7/7

N2 - It is known that the interval type-2 (IT2) fuzzy controllers are superior compared to their type-1 counterparts in terms of robustness, flexibility, etc. However, how to conduct the type reduction optimally with the consideration of system stability under the fuzzy-model-based (FMB) control framework is still an open problem. To address this issue, we present a new approach through the membership-function-dependent (MFD) and deep reinforcement learning (DRL) approaches. In the proposed approach, the reduction of IT2 membership functions of the fuzzy controller is completing during optimizing the control performance. Another fundamental issue is that the stability conditions must hold subject to different type-reduction methods. It is tedious and impractical to resolve the stability conditions according to different type-reduction methods, which could lead to infinite possibility. It is more practical to guarantee the holding of stability conditions during type-reduction rather than resolving the stability conditions, the MFD approach is proposed with the imperfect premise matching (IPM) concept. Thanks to the unique merit of the MFD approach, the stability conditions according to all the different embedded type-1 membership functions within the footprint of uncertainty (FOU) are guaranteed to be valid. During the control processes, the state transitions associated with properly engineered cost/reward function can be used to approximately calculate the deterministic policy gradient to optimize the acting policy and then to improve the control performance through determining the grade of IT2 membership functions of the fuzzy controller. The detailed simulation example is provided to verify the merits of the proposed approach.

AB - It is known that the interval type-2 (IT2) fuzzy controllers are superior compared to their type-1 counterparts in terms of robustness, flexibility, etc. However, how to conduct the type reduction optimally with the consideration of system stability under the fuzzy-model-based (FMB) control framework is still an open problem. To address this issue, we present a new approach through the membership-function-dependent (MFD) and deep reinforcement learning (DRL) approaches. In the proposed approach, the reduction of IT2 membership functions of the fuzzy controller is completing during optimizing the control performance. Another fundamental issue is that the stability conditions must hold subject to different type-reduction methods. It is tedious and impractical to resolve the stability conditions according to different type-reduction methods, which could lead to infinite possibility. It is more practical to guarantee the holding of stability conditions during type-reduction rather than resolving the stability conditions, the MFD approach is proposed with the imperfect premise matching (IPM) concept. Thanks to the unique merit of the MFD approach, the stability conditions according to all the different embedded type-1 membership functions within the footprint of uncertainty (FOU) are guaranteed to be valid. During the control processes, the state transitions associated with properly engineered cost/reward function can be used to approximately calculate the deterministic policy gradient to optimize the acting policy and then to improve the control performance through determining the grade of IT2 membership functions of the fuzzy controller. The detailed simulation example is provided to verify the merits of the proposed approach.

U2 - 10.1109/TAI.2022.3187951

DO - 10.1109/TAI.2022.3187951

M3 - Journal article

JO - IEEE Transactions on Artificial Intelligence

JF - IEEE Transactions on Artificial Intelligence

SN - 2691-4581

ER -