Home > Research > Publications & Outputs > Global and Local Texture Randomization for Synt...

Links

Text available via DOI:

View graph of relations

Global and Local Texture Randomization for Synthetic-to-Real Semantic Segmentation

Research output: Contribution to Journal/MagazineJournal articlepeer-review

Published

Standard

Global and Local Texture Randomization for Synthetic-to-Real Semantic Segmentation. / Peng, Duo; Lei, Yinjie; Liu, Lingqiao et al.
In: IEEE Transactions on Image Processing, Vol. 30, 31.12.2021, p. 6594-6608.

Research output: Contribution to Journal/MagazineJournal articlepeer-review

Harvard

Peng, D, Lei, Y, Liu, L, Zhang, P & Liu, J 2021, 'Global and Local Texture Randomization for Synthetic-to-Real Semantic Segmentation', IEEE Transactions on Image Processing, vol. 30, pp. 6594-6608. https://doi.org/10.1109/TIP.2021.3096334

APA

Peng, D., Lei, Y., Liu, L., Zhang, P., & Liu, J. (2021). Global and Local Texture Randomization for Synthetic-to-Real Semantic Segmentation. IEEE Transactions on Image Processing, 30, 6594-6608. https://doi.org/10.1109/TIP.2021.3096334

Vancouver

Peng D, Lei Y, Liu L, Zhang P, Liu J. Global and Local Texture Randomization for Synthetic-to-Real Semantic Segmentation. IEEE Transactions on Image Processing. 2021 Dec 31;30:6594-6608. Epub 2021 Jul 16. doi: 10.1109/TIP.2021.3096334

Author

Peng, Duo ; Lei, Yinjie ; Liu, Lingqiao et al. / Global and Local Texture Randomization for Synthetic-to-Real Semantic Segmentation. In: IEEE Transactions on Image Processing. 2021 ; Vol. 30. pp. 6594-6608.

Bibtex

@article{e2642a32efbe451282be9e55f69859e9,
title = "Global and Local Texture Randomization for Synthetic-to-Real Semantic Segmentation",
abstract = "Semantic segmentation is a crucial image understanding task, where each pixel of image is categorized into a corresponding label. Since the pixel-wise labeling for ground-truth is tedious and labor intensive, in practical applications, many works exploit the synthetic images to train the model for real-word image semantic segmentation, i.e., Synthetic-to-Real Semantic Segmentation (SRSS). However, Deep Convolutional Neural Networks (CNNs) trained on the source synthetic data may not generalize well to the target real-world data. To address this problem, there has been rapidly growing interest in Domain Adaption technique to mitigate the domain mismatch between the synthetic and real-world images. Besides, Domain Generalization technique is another solution to handle SRSS. In contrast to Domain Adaption, Domain Generalization seeks to address SRSS without accessing any data of the target domain during training. In this work, we propose two simple yet effective texture randomization mechanisms, Global Texture Randomization (GTR) and Local Texture Randomization (LTR), for Domain Generalization based SRSS. GTR is proposed to randomize the texture of source images into diverse unreal texture styles. It aims to alleviate the reliance of the network on texture while promoting the learning of the domain-invariant cues. In addition, we find the texture difference is not always occurred in entire image and may only appear in some local areas. Therefore, we further propose a LTR mechanism to generate diverse local regions for partially stylizing the source images. Finally, we implement a regularization of Consistency between GTR and LTR (CGL) aiming to harmonize the two proposed mechanisms during training. Extensive experiments on five publicly available datasets (i.e., GTA5, SYNTHIA, Cityscapes, BDDS and Mapillary) with various SRSS settings (i.e., GTA5/SYNTHIA to Cityscapes/BDDS/Mapillary) demonstrate that the proposed method is superior to the state-of-the-art methods for domain generalization based SRSS.",
author = "Duo Peng and Yinjie Lei and Lingqiao Liu and Pingping Zhang and Jun Liu",
year = "2021",
month = dec,
day = "31",
doi = "10.1109/TIP.2021.3096334",
language = "English",
volume = "30",
pages = "6594--6608",
journal = "IEEE Transactions on Image Processing",
issn = "1057-7149",
publisher = "Institute of Electrical and Electronics Engineers Inc.",

}

RIS

TY - JOUR

T1 - Global and Local Texture Randomization for Synthetic-to-Real Semantic Segmentation

AU - Peng, Duo

AU - Lei, Yinjie

AU - Liu, Lingqiao

AU - Zhang, Pingping

AU - Liu, Jun

PY - 2021/12/31

Y1 - 2021/12/31

N2 - Semantic segmentation is a crucial image understanding task, where each pixel of image is categorized into a corresponding label. Since the pixel-wise labeling for ground-truth is tedious and labor intensive, in practical applications, many works exploit the synthetic images to train the model for real-word image semantic segmentation, i.e., Synthetic-to-Real Semantic Segmentation (SRSS). However, Deep Convolutional Neural Networks (CNNs) trained on the source synthetic data may not generalize well to the target real-world data. To address this problem, there has been rapidly growing interest in Domain Adaption technique to mitigate the domain mismatch between the synthetic and real-world images. Besides, Domain Generalization technique is another solution to handle SRSS. In contrast to Domain Adaption, Domain Generalization seeks to address SRSS without accessing any data of the target domain during training. In this work, we propose two simple yet effective texture randomization mechanisms, Global Texture Randomization (GTR) and Local Texture Randomization (LTR), for Domain Generalization based SRSS. GTR is proposed to randomize the texture of source images into diverse unreal texture styles. It aims to alleviate the reliance of the network on texture while promoting the learning of the domain-invariant cues. In addition, we find the texture difference is not always occurred in entire image and may only appear in some local areas. Therefore, we further propose a LTR mechanism to generate diverse local regions for partially stylizing the source images. Finally, we implement a regularization of Consistency between GTR and LTR (CGL) aiming to harmonize the two proposed mechanisms during training. Extensive experiments on five publicly available datasets (i.e., GTA5, SYNTHIA, Cityscapes, BDDS and Mapillary) with various SRSS settings (i.e., GTA5/SYNTHIA to Cityscapes/BDDS/Mapillary) demonstrate that the proposed method is superior to the state-of-the-art methods for domain generalization based SRSS.

AB - Semantic segmentation is a crucial image understanding task, where each pixel of image is categorized into a corresponding label. Since the pixel-wise labeling for ground-truth is tedious and labor intensive, in practical applications, many works exploit the synthetic images to train the model for real-word image semantic segmentation, i.e., Synthetic-to-Real Semantic Segmentation (SRSS). However, Deep Convolutional Neural Networks (CNNs) trained on the source synthetic data may not generalize well to the target real-world data. To address this problem, there has been rapidly growing interest in Domain Adaption technique to mitigate the domain mismatch between the synthetic and real-world images. Besides, Domain Generalization technique is another solution to handle SRSS. In contrast to Domain Adaption, Domain Generalization seeks to address SRSS without accessing any data of the target domain during training. In this work, we propose two simple yet effective texture randomization mechanisms, Global Texture Randomization (GTR) and Local Texture Randomization (LTR), for Domain Generalization based SRSS. GTR is proposed to randomize the texture of source images into diverse unreal texture styles. It aims to alleviate the reliance of the network on texture while promoting the learning of the domain-invariant cues. In addition, we find the texture difference is not always occurred in entire image and may only appear in some local areas. Therefore, we further propose a LTR mechanism to generate diverse local regions for partially stylizing the source images. Finally, we implement a regularization of Consistency between GTR and LTR (CGL) aiming to harmonize the two proposed mechanisms during training. Extensive experiments on five publicly available datasets (i.e., GTA5, SYNTHIA, Cityscapes, BDDS and Mapillary) with various SRSS settings (i.e., GTA5/SYNTHIA to Cityscapes/BDDS/Mapillary) demonstrate that the proposed method is superior to the state-of-the-art methods for domain generalization based SRSS.

U2 - 10.1109/TIP.2021.3096334

DO - 10.1109/TIP.2021.3096334

M3 - Journal article

VL - 30

SP - 6594

EP - 6608

JO - IEEE Transactions on Image Processing

JF - IEEE Transactions on Image Processing

SN - 1057-7149

ER -