Aliaksei Petsiuk

Additive Manufacturing R&D

Synthetic-to-real Composite Semantic Segmentation in Additive Manufacturing


Journal article


Aliaksei L. Petsiuk, Harnoor Singh, Himanshu Dadhwal, Joshua M. Pearce
Journal of Manufacturing and Materials Processing, 2022

Semantic Scholar ArXiv DBLP DOI
Cite

Cite

APA   Click to copy
Petsiuk, A. L., Singh, H., Dadhwal, H., & Pearce, J. M. (2022). Synthetic-to-real Composite Semantic Segmentation in Additive Manufacturing. Journal of Manufacturing and Materials Processing.


Chicago/Turabian   Click to copy
Petsiuk, Aliaksei L., Harnoor Singh, Himanshu Dadhwal, and Joshua M. Pearce. “Synthetic-to-Real Composite Semantic Segmentation in Additive Manufacturing.” Journal of Manufacturing and Materials Processing (2022).


MLA   Click to copy
Petsiuk, Aliaksei L., et al. “Synthetic-to-Real Composite Semantic Segmentation in Additive Manufacturing.” Journal of Manufacturing and Materials Processing, 2022.


BibTeX   Click to copy

@article{aliaksei2022a,
  title = {Synthetic-to-real Composite Semantic Segmentation in Additive Manufacturing},
  year = {2022},
  journal = {Journal of Manufacturing and Materials Processing},
  author = {Petsiuk, Aliaksei L. and Singh, Harnoor and Dadhwal, Himanshu and Pearce, Joshua M.}
}

Abstract

The application of computer vision and machine learning methods for semantic segmentation of the structural elements of 3D-printed products in the field of additive manufacturing (AM) can improve real-time failure analysis systems and potentially reduce the number of defects by providing additional tools for in situ corrections. This work demonstrates the possibilities of using physics-based rendering for labeled image dataset generation, as well as image-to-image style transfer capabilities to improve the accuracy of real image segmentation for AM systems. Multi-class semantic segmentation experiments were carried out based on the U-Net model and the cycle generative adversarial network. The test results demonstrated the capacity of this method to detect such structural elements of 3D-printed parts as a top (last printed) layer, infill, shell, and support. A basis for further segmentation system enhancement by utilizing image-to-image style transfer and domain adaptation technologies was also considered. The results indicate that using style transfer as a precursor to domain adaptation can improve real 3D printing image segmentation in situations where a model trained on synthetic data is the only tool available. The mean intersection over union (mIoU) scores for synthetic test datasets included 94.90% for the entire 3D-printed part, 73.33% for the top layer, 78.93% for the infill, 55.31% for the shell, and 69.45% for supports.