Nothing Special   »   [go: up one dir, main page]

Next Article in Journal
Design and Feasibility Verification of Novel AC/DC Hybrid Microgrid Structures
Previous Article in Journal
Hyperspectral Characteristics and SPAD Estimation of Wheat Leaves under CO2 Microleakage Stress
Previous Article in Special Issue
Railway Catenary Condition Monitoring: A Systematic Mapping of Recent Research
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Effect of Annotation Quality on Wear Semantic Segmentation by CNN

1
Digital Production, AImotion Bavaria, Technische Hochschule Ingolstadt, 85049 Ingolstadt, Germany
2
Department of Physics, Faculty of Science, Menoufia University, Menoufia 32511, Egypt
3
Institute of Optical and Electronic Materials, Hamburg University of Technology, 21073 Hamburg, Germany
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(15), 4777; https://doi.org/10.3390/s24154777
Submission received: 13 June 2024 / Revised: 15 July 2024 / Accepted: 16 July 2024 / Published: 23 July 2024
(This article belongs to the Special Issue Sensors for Non-Destructive Testing and Structural Health Monitoring)
Figure 1
<p>(<b>a</b>) Important geometric parameters of an end mill, such as relief face, end angle on the axial rake, rake face, axial relief face, and helix angle. (<b>b</b>) Illustration of light reflection on a TiN-coated end mill when illuminated by standard direct diffuse lighting. Notably, the most intense reflection is observed along the cutter’s edges, while shadowing is evident within the inner rake space. (<b>c</b>) Image captured by the IAS.</p> ">
Figure 2
<p>Schematic representation of the measurement setup for capturing high-quality images of end mills. The acquisition system consists of (<b>1</b>) a hemisphere with barium sulfate coating, (<b>2</b>) 12 LEDs located at the edge of the hemisphere, (<b>3</b>) a tool to be examined, which is held by (<b>4</b>) a three-jaw chuck, (<b>5</b>) represents the rotating plate for a 360° recording, (<b>6</b>) is a camera with an interface (<b>7</b>) connected to the computer.</p> ">
Figure 3
<p>Positive annotation examples. The wear classification includes two primary categories: “yellow” represents typical wear and “green” denotes abnormal wear and the two additional categories: “red” for the background and “black” for the tool.</p> ">
Figure 4
<p>Negative examples of annotations. Each sub-figure highlights a distinct type of incorrect annotation. The classifications include: “yellow” for normal wear, “green” for abnormal wear, “red” for the background, and “black” for the tool itself. (<b>a</b>) shows mislabeling abnormal wear as normal wear, (<b>b</b>) incorrect annotations mistakenly marking the background as a part of the tool surface, (<b>c</b>) impurities have been labeled as abnormal wear, (<b>d</b>) material removal due to chip residue at chipping space has been marked as normal wear.</p> ">
Figure 5
<p>CNN architecture for normal and abnormal wear segmentation. Encoding blocks are colored in blue. The bottleneck layer is colored in brown. Decoding blocks are colored in green.</p> ">
Figure 6
<p>Illustrative images captured via the acquisition system: (<b>a</b>) TiCN-coated endmill and (<b>b</b>) TiN-coated endmill.</p> ">
Figure 7
<p>Masks of endmill wear annotations for comparison of a TiCN-coated endmill dataset. The annotations were performed by three annotators: Annotator 1, Annotator 2, and Annotator 3. The dataset includes four classes: normal wear in green, abnormal wear in yellow, background in red, and tool in black. The critical annotations have been marked in red.</p> ">
Figure 8
<p>Masks of Endmill wear annotations for comparison of a TiN-coated endmill dataset. The annotations were performed by three annotators: Annotator 1, Annotator 2, and Annotator 3. The dataset includes four classes: normal wear in green, abnormal wear in yellow, background in red, and tool in black. The critical annotations have been marked in red.</p> ">
Figure 9
<p>mIoU results of various models for classes of interest: (<b>a</b>) normal wear and (<b>b</b>) abnormal wear. These models were trained using the same dataset but labeled by different annotators. The LR was set to 0.001 and 0.0001, and hyperparameters such as BS and LR varied, as detailed in <a href="#sensors-24-04777-t0A1" class="html-table">Table A1</a>. The dataset originates from a TiN-coated end mill. The standard deviation is depicted to illustrate the performance variation among annotators 1, 2, and 3.</p> ">
Figure 9 Cont.
<p>mIoU results of various models for classes of interest: (<b>a</b>) normal wear and (<b>b</b>) abnormal wear. These models were trained using the same dataset but labeled by different annotators. The LR was set to 0.001 and 0.0001, and hyperparameters such as BS and LR varied, as detailed in <a href="#sensors-24-04777-t0A1" class="html-table">Table A1</a>. The dataset originates from a TiN-coated end mill. The standard deviation is depicted to illustrate the performance variation among annotators 1, 2, and 3.</p> ">
Figure 10
<p>mIoU results of various models for classes of interest: (<b>a</b>) normal wear and (<b>b</b>) abnormal wear. These models were trained using the same dataset but labeled by different annotators. The LR was set to 0.001 and 0.0001, and hyperparameters such as BS and LR varied, as detailed in <a href="#sensors-24-04777-t0A2" class="html-table">Table A2</a>. The dataset originates from a TiCN-coated end mill. The standard deviation is depicted to illustrate the performance variation among annotators 1, 2, and 3.</p> ">
Figure 11
<p>Prediction results and corresponding masks on test images from a TiN-coated milling tool, predicted by the best-performing models of the three annotators, 1, 2 and 3, as detailed and bold in <a href="#sensors-24-04777-t0A1" class="html-table">Table A1</a>. The prediction includes the four classes: normal wear in green, abnormal wear in yellow, background in red, and tool in black. Wrong annotations in the GT Mask are marked red.</p> ">
Figure 12
<p>Prediction results and corresponding masks on test images from a TiN-coated milling tool, predicted by the best-performing models of the three annotators, 1, 2 and 3, as detailed and bold in <a href="#sensors-24-04777-t0A2" class="html-table">Table A2</a>. The prediction includes the four classes: normal wear in green, abnormal wear in yellow, background in red, and tool in black. Critical regions, such as wrong predictions or missed wear recognition, are marked in red.</p> ">
Versions Notes

Abstract

:
In this work, we investigate the impact of annotation quality and domain expertise on the performance of Convolutional Neural Networks (CNNs) for semantic segmentation of wear on titanium nitride (TiN) and titanium carbonitride (TiCN) coated end mills. Using an innovative measurement system and customized CNN architecture, we found that domain expertise significantly affects model performance. Annotator 1 achieved maximum mIoU scores of 0.8153 for abnormal wear and 0.7120 for normal wear on TiN datasets, whereas Annotator 3 with the lowest expertise achieved significantly lower scores. Sensitivity to annotation inconsistencies and model hyperparameters were examined, revealing that models for TiCN datasets showed a higher coefficient of variation (CV) of 16.32% compared to 8.6% for TiN due to the subtle wear characteristics, highlighting the need for optimized annotation policies and high-quality images to improve wear segmentation.

1. Introduction

Deep convolutional neural networks (DCNN) are revolutionizing visual inspection in manufacturing industries. For supervised learning, high-quality annotated datasets are crucial, as the quality of annotations significantly influences model performance [1,2,3,4]. However, many available datasets suffer from improper annotations and instance labeling errors, adversely impacting the performance of learning algorithms [5,6].
Restricting annotations to a single label, similar to ImageNet, can result in inaccuracies because images might encompass multiple objects, which in turn may contain sub-objects or different classes. Furthermore, encouraging users to annotate images that should not be included in the dataset introduces inconsistencies and biases [7,8]. Growing skepticism surrounding datasets derived from user-generated content on the Internet has led to discontinuation or revision of several popular benchmarks. The ongoing use and distribution of these datasets in the form of duplicates or subsets also raise concerns. AI-based quality control in manufacturing faces similar challenges [9]. A survey [10] that interviewed 227 participants from five continents and 20 different industries found that 76% of the participants agree that training data quality and its labeling represent significant challenges in AI projects. To get AI systems off the ground, training data must be extensive and precisely labeled and commented. The use of AI is becoming an increasing priority for companies. Data scientists are under pressure to deliver projects but often need to provide training data of the required size and quality. In the manufacturing sector, the challenges extend beyond data aggregation to dataset selection and labeling, introducing potential biases. In object recognition tasks, even ensuring accurate and consistent placement of bounding boxes around objects is challenging in terms of labeling quality. The authors of [11] observed that sketching a bounding box is both more intricate and time-consuming than annotating classification labels, typically achieved via multiple-choice queries. The author showed that labeling through a suitable process and guideline, in several steps from quality control and training the performance model, can be significantly improved. Therefore, it can be strongly expected that the task of annotating different types of wear presents a higher number of challenges. Tool wear is divided into two main categories: normal wear and abnormal wear. Normal wear occurs as a normal consequence of machining and is influenced by factors such as cutting parameters, tool material and workpiece material [12,13]. However, abnormal wear occurs due to unfavorable reasons such as excessive cutting forces, poor lubrication, material adhesion or complete tool failure. The identification of abnormal wear can be used to take appropriate measures to optimize the geometric parameters of a tool and to ensure the quality of the workpieces.
In order to distinguish between normal and abnormal wear, the neural network must be fed a dataset with consistent annotation to differentiate between the different wear categories for semantic segmentation. In the standard semantic segmentation approach, including our CNN architecture, each pixel is assigned to a single category (e.g., normal wear, abnormal wear, tool).
In this paper, in addition to the deep approach of wear detection, we will also analyze in detail the difficulties that arise in labeling with respect to the normal and abnormal wear of geometrically complex cutting tools by using our CNN systematically.
To the best of our knowledge, there is no previous work that specifically addresses this problem of annotating datasets for wear detection and systematically addresses the effects of different annotations on the performance of the model. Here, we clearly show the challenges in wear detection reduction and at the same time the role of expertise in developing an AI-based wear detection model with respect to the four classes: “Abnormal Wear”, “Normal Wear”, “Tool” and “Background”. This is only feasible due to our innovative detection system, which has already been registered and published at the European Patent Office [14]. This system enables reproducible imaging of the tool and signs of wear, thus reducing effects such as reflections on the surface and fluctuations in the light source. Otherwise, it would not be possible to compare the models from different annotators using conventional imaging methods.
The article is organized as follows: Section 2 discusses the state of the art regarding the relevance of marking quality and CNN-based approaches for AI applications and research. Section 3 introduces the methods and discusses the background and the challenge of tool manufacturers in image-based inspection of cutting tools. The focus is on their optical properties, especially their technical and geometrical parameters. Section 3 also presents the CNN-based methods and the guidelines for the detection of normal and abnormal wear. Section 4 discusses the performance of the different modelers trained by different annotators. Finally, the wear segmentation results of the different annotators with different grades of expertise are reviewed using the CV to identify the main inconsistencies during labeling and to evaluate the performance of the model on two different types of end mill datasets.

2. State of the Art

In this section, we consider the related studies, which cover two main areas. First, we deal with data-based AI. This includes an examination of existing annotations for benchmark datasets. We then look at new AI-based methods for wear detection. We then give a brief overview of the state of the art in artificial neural network techniques such as semantic segmentation and object recognition, where we also focus on the quality of labeling in terms of wear detection.
It is well known that the importance of data quality and preparation is of particular interest in the development of artificial intelligence models. Data-centric AI and the improvement of datasets are not equivalent. A data-centric AI competition focuses on the quality, relevance, and robustness of the data used to train AI models. Data-centric AI approaches focus mostly on systematically improving data to achieve a model with the best performance. Meanwhile, model-centric approaches focus on code or model architecture improvement to enhance performance. Both methods can be balanced well to provide a robust AI solution [15]. There has been a significant effort to improve the performance of CNN models using data-centric approaches. The authors in [16] discuss the relevance of data-centric methods for structural health monitoring and resilience.
In computer vision, most research on dataset validation has traditionally focused on validation against the image database such as [17] and the verification of associated annotations. ImageNet has been the subject of numerous studies recently. ImageNetV2, which is described in detail in [18], experienced a significant drop in performance for numerous classification models, as [18,19] shows. Labeling discrepancies were found in several prominent datasets, leading to performance degradation, especially in DCNN [5]. In a study by [20], bird specialists discovered about 4% of annotation errors in bird images from the CUB-200-2011 [21] and ImageNet datasets. In principle, biases within datasets can lead to reduced model performance. According to Esteva et al. [22], the lack of extensive, high-quality labeled datasets is a major obstacle to the use of supervised deep learning for medical imaging. Taran and colleagues [23] used the Cityscapes dataset, which contains both fine and coarse annotated ground truth data, to investigate the effects of annotation quality on the performance of semantic image segmentation in traffic conditions [24]. The authors investigated two scenarios: first, using the fine ground truth annotations for both training and inference; second, training with the fine annotations followed by inference using the coarse ground truth annotations. For the semantic segmentation model, the research group in [25] used a Pyramid Scene Parsing Network (PSPNet), and they analyzed a subset of the Cityscapes dataset, which included data from three different cities and the following classes: roads, cars, pedestrians, traffic lights, and traffic signs. The dataset intentionally contains 20,000 additional images with coarse annotations to support methods that utilize large amounts of weakly annotated data. The authors used average IoU as a metric. In their results, they found that IoU values for training data using both fine and coarse labeled data were generally higher than those for images with fine ground truth. Based on the results of these comparisons between fine and coarse ground truth annotations, the authors suggested that deep neural networks could be used to generate datasets with coarse ground truth. These could then be modified and used to fine-tune pre-trained models for specific applications.
Currently, to our knowledge, there is no research on the influence of annotation quality and skill level of a worker on model performance for identifying various wear types of cutting tools, despite the high demand for AI-based tool wear inspection systems. Recent research in tool condition monitoring (TCM) has widely been focused on wear detection using various CNN architectures such as [26,27]. Employing the automatic convolutional encoder (CAE), Xuefeng Wu and colleagues adapted a network model specifically for wear detection, refining model parameters through the backpropagation method in tandem with the stochastic gradient descent (SGD) algorithm [26]. In a parallel vein, Thomas Bergs and team employed the Fully Convolutional Network (FCN) and U-Net for the semantic segmentation of individual tool datasets. Their objective was to identify wear on a microscopic scale. For the labeling process, both [26,27] use a standard direct light source to generate high-resolution images from optical microscopy for tool wear labeling and inspection. In contrast, our approach ensures proper illumination of the tool to obtain high-quality images without reflections from the entire tool. While the referenced methods use a microscope and capture images under a single lighting condition, making it difficult to inspect the entire tool and detect wear in different shapes and forms, our approach overcomes these limitations by eliminating illumination artifacts and thereby improving wear detection. The results for a limited tool scene, due to changes such as light exposure, yield a mean IoU coefficient of 0.73 [28]. However, due to the complicated geometry and structure of the tools, capturing suitable images for advanced AI applications remains a challenge [29].
U-Net [30] and Mask R-CNN [31] are two of the leading deep learning frameworks known for their superior performance in instance segmentation. Both achieved remarkable results during the 2018 Kaggle Data Science Bowl [32,33]. Mask R-CNN utilizes multiscale feature maps to capture robust semantic features, with the aim of effectively delineating the boundaries of the cervical nuclei [34]. However, Mask R-CNN demands significant computational power, rendering it less suitable for large-scale classification tasks in industrial inspections.
In the context of semantic segmentation, the U-Net has demonstrated superior performance compared to conventional convolutional networks by using the sliding window approach. U-Net architecture has been used for various applications such as medical and biomedical imaging and remote sensing image analysis [35,36]. This method entails applying a fixed-size window across different segments of the input image to discern features. U-Net’s distinctive “U-shaped” architecture enables it to adeptly capture context and precisely localize it, presenting a more efficient strategy than the sliding-window technique. U-Net’s architecture comprises encoder blocks and decoder blocks. The encoder blocks condense the input image to discern features across varied scales via convolutional layers. Max-pooling operations further diminish the spatial dimensions of these feature maps. In contrast, the decoder blocks are tasked with upsampling the feature maps. Transposed convolutions are used for this upsampling process to recapture spatial information that was previously lost during the downsampling phase. Skip connections are integrated to merge the feature maps from both encoder and decoder blocks, ensuring the network preserves crucial details during the upsampling process. For our analysis, inspired by the U-Net presented in [30], we developed a CNN architecture with three encoding and decoding blocks utilized for images with three channels (RGB) to maintain visual information. The bottleneck layer between the encoding and the decoding blocks represents the latent space that holds the most compressed representation of the training dataset.

3. Materials and Methods

3.1. Structure Parameter-Related Annotation Challenges

In this section, we discuss the challenges associated with annotating images from integral spiral cutters, focusing on their pivotal technical and geometric parameters. We then introduce the innovative Image Acquisition System (IAS), designed to capture no-reflection images of tools, ensuring optimal visibility of wear for annotation purposes.
Integral spiral milling cutters are routinely used in milling processes to machine complicated workpieces made of different materials. The geometry of the tool is crucial for the efficiency and quality of the milling process. Different geometric properties of the cutting edges have a major influence on the overall quality of the milled product. Figure 1a,b show the complexity of describing the wear characteristics of these tools, while taking into account their specific geometric parameters and the resulting reflection behavior on the end mill surface, which makes the annotation process difficult. We solve this problem by capturing images with IAS (Section 3.2), which avoid reflection in imaging the tools, as shown in Figure 1c.
In addition to the optical properties, the complex geometry of these tools makes it difficult to identify wear patterns and even more difficult to differentiate between normal and abnormal wear. Fluctuating light conditions further complicate the detection and differentiation of signs of wear [37]. Consequently, such complex-shaped and optically critical tools provide an ideal way to investigate the role of expertise in providing labeled datasets and its impact on the quality of the annotations when evaluating the performance of different models.

3.2. Acquisition System

To ensure high reproducibility when capturing images of end mills, we utilized the acquisition system depicted in Figure 2. The system employs a centrally-aligned three-jaw clamping chuck (4) to secure the tool (3) precisely at the center of the housing cavity (1). An LED ring, positioned on the inner surface of the hemisphere, disperses the electromagnetic radiation emitted from the emitter uniformly across the object. The housing’s (1) design includes a radially symmetrical segment, further promoting multidirectional light scattering. This layout allows for the electromagnetic radiation to undergo multiple reflections within a U-shaped region, thereby amplifying the diffusion effect produced by the diffuser.
For a complete tool inspection, a motorized rotating plate (5) holds the three-jaw chuck. This arrangement allows for the continuous capture of the end mill from various angles using a single camera. While parts of the structure reflect the electromagnetic radiation, they incorporate openings or transparent sections, enabling the radiation to traverse from the interior to the camera system (6) and interface for data transfer of the images (7). The images were taken with a commercial Nikon camera (Nikon D800E, Nikon Corporation, Tokyo, Japan), with a 105 mm lens. The tool was a four-edged end mill (106.5 mm in length, 40 mm in flute length, 15.4 mm in diameter, 16 mm in shank diameter). For the TiCN dataset, an aperture of f/29 was used with a similar four-edged end mill (93 mm length, 37 mm flute length, 16 mm diameter, 16 mm shank diameter). Both sets of data were taken in equidistant angular steps of 15° from 0° to 360° with white illumination. The field of view of the 105 mm lens on a full-frame sensor is approximately 23.3 degrees diagonal. Smaller apertures such as f/32 and f/29 provide a large depth of field, which is beneficial for capturing detailed images of tool wear. The images have been cropped to focus on the tool and minimize background noise. The Nikon D800E’s pixel pitch of 4.87 microns ensures that each pixel corresponds to approximately 4.87 microns on the tool surface. Given the importance of high-resolution images for effective model training and enhanced wear and damage detection, it is essential to capture high-quality images with great resolution. Consequently, high-resolution images were segmented into 32 discrete smaller images. This not only bolsters model training but also refines the detection of wear and damage on a microscopic scale.

3.3. Annotation Guideline

The annotation of our dataset was carried out by three annotators, each having varying levels of experience in machining. They all commenced their annotation tasks concurrently. Annotation was performed on the full images. We divided the images into 32 small fragments. Prior to the main annotation task, each was required to undergo training on a predefined dataset, ensuring their annotations aligned with established ground truths. They were also provided with examples of both normal and abnormal wear conditions to achieve a consistent annotation baseline. To ensure high-quality annotations, we formulated detailed instructions for the annotations:
  • Definition: Normal wear is characterized by wear without fractures. In contrast, abnormal wear signifies wear with fractures. Both types of wear are considered contiguous surfaces.
  • Positive Examples (refer to Figure 3):
  • Negative Examples (please see Figure 4):
    (a)
    Mislabeling abnormal wear as normal wear (Figure 4a)
    (b)
    Annotations mistakenly marking the background as a part of the tool (Figure 4b)
    (c)
    Incorrect annotations marking impurities as abnormal wear (Figure 4c)
    (d)
    Misidentifying worn regions within the chipping space as normal wear (Figure 4d)
  • Additional Guidelines:
    (a)
    Only label damage present on the cutting edges or phase, excluding the chipping space.
    (b)
    Wear that is ambiguous and cannot be distinctly labeled should be excluded from the dataset.
    (c)
    Instances can appear overlapped, but in effect, they do have finer boundaries that can merge into one another, especially at the cutting edges. Here, careful annotation is required.
Before the annotators started working on a new dataset, they performed a trial run in which they selected 10 difficult image examples from the dataset. The dataset of two tool types with two different coatings contains four instances assigned to four target classes: normal wear conditions, abnormal wear conditions, tools, and background. During the annotation process, it was discovered that certain images in the original dataset did not contain clear recognizable wear patterns. These images were subsequently removed from our dataset.

3.4. Cnn Model

A CNN architecture with three encoding and decoding blocks has been used to train models for normal and abnormal wear detection. The detailed architecture is shown in Figure 5. The CNN architecture consists of three encoding and decoding blocks utilized for images with three channels (RGB) to maintain visual information. The bottleneck layer between the encoding and the decoding blocks represents the latent space that holds the most compressed representation of the training dataset. The training parameters have been reduced to 2,140,740 and are listed with other relevant parameters below in Table 1. BS stands for batch size, which refers to the number of training examples utilized in one iteration. DO stands for dropout rate, which is a regularization technique used to prevent overfitting in neural networks by randomly dropping units during training.

3.5. Dataset Characteristics

A total of 24 high-resolution images of each tool were captured in 15° angle increments for this study. Each image was finally split into 32 fragments of pixel size 512 × 512. For this purpose, only the cutting area of the tools was taken into account. Thus, a total of 768 images were generated for training and testing the neural network for each tool. The participating employees came from Linner Werkzeug Schleif Fabrik GmbH (https://herionlinner.com/linner-gmbh-werkzeugfabrik/ (accessed on 10 May 2024)), a company specializing in tool regrinding, with varying work experience ranging from 1 to 20 years. The tools have been used on CNC machines to produce gear racks by CNC finishing at WMH Herion Antriebstechnik GmbH (https://herionlinner.com/antriebstechnik/ (accessed on 10 May 2024)).
The average time taken to annotate an image was 45 min for a whole image, which is approximately 1.5 min per image fragment. Annotation was performed on the full images. The annotators used LabelMe software (v5.0.1) [38] (https://pypi.org/project/labelme/ (accessed on 10 May 2024)) to label sample images for this study. To qualify as proficient annotators, each candidate was required to complete an image annotation training program. This program consisted of three steps: tutorials on how to use the software for annotation using polygons, distinguishing different types of wear, and adhering to the guidelines outlined in Section 3.3.
We compare a carefully annotated dataset with alternative annotations created by individuals from diverse professional backgrounds. The three people did not follow the same annotation instructions. The resulting models must be able to differentiate various wear patterns in complex end mills. The annotation process involves participants with varying skill sets, ranging from novices to experts. We expect that, even when detailed guidelines are provided, annotations produced by experienced professionals will exhibit a higher level of precision and consistency compared to those produced by their less experienced colleagues. The main relevant aspects that can affect CNN model performance by the dataset are:
  • Tool Diversity and Wear Patterns: Our experimental framework leverages two distinct datasets to ensure a comprehensive evaluation of various wear patterns.
    • Dataset 1: encompasses tools coated with Titanium Nitride (TiN).
    • Dataset 2: incorporates tools coated with Titanium Carbonitride (TiCN).
  • Optimizing CNN Models: Images from the datasets were strategically resized to dimensions of 512 × 512 pixels, facilitating compatibility with our CNN model and optimizing computational performance.
  • Data Partitioning: The assembled images are systematically divided into training, validation, and testing segments, following a 8:1:1 distribution. A detailed enumeration of the instances in the dataset is presented in Table 2.
To obtain an estimation of the instances, the number of individual instances labeled as polygons by annotator 1 was calculated. The number of instances of Background and Tool is easy to check here. Since each annotator labels its own dataset to independently train its own model for wear detection, the number of instances and pixel sizes of normal and abnormal wear may vary for each dataset of distinct annotators.
Marking wear phenomena requires a great deal of precision. Considering the time involved in generating and annotating datasets, our efforts were focused on two particular tool coatings: TiCN (Figure 6a) and TiN (Figure 6b). Apart from the differences in coating, they have different wear patterns, wear contamination, and specific applications. Both coatings have their unique strengths: TiCN, prevalent in end mills, is robust and widely used for machining steel and cast iron. On the other hand, TiN is renowned for its wear resistance and low friction coefficient, making it a popular choice for various cutting applications [39].
To expedite model training without compromising on image quality, we segmented the original images. This ensured quick training and preserved critical visual data that could otherwise be lost by compression. For our ablation study, we used the CNN architecture presented in Figure 5. Through an examination of various hyperparameters, such as Learning rate (LR), Batch size (BS), and dropout rate (DO), we determined the optimal settings for the multiple-class segmentation results.

3.6. Annotators

For a comprehensive evaluation of the impact of annotation quality on modeling results, we assigned several annotators with different levels of expertise to annotate each dataset.
  • Annotator 1: with more than two decades of experience in the field, this person embodies the highest level of expertise and experienced insight into this topic.
  • Annotator 2: with 2 years of hands-on experience, this participant represents the middle tier, bridging the gap between novices and veterans.
  • Annotator 3: as a newcomer to the field of machining technology, this participant offered a fresh perspective without deep-rooted biases or ingrained expertise.

3.7. Evaluation Indicators

To evaluate and compare segmentation models, we employ the accuracy metric known as Intersection over Union (IoU). The Jaccard Index is used as a metric to investigate the similarity in pixel-wise matter between Ground Truth (GT) and prediction.
J ( A , B ) = | A B | | A B | = | A B | | A | + | B | | A B |
In this formula, J ( A , B ) represents the Jaccard Index between sets A and B. The numerator | A B | is the size (cardinality) of the intersection of sets A and B, and the denominator | A B | is the size of the union of sets A and B. This metric provides a measure of the overlap or similarity between the two sets, with values ranging from 0 (no overlap) to 1 (complete overlap or similarity).
To evaluate the models on their overall performance, we introduce a weight adjustment of the under-represented classes such as normal and abnormal wear compared to the over-represented classes such as background and damage-free tool surface. The formula to determine the weights for wmIoU are described below:
  • Determine the class frequencies by counting the occurrences of each class in the dataset to obtain N 1 , N 2 , N 3 , and N 4 .
  • Calculate the inverse frequencies for each class as follows:
    1 N 1 , 1 N 2 , 1 N 3 , and 1 N 4 .
  • Normalize the weights by summing all the inverse frequencies and then divide each inverse frequency by this sum to obtain weights w 1 , w 2 , w 3 , and w 4 that add up to 1:
    w i = 1 N i j = 1 4 1 N j
  • Apply the weights to calculate the weighted mean IoU:
    wmIoU = w 1 · I o U 1 + w 2 · I o U 2 + w 3 · I o U 3 + w 4 · I o U 4
Using inverse frequencies, we ensure that underrepresented classes (with a lower frequency N j ) are given more weight in the calculation. This increases the influence of the underrepresented class on the average performance evaluation of the model. In addition to our analysis, we used the CV to assess the relative variability of the mIoU model performance of the three annotators. We use the CV as a standardized metric of dispersion that is particularly useful for comparing the degree of variation in prediction performance of the models that have been trained on different labeled datasets. The formula for calculating the coefficient of variation is as follows:
C V I o U i = σ I o U i j μ I o U i j × 100 %
where σ represents the standard deviation and μ the mean of the segmentation results for the class i of all annotators j = 1, 2, 3.

4. Results and Discussion

In this research paper, we address the following main goals. First, we investigate the challenge of annotating wear of different types to improve our annotation guidelines for wear segmentation in its different forms. Second, we compare the performance of the different models from the different annotators and investigate the impact of annotation quality on the performance of our proposed CNN models. Then we verify the impact of the hyperparameters with respect to both datasets—TiN and TiCN—which have been annotated by different annotators. Here, we investigate which of the hyperparameter combinations are particularly sensitive to varying annotation quality.

4.1. Comparison of Annotation by Different Annotators

A comparative analysis of the annotations revealed interesting patterns. Although the annotations for the categories “tools” and “background” show good agreement, there is a noticeable variability in the labeling of “normal” and “abnormal” wear. This variability can be clearly seen in Figure 7 and Figure 8. The critical annotations have been marked in red.
The reasons for these discrepancies are complex:
  • Ambiguity in wear assessment: in particular, minute wear features on cutting edges, such as on the edge of the TiCN cutter, presented a challenge in definitive categorization, but still shows consistency in annotation (marked green and yellow in Figure 7b).
  • Concentration loss: as can be seen in Figure 7b,c, noticeable wear patterns (marked in red at the top left part of Figure 7d) were occasionally missed. This oversight could be due to diminishing concentration during the annotation process.
Given these findings, it is of crucial importance that we include an additional control level in the annotation process workflow. We propose an additional annotation check aimed at preemptively identifying obvious inconsistencies in the annotation and taking measures at an early stage.

4.2. Performance Comparison of Various CNN Models on Diverse Datasets from Multiple Annotators

In this subsection, we present a detailed evaluation of the performance of our proposed CNN model by varying hyperparameter combinations. The models have been trained on both datasets, i.e., the TiN- and TiCN-coated ones. Each training dataset has been labeled by one of the annotators, resulting in a total of six annotated datasets. We evaluate the adaptability and performance of these wear inspection methods with respect to the labeling quality and identify subtleties that arise from different data, hyperparameters, and different annotations. All of these factors can affect the performance of a model. We will investigate the factors that influence performance and conclude with potential strategies to improve the robustness and generalization of models for wear detection tasks.
To examine the role of labeling in affecting the effect of annotation quality, Table A1 and Table A2 present the mean IoU in TiN and TiCN inference data predicted by models trained on different datasets labeled by annotators 1, 2, and 3. We consider the mIoU of the distinct classes: “Abnormal Wear”, “Normal Wear”, “Tool”, and “Background”, as well as the overall performance wmIoU. The models have been trained with different hyperparameters. The LR has been set to 0.001 and 0.0001. We also varied the DO rate for each layer: 0.0, 0.3, 0.5, and experimented with different BS: 8 and 16. We employed the sparse categorical cross-entropy loss function to train the model for all combinations of parameters.
It can be seen from Table A1 and Table A2 that all annotators achieve remarkable results for the class “Background”, with a high mIoU of 0.99 for this class in both datasets, coated with TiN and TiCN. Taking into account the class “Tool”, the annotators performed better on the TiN dataset (Table A1) compared to the TiCN dataset (Table A2).
While Annotators 1, 2, and 3 achieved a higher mIoU of nearly 1.0 for multiple models (except for Annotator 2) for the class “Tool” on the TiN dataset, the models from annotators predicted on the TiCN-coated dataset achieved for the class “Tool” a maximum mIoU of 0.96 for Annotator 1 (A1MTiCN 3, LR: 0.001, BS: 16, DO: 0) and 0.94 for Annotator 2 (A2MTiCN 2, LR: 0.001, BS: 8, DO: 0.5). The model A3MTiCN 1 (LR: 0.001, BS: 8, DO: 0.3) of Annotator 3 achieved a maximum mIoU of 0.97. This suggests that the complexity of tool features, possibly combined with variations in annotator labeling, affects model performance, even for the class “Tool”.
For research interest, the classes “Normal Wear” and “Abnormal Wear” are mostly relevant since their labeling quality and impact on performance can be dependent on the level of expertise the annotators have. Regarding normal and abnormal wear, the TiN-coated milling tool compared to the TiCN-coated milling tool achieved a significantly higher mIoU value for almost all combinations of hyperparameters, as seen in Table A1. The model A1MTiN 1 (LR: 0.001, BS: 8, DO: 0.3) of Annotator 1 achieved a remarkable mIoU of 0.82 on the TiN-coated dataset for the class “Abnormal Wear”, and for the class “Normal Wear”, an mIoU of 0.71. While model A2MTiN 6 (LR: 0.0001, BS: 8, DO: 0.0) from Annotator 2 achieved similar high performance with an mIoU of 0.81 for abnormal wear but only 0.46 for normal. The best model trained on the Annotator 3 dataset is A3MTiN 1 (LR: 0.001, BS: 8, DO: 0.3), achieving a maximum mIoU of 0.75 for the class “Abnormal Wear” but only a poor mIoU of 0.57 for the class “Normal Wear”.
For comparison, ref. [40] achieved the highest score of 0.55 with LinkNet for flank wear (normal wear), and for the class groove (abnormal wear), achieved the highest score of 0.80 with U-Net. It must be mentioned that these results [40] stem from optical microscopic images that consider only a small region of the tool. In our case, our results come from the entire tool itself.
Considering the TiCN-coated endmill, the overall IoU results are rather poor, as can be seen in Table A2 for abnormal wear and for normal wear classes.
Annotator 1 achieved the best wear segmentation results of mIoU = 0.66 for abnormal wear and 0.59 for normal wear with the hyperparameter combination A1MTiCN 1 (LR: 0.001, BS: 8, DO: 0.3). Annotator 2, in model A2MTiCN 2 (LR: 0.001, BS: 8, DO: 0.5), achieved the best IoU segmentation results of 0.60 for abnormal wear and 0.56 for normal wear. Although Annotator 3’s performance in A3MTiCN 2 (LR: 0.001, BS: 8, DO: 0.5) for abnormal wear segmentation was better than Annotator 2, the segmentation results for normal wear were poorer, with an mIoU of 0.40.
Regarding the hyperparameter tuning, we can observe that models with adjusted DO, in particular those at 0.3 and 0.5, tend to deliver the best performance, suggesting that regularization via dropout could be impacting the model’s ability to generalize from training data.
For generalization, the use of dropout layers as a regularization method, especially with a DO of 0.3, generally seems to improve the wmIoU across all annotators. A higher BS = 16 also appears to result in a slightly lower wmIoU for all models and datasets compared to a smaller BS = 8. In comparison, wmIoU values tend to perform better with LR = 0.001 than with LR = 0.0001.

4.3. Impact of Hyperparameters on Model Sensitivity to Annotation Quality

To consider the differences and sensitivity of the model with regard to possible annotation errors, the figures below present the segmentation results for “Normal Wear” and “Abnormal Wear” classes from the TiN (Figure 9) and and TiCN (Figure 10) datasets, evaluated using the mIoU and the standard deviation between the annotators. Each figure compares the performance of various models, each defined by specific hyperparameters: LR, BS, and DO. The results are evaluated by three different annotators and the variability between their annotations is shown through standard deviation error bars.
The presence of a higher standard deviation in some models suggests that these models are more sensitive to annotation differences. It can be observed that certain hyperparameters can make a model more sensible to incorrect annotations. This sensitivity means that the performance of the model can vary significantly depending on the quality and consistency of the annotations. The mIou results of normal and abnormal wear of the TiN tool (Figure 9a,b) and normal wear (Figure 10b) of the TiCN tool show a similar trend where the models with LR = 0.001 outperform the models with LR = 0.0001. The model with hyperparameters LR = 0.0001, BS = 16, and DO = 0 shows greater performance variability between the annotators in the class, as shown by the larger standard deviation error bars. This trend cannot be observed for the abnormal wear of TiCN in Figure 10b. In general, models with dropout rates of 0.3 and 0.5 tend to perform better. Annotator 1 consistently yields higher mIoU values, but there is obvious variability between the annotators.

4.4. Visual Analysis

The segmentation result of the best-performing model of each annotator is visually shown in Figure 11 for the TiN-coated dataset and in Figure 12 for the TiCN dataset. It can be clearly seen that the reference annotation in the GT in Figure 11b was not correct, and abnormal wear was mistakenly annotated as normal wear, highlighting the relevance of an annotation guideline. For the TiN tool, all models that performed best for the annotators were able to predict wear correctly (Figure 11c,e,g). In contrast for the TiCN tool, in Figure 12c,e,h, it can be seen that abnormal wear was partially predicted as a background. While the acquisition systems reduce reflections, the top of the tool still shows light artifacts, especially in combination with wear. This is because the wear can behave as a scattering source that can appear brighter at a certain angle to the observer, leading to misclassification of wear as a background, as seen in Figure 12c,e,g. This can be improved by increasing the number of training datasets or adjusting the lighting intensity or integrating time of the camera sensor. Furthermore, it can be seen from Figure 12e,f that normal and abnormal wear at the cutting edges remains completely undetected in the A2MTiCN model (marked in red), while the other models of the two annotators detect wear but have difficulties, especially in the subtle transition from normal to abnormal wear. Nevertheless, the results are considered good in quantitative terms.

4.5. Coefficient of Variation Analysis of the Segmentation Results across Annotators, Classes, and Hyperparameter Variations

In this section, we present the results of our analysis aimed at validating the inconsistencies in the labeling process and its impact on model performance, which was performed by three annotators with different levels of expertise. The analysis focuses on four different classes: background, tool, normal wear, and abnormal wear. We used the coefficient of variation of mIoU (Equation (5)) to assess the consistency and reliability of the method, as well as evaluate the variation in the readings of individual classes by Annotator 1, Annotator 2, and Annotator 3. Additionally, we investigated the CV associated with different hyperparameters, including DO, LR, and BS, which affected model performance. Our goal was to identify reference points for improving the annotation process and to understand the influence of annotator expertise and model hyperparameters on the quality of dataset annotations.
The following tables (Table 3 for the TiN-coated dataset and Table 4 for the TiCN-coated dataset) present the CV of the IoU values of different models across the different classes, including the wmIoU. The mIoU CV values provide insights into the relative variability of the IoU values for each category across different models. A higher CV indicates greater dispersion around the mean, suggesting that the performance of the models is less consistent in that category.
For the TiN-coated tool, the highest mean CV values are observed in the class “Normal Wear” at 9.48% followed by “Abnormal Wear” at 8.61%, indicating significant variability in model performance in these categories. In contrast, the “Background” class, with a CV value of 0.07%, shows the least variability, followed by the class “Tool” with only 2.75%, suggesting consistent performance across models. It seems that the novel acquisition system enables the model with all hyperparameter combinations to segment the tool from the background effectively, and annotations seem to be performed well by all annotators.
For the TiCN-coated tool, the background has a small mean CV value of 0.76%, but the mean CV value of 9.97% is high compared to the one for the TiN-coated tool, indicating that certain hyperparameters can be beneficial for enhancing the extraction of the tool pixel-wise from the other classes. The mean CV for “Normal Wear” of 20.24% and for the class “Abnormal Wear” of 18.53% exhibits the highest mean CV values, indicating notable variability.

5. Conclusions

In this study, we presented an approach to compare the annotation quality and consequent wear detection performance of different CNN models, each trained on datasets created by annotators with varying levels of expertise. The images are derived from TiN- and TiCN-coated milling tools. To achieve this, we utilized a new imaging system designed to minimize reflection and produce high-quality images. Additionally, we analyzed the influence of various hyperparameters to generalize the test datasets and discussed the sensitivity of potentially inconsistent annotations.
The hyperparameters of DO of 0.3 and LR of 0.001 showed consistent model performance in terms of wear detection across all annotations. Annotator 1 achieved a maximum mIoU of 0.8153 for abnormal wear and 0.7120 for normal wear on the TiN datasets. Annotator 3’s models delivered an mIoU of 0.7538 for abnormal wear and 0.5679 for normal wear, with Annotator 2’s performance falling in between these values. The TiCN dataset exhibited a similar trend but with significantly poorer results, indicating annotation challenges due to the subtle wear nature of the tool. This was further demonstrated through the coefficient of variation (CV). The TiN tool showed a low mean CV for overall wmIoU performance at 8.6%, while the TiCN dataset performed significantly worse, with a mean CV of 16.32% for wmIoU.
The results demonstrate the complexity of wear annotation challenges. These findings underscore the importance of professional annotation guidelines, high-resolution images, and large datasets encompassing various types of wear. The three annotators illustrated that specific expertise in machining technology is crucial for the labeling process.

6. Patents

The illumination technique used in our research for the wear inspection system is based on the European patent EP1430720, developed by Mühenad Bilal and Christian Mayer. This specific illumination approach has been modified for wear characterization and enables the identification of tiny wear features that cannot be detected with conventional inspection systems.

Author Contributions

Conceptualization, M.B. (Markus Bregulla); methodology, M.B. (Mühenad Bilal); validation, M.B. (Mühenad Bilal), R.P.; formal analysis, M.B. (Mühenad Bilal); investigation, M.B. (Mühenad Bilal); resources, D.G.; data creation, M.B. (Mühenad Bilal) and R.P.; image processing, M.B. (Mühenad Bilal) and R.P.; model training, M.B. (Mühenad Bilal) and R.P.; writing—original draft preparation, M.B. (Mühenad Bilal); writing—review and editing, M.B. (Markus Bregulla), R.P., L.K., M.A.G., and D.G.; visualization, M.B. (Mühenad Bilal); supervision, D.G. and M.B. (Markus Bregulla); project administration, M.B. (Markus Bregulla); funding acquisition, L.K. and M.B. (Markus Bregulla). All authors have read and agreed to the published version of the manuscript.

Funding

The authors gratefully acknowledge the financial support from the Research and Development (R&D) program “FuE Programm Informations- und Kommunikationstechnik Bayern” of the Free State of Bavaria under the funding code IUK578/001 and IUK578/002.

Data Availability Statement

The data supporting the findings of this study are owned by WMH Herion Antriebstechnik GmbH and are not publicly available due to proprietary restrictions. However, data may be available from the authors upon reasonable request and with permission of WMH Herion Antriebstechnik GmbH.

Acknowledgments

Special thanks are extended to the consortium partners for their indispensable collaboration in this research project. Linner GmbH Tool Factory provided invaluable industry insights, with Ludwig Linner the primary responsible contact for the project. Technical guidance and academic supervision were generously offered by Markus Bregulla from the AIMotion Bavaria Department at the Technical University of Ingolstadt (THI). The authors also acknowledge the commitment and support of all team members and contributors who have facilitated the merging of high-quality research with practical application. We would like to thank you for the support provided by the Open Access Publication Fund of the Technische Hochschule Ingolstadt.

Conflicts of Interest

There are no conflicts of interest in connection with the publication of this article.

Abbreviations

The following abbreviations are used in this manuscript:
(CNN)Convolutional Neural Networks
(TiN)Titanium Nitride
(TiCN)Titanium Carbonitride
(CV)Coefficient of Cariation
(DCNN)Deep Convolutional Neural Networks
(IAS)Image Acquisition System
(LR)Learning Rate
(BS)Batch Size
(DO)Dropout Rate
(mIoU)mean Intersection over Union
(wmIoU)weighted mean Intersection over Union

Appendix A

Table A1. Comparison of the segmentation results as mIoU for the four classes: background, tool, abnormal wear, normal wear, and the overall performance wmIoU using the test dataset of a TiN-coated milling cutter labeled by Annotator 1, Annotator 2, and Annotator 3 with variation of the hyperparameters: LR, DO and BS.
Table A1. Comparison of the segmentation results as mIoU for the four classes: background, tool, abnormal wear, normal wear, and the overall performance wmIoU using the test dataset of a TiN-coated milling cutter labeled by Annotator 1, Annotator 2, and Annotator 3 with variation of the hyperparameters: LR, DO and BS.
Annotator 1Background [mIoU]Tool [mIoU]Abnormal Wear [mIoU]Normal Wear [mIoU]wmIoU [mIoU]LRBSDO
A1MTiN 00.99870.98950.75370.61340.64720.00180
A1MTiN 10.99860.99790.81530.71200.73690.00180.3
A1MTiN 20.99800.98900.68660.71200.70670.00180.5
A1MTiN 30.99830.94590.63610.71200.69470.001160
A1MTiN 40.99880.99820.81050.59350.64530.001160.3
A1MTiN 50.99780.99620.51220.70390.65960.001160.5
A1MTiN 60.99860.94100.66860.35970.43350.000180
A1MTiN 70.99780.97240.62430.58750.59700.000180.3
A1MTiN 80.99790.98880.66230.66190.66270.000180.5
A1MTiN 90.99860.98160.71190.47920.53500.0001160
A1MTiN 100.99810.98000.64580.51610.54760.0001160.3
A1MTiN 110.99810.96400.56710.43120.46430.0001160.5
Annotator 2Background [mIoU]Tool [mIoU]Abnormal Wear [mIoU]Normal Wear [mIoU]wmIoU [mIoU]LRBSDO
A2MTiN 00.99580.98450.62880.48530.52010.00180
A2MTiN 10.99640.95180.68700.59330.61480.00180.3
A2MTiN 20.99700.99690.67940.59330.61440.00180.5
A2MTiN 30.99450.64530.58930.59330.59260.001160
A2MTiN 40.99700.96260.67800.59330.61410.001160.3
A2MTiN 50.99710.99690.67800.59330.61410.001160.5
A2MTiN 60.99770.99800.80820.46180.54430.000180
A2MTiN 70.99800.99800.77780.58810.63350.000180.3
A2MTiN 80.99720.93860.60730.57750.58530.000180.5
A2MTiN 90.99790.98180.72340.45040.51570.0001160
A2MTiN 100.99730.96470.59080.46460.49540.0001160.3
A2MTiN 110.99720.97170.53340.39490.42870.0001160.5
Annotator 3Background [mIoU]Tool [mIoU]Abnormal Wear [mIoU]Normal Wear [mIoU]wmIoU [mIoU]LRBSDO
A3MTiN 00.99820.98950.70470.54000.57970.00180
A3MTiN 10.99850.99810.75380.56790.61250.00180.3
A3MTiN 20.99740.99680.64350.56790.58660.00180.5
A3MTiN 30.99830.94700.73470.44780.51630.001160
A3MTiN 40.99830.99760.69540.56810.59900.001160.3
A3MTiN 50.99760.98830.58520.55990.56680.001160.5
A3MTiN 60.99830.98990.65080.37570.44170.000180
A3MTiN 70.99810.98030.64320.49180.52850.000180.3
A3MTiN 80.99780.92990.55340.55990.55920.000180.5
A3MTiN 90.99830.96490.44520.31120.34420.0001160
A3MTiN 100.99800.94670.63810.50370.53630.0001160.3
A3MTiN 110.99770.93020.49660.42070.43970.0001160.5
Table A2. Comparison of the segmentation results as mIoU for the four classes: background, tool, abnormal wear, normal wear, and the overall performance wmIoU using the test dataset of a TiCN-coated milling cutter labeled by Annotator 1, Annotator 2, and Annotator 3 with variation of the hyperparameters: LR, DO and BS.
Table A2. Comparison of the segmentation results as mIoU for the four classes: background, tool, abnormal wear, normal wear, and the overall performance wmIoU using the test dataset of a TiCN-coated milling cutter labeled by Annotator 1, Annotator 2, and Annotator 3 with variation of the hyperparameters: LR, DO and BS.
Annotator 1Background [mIoU]Tool [mIoU]Abnormal Wear [mIoU]Normal Wear [mIoU]wmIoU [mIoU]LRBSDO
A1MTiCN 00.99610.84390.52290.45860.53050.00180
A1MTiCN 10.99640.95480.65680.58470.65360.00180.3
A1MTiCN 20.99580.92620.45960.50720.58420.00180.5
A1MTiCN 30.99690.95510.61360.54450.62090.001160
A1MTiCN 40.99590.81730.29340.21740.32900.001160.3
A1MTiCN 50.99550.83220.56380.59340.63750.001160.5
A1MTiCN 60.99750.82190.45850.45610.52390.000180
A1MTiCN 70.99750.89840.57590.47940.55760.000180.3
A1MTiCN 80.99660.80500.41000.44020.50760.000180.5
A1MTiCN 90.99710.70310.38860.36750.43010.0001160
A1MTiCN 100.99640.59830.34860.36720.41040.0001160.3
A1MTiCN 110.99630.75070.31190.29700.38120.0001160.5
Annotator 2Background [mIoU]Tool [mIoU]Abnormal Wear [mIoU]Normal Wear [mIoU]wmIoU [mIoU]LRBSDO
A2MTiCN 00.89830.35330.15990.25710.27500.00180
A2MTiCN 10.99610.79110.48700.42490.49330.00180.3
A2MTiCN 20.99510.94070.59470.56170.63200.00180.5
A2MTiCN 30.99620.76670.40760.18810.29700.001160
A2MTiCN 40.99700.93780.58780.55550.62640.001160.3
A2MTiCN 50.99580.90050.37050.59340.64850.001160.5
A2MTiCN 60.99550.74280.35190.19570.29840.000180
A2MTiCN 70.99740.93960.58530.35470.46460.000180.3
A2MTiCN 80.99700.89600.56990.44990.53330.000180.5
A2MTiCN 90.99680.79050.49860.26000.36010.0001160
A2MTiCN 100.99590.90180.50480.36170.46270.0001160.3
A2MTiCN 110.99650.80190.33850.34340.42830.0001160.5
Annotator 3Background [mIoU]Tool [mIoU]Abnormal Wear [mIoU]Normal Wear [mIoU]wmIoU [mIoU]LRBSDO
A3MTiCN 00.99150.86270.41240.53630.59570.00180
A3MTiCN 10.98960.97160.59220.53140.61320.00180.3
A3MTiCN 20.98590.89330.64100.39690.49060.00180.5
A3MTiCN 30.98930.77730.48760.32450.40970.001160
A3MTiCN 40.98660.83500.57730.31290.41160.001160.3
A3MTiCN 50.98430.77440.48340.34750.42770.001160.5
A3MTiCN 60.99040.63800.22490.31510.37460.000180
A3MTiCN 70.99090.89760.63250.48540.56280.000180.3
A3MTiCN 80.98800.76840.42720.25610.35230.000180.5
A3MTiCN 90.99140.76910.23610.38420.45450.0001160
A3MTiCN 100.98860.75100.53410.43910.49780.0001160.3
A3MTiCN 110.98700.79210.26460.28600.37960.0001160.5

References

  1. Shad, R.; Cunningham, J.P.; Ashley, E.A.; Langlotz, C.P.; Hiesinger, W. Designing clinically translatable artificial intelligence systems for high-dimensional medical imaging. Nat. Mach. Intell. 2021, 3, 929–935. [Google Scholar] [CrossRef]
  2. Peiffer-Smadja, N. Machine learning for COVID-19 needs global collaboration and data-sharing. Nat. Mach. Intell. 2020, 2, 293–294. [Google Scholar] [CrossRef]
  3. Hu, Y. The challenges of deploying artificial intelligence models in a rapidly evolving pandemic. Nat. Mach. Intell. 2020, 2, 298–300. [Google Scholar] [CrossRef]
  4. Willemink, M.J. Preparing medical imaging data for machine learning. Radiology 2020, 295, 4–15. [Google Scholar] [CrossRef] [PubMed]
  5. Northcutt, C.G.; Athalye, A.; Mueller, J. Pervasive label errors in test sets destabilize machine learning benchmarks. arXiv 2021, arXiv:2103.14749. [Google Scholar]
  6. Rottmann, M.; Reese, M. Automated detection of label errors in semantic segmentation datasets via deep learning and uncertainty quantification. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision 2023, Waikoloa, HI, USA, 2–7 January 2023; pp. 3214–3223. [Google Scholar]
  7. Paullada, A.; Raji, I.D.; Bender, E.M.; Denton, E.; Hanna, A. Data and its (dis)contents: A survey of dataset development and use in machine learning research. Patterns 2021, 2, 100336. [Google Scholar] [CrossRef] [PubMed]
  8. Peng, R.; Liu, J.; Fu, X.; Liu, C.; Zhao, L. Application of machine vision method in tool wear monitoring. Int. J. Adv. Manuf. Technol. 2021, 116, 1357–1372. [Google Scholar] [CrossRef]
  9. Peres, R.S.; Guedes, M.; Miranda, F.; Barata, J. Simulation-Based Data Augmentation for the Quality Inspection of Structural Adhesive With Deep Learning. IEEE Access 2021, 9, 76532–76541. [Google Scholar] [CrossRef]
  10. Survey: 96% Enterprises Encounter Training Data Quality. Available online: https://www.businesswire.com/news/home/20190523005183/en/Survey-96-Enterprises-Encounter-Training-Data-Quality (accessed on 10 May 2024).
  11. Su, H.; Deng, J.; Fei-Fei, L. Crowdsourcing annotations for visual object detection. In Proceedings of the Workshops at the Twenty-Sixth AAAI Conference on Artificial Intelligence, Toronto, ON, Canada, 22–26 July 2012. [Google Scholar]
  12. Bhuiyan, M.; Choudhury, I.A.; Dahari, M.; Nukman, Y.; Dawal, S.Z. Application of acoustic emission sensor to investigate the frequency of tool wear and plastic deformation in tool condition monitoring. Measurement 2016, 92, 208–217. [Google Scholar] [CrossRef]
  13. Sun, W.H.; Yeh, S.S. Using the Machine Vision Method to Develop an On-machine Insert Condition Monitoring System for Computer Numerical Control Turning Machine Tools. Materials 2018, 11, 1977. [Google Scholar] [CrossRef]
  14. Bilal, M.; Mayer, C. Objektbeleuchtung; EP4130720; European Patent Office: Munich, Germany, 2023. [Google Scholar]
  15. Kumar, S.; Datta, S.; Singh, V.; Singh, S.K.; Sharma, R. Opportunities and Challenges in Data-Centric AI. IEEE Access 2024, 12, 33173–33189. [Google Scholar] [CrossRef]
  16. Chen, W.; Iyer, A.; Bostanabad, R. Data centric design: A new approach to design of microstructural material systems. Engineering 2022, 10, 89–98. [Google Scholar] [CrossRef]
  17. Fang, C.; Xu, Y.; Rockmore, D.N. Unbiased metric learning: On the utilization of multiple datasets and web images for softening bias. In Proceedings of the IEEE International Conference on Computer Vision 2013, Sydney, NSW, Australia, 1–8 December 2013; pp. 1657–1664. [Google Scholar]
  18. Recht, B.; Roelofs, R.; Schmidt, L.; Shankar, V. Do imagenet classifiers generalize to imagenet? In Proceedings of the International Conference on Machine Learning. PMLR 2019, Long Beach, CA, USA, 9–15 June 2019; pp. 5389–5400. [Google Scholar]
  19. Shankar, V.; Roelofs, R.; Mania, H.; Fang, A.; Recht, B.; Schmidt, L. Evaluating machine accuracy on imagenet. In Proceedings of the International Conference on Machine Learning. PMLR, 2020, Virtual, 13–18 July 2020; pp. 8634–8644. [Google Scholar]
  20. van Horn, G.; Branson, S.; Farrell, R.; Haber, S.; Barry, J.; Ipeirotis, P.; Perona, P.; Belongie, S. Building a bird recognition app and large scale dataset with citizen scientists: The fine print in fine-grained dataset collection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2015, Boston, MA, USA, 7–12 June 2015; pp. 595–604. [Google Scholar]
  21. Wah, C.; Branson, S.; Welinder, P.; Perona, P.; Belongie, S. The Caltech-ucsd Birds-200-2011 Dataset; Caltech: Pasadena, CA, USA, 2011. [Google Scholar]
  22. Esteva, A.; Robicquet, A.; Ramsundar, B.; Kuleshov, V.; DePristo, M.; Chou, K.; Cui, C.; Corrado, G.; Thrun, S.; Dean, J. A guide to deep learning in healthcare. Nat. Med. 2019, 25, 24–29. [Google Scholar] [CrossRef] [PubMed]
  23. Cordts, M.; Omran, M.; Ramos, S.; Rehfeld, T.; Enzweiler, M.; Benenson, R.; Franke, U.; Roth, S.; Schiele, B. The cityscapes dataset for semantic urban scene understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2016, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 3213–3223. [Google Scholar]
  24. Taran, V.; Gordienko, Y.; Rokovyi, A.; Alienin, O.; Stirenko, S. Impact of ground truth annotation quality on performance of semantic image segmentation of traffic conditions. In Advances in Computer Science for Engineering and Education II; Springer: Berlin/Heidelberg, Germany, 2020; pp. 183–193. [Google Scholar]
  25. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid scene parsing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2017, Honolulu, HI, USA, 21–26 July 2017; pp. 2881–2890. [Google Scholar]
  26. Wu, X.; Liu, Y.; Zhou, X.; Mou, A. Automatic identification of tool wear based on convolutional neural network in face milling process. Sensors 2019, 19, 3817. [Google Scholar] [CrossRef]
  27. Holst, C.; Yavuz, T.B.; Gupta, P.; Ganser, P.; Bergs, T. Deep learning and rule-based image processing pipeline for automated metal cutting tool wear detection and measurement. IFAC-PapersOnLine 2022, 55, 534–539. [Google Scholar] [CrossRef]
  28. Bergs, T.; Holst, C.; Gupta, P.; Augspurger, T. Digital image processing with deep learning for automated cutting tool wear detection. Procedia Manuf. 2020, 48, 947–958. [Google Scholar] [CrossRef]
  29. Lutz, B.; Kisskalt, D.; Regulin, D.; Hauser, T.; Franke, J. Material Identification for Smart Manufacturing Systems: A Review. In Proceedings of the 2021 4th IEEE International Conference on Industrial Cyber-Physical Systems (ICPS), Victoria, BC, Canada, 10–12 May 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 353–360. [Google Scholar] [CrossRef]
  30. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015; Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2015; Volume 9351, pp. 234–241. [Google Scholar] [CrossRef]
  31. He, K.; Gkioxari, G.; Dollar, P.; Girshick, R. Mask R-CNN. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 386–397. [Google Scholar] [CrossRef]
  32. Allen Goodman, Anne Carpenter, Elizabeth Park, jlefman-nvidia, Josette BoozAllen, Kyle, Maggie, Nilofer, Peter Sedivec, Will Cukierski. 2018 Data Science Bowl. Kaggle. 2018. Available online: https://kaggle.com/competitions/data-science-bowl-2018 (accessed on 1 May 2024).
  33. Jacobkie. Data Science Bowl 2nd Place Solution. 2018. Available online: https://github.com/jacobkie/2018DSB (accessed on 12 June 2024).
  34. Cai, L.; Long, T.; Dai, Y.; Huang, Y. Mask R-CNN-Based Detection and Segmentation for Pulmonary Nodule 3D Visualization Diagnosis. IEEE Access 2020, 8, 44400–44409. [Google Scholar] [CrossRef]
  35. Jain, A.K.; Lad, B.K. A novel integrated tool condition monitoring system. J. Intell. Manuf. 2019, 30, 1423–1436. [Google Scholar] [CrossRef]
  36. Su, Z.; Li, W.; Ma, Z.; Gao, R. An improved U-Net method for the semantic segmentation of remote sensing images. Appl. Intell. 2022, 52, 3276–3288. [Google Scholar] [CrossRef]
  37. Pfeifer, T.; Wiegers, L. Reliable tool wear monitoring by optimized image and illumination control in machine vision. Measurement 2000, 28, 209–218. [Google Scholar] [CrossRef]
  38. Russell, B.C.; Torralba, A.; Murphy, K.P.; Freeman, W.T. LabelMe: A database and web-based tool for image annotation. Int. J. Comput. Vis. 2008, 77, 157–173. [Google Scholar] [CrossRef]
  39. Grigoriev, S.N.; Migranov, M.S.; Melnik, Y.A.; Okunkova, A.A.; Fedorov, S.V.; Gurin, V.D.; Volosova, M.A. Application of adaptive materials and coatings to increase cutting tool performance: Efficiency in the case of composite powder high speed steel. Coatings 2021, 11, 855. [Google Scholar] [CrossRef]
  40. Lutz, B.; Reisch, R.; Kisskalt, D.; Avci, B.; Regulin, D.; Knoll, A.; Franke, J. Benchmark of Automated Machine Learning with State-of-the-Art Image Segmentation Algorithms for Tool Condition Monitoring. Procedia Manuf. 2020, 51, 215–221. [Google Scholar] [CrossRef]
Figure 1. (a) Important geometric parameters of an end mill, such as relief face, end angle on the axial rake, rake face, axial relief face, and helix angle. (b) Illustration of light reflection on a TiN-coated end mill when illuminated by standard direct diffuse lighting. Notably, the most intense reflection is observed along the cutter’s edges, while shadowing is evident within the inner rake space. (c) Image captured by the IAS.
Figure 1. (a) Important geometric parameters of an end mill, such as relief face, end angle on the axial rake, rake face, axial relief face, and helix angle. (b) Illustration of light reflection on a TiN-coated end mill when illuminated by standard direct diffuse lighting. Notably, the most intense reflection is observed along the cutter’s edges, while shadowing is evident within the inner rake space. (c) Image captured by the IAS.
Sensors 24 04777 g001
Figure 2. Schematic representation of the measurement setup for capturing high-quality images of end mills. The acquisition system consists of (1) a hemisphere with barium sulfate coating, (2) 12 LEDs located at the edge of the hemisphere, (3) a tool to be examined, which is held by (4) a three-jaw chuck, (5) represents the rotating plate for a 360° recording, (6) is a camera with an interface (7) connected to the computer.
Figure 2. Schematic representation of the measurement setup for capturing high-quality images of end mills. The acquisition system consists of (1) a hemisphere with barium sulfate coating, (2) 12 LEDs located at the edge of the hemisphere, (3) a tool to be examined, which is held by (4) a three-jaw chuck, (5) represents the rotating plate for a 360° recording, (6) is a camera with an interface (7) connected to the computer.
Sensors 24 04777 g002
Figure 3. Positive annotation examples. The wear classification includes two primary categories: “yellow” represents typical wear and “green” denotes abnormal wear and the two additional categories: “red” for the background and “black” for the tool.
Figure 3. Positive annotation examples. The wear classification includes two primary categories: “yellow” represents typical wear and “green” denotes abnormal wear and the two additional categories: “red” for the background and “black” for the tool.
Sensors 24 04777 g003
Figure 4. Negative examples of annotations. Each sub-figure highlights a distinct type of incorrect annotation. The classifications include: “yellow” for normal wear, “green” for abnormal wear, “red” for the background, and “black” for the tool itself. (a) shows mislabeling abnormal wear as normal wear, (b) incorrect annotations mistakenly marking the background as a part of the tool surface, (c) impurities have been labeled as abnormal wear, (d) material removal due to chip residue at chipping space has been marked as normal wear.
Figure 4. Negative examples of annotations. Each sub-figure highlights a distinct type of incorrect annotation. The classifications include: “yellow” for normal wear, “green” for abnormal wear, “red” for the background, and “black” for the tool itself. (a) shows mislabeling abnormal wear as normal wear, (b) incorrect annotations mistakenly marking the background as a part of the tool surface, (c) impurities have been labeled as abnormal wear, (d) material removal due to chip residue at chipping space has been marked as normal wear.
Sensors 24 04777 g004
Figure 5. CNN architecture for normal and abnormal wear segmentation. Encoding blocks are colored in blue. The bottleneck layer is colored in brown. Decoding blocks are colored in green.
Figure 5. CNN architecture for normal and abnormal wear segmentation. Encoding blocks are colored in blue. The bottleneck layer is colored in brown. Decoding blocks are colored in green.
Sensors 24 04777 g005
Figure 6. Illustrative images captured via the acquisition system: (a) TiCN-coated endmill and (b) TiN-coated endmill.
Figure 6. Illustrative images captured via the acquisition system: (a) TiCN-coated endmill and (b) TiN-coated endmill.
Sensors 24 04777 g006
Figure 7. Masks of endmill wear annotations for comparison of a TiCN-coated endmill dataset. The annotations were performed by three annotators: Annotator 1, Annotator 2, and Annotator 3. The dataset includes four classes: normal wear in green, abnormal wear in yellow, background in red, and tool in black. The critical annotations have been marked in red.
Figure 7. Masks of endmill wear annotations for comparison of a TiCN-coated endmill dataset. The annotations were performed by three annotators: Annotator 1, Annotator 2, and Annotator 3. The dataset includes four classes: normal wear in green, abnormal wear in yellow, background in red, and tool in black. The critical annotations have been marked in red.
Sensors 24 04777 g007
Figure 8. Masks of Endmill wear annotations for comparison of a TiN-coated endmill dataset. The annotations were performed by three annotators: Annotator 1, Annotator 2, and Annotator 3. The dataset includes four classes: normal wear in green, abnormal wear in yellow, background in red, and tool in black. The critical annotations have been marked in red.
Figure 8. Masks of Endmill wear annotations for comparison of a TiN-coated endmill dataset. The annotations were performed by three annotators: Annotator 1, Annotator 2, and Annotator 3. The dataset includes four classes: normal wear in green, abnormal wear in yellow, background in red, and tool in black. The critical annotations have been marked in red.
Sensors 24 04777 g008
Figure 9. mIoU results of various models for classes of interest: (a) normal wear and (b) abnormal wear. These models were trained using the same dataset but labeled by different annotators. The LR was set to 0.001 and 0.0001, and hyperparameters such as BS and LR varied, as detailed in Table A1. The dataset originates from a TiN-coated end mill. The standard deviation is depicted to illustrate the performance variation among annotators 1, 2, and 3.
Figure 9. mIoU results of various models for classes of interest: (a) normal wear and (b) abnormal wear. These models were trained using the same dataset but labeled by different annotators. The LR was set to 0.001 and 0.0001, and hyperparameters such as BS and LR varied, as detailed in Table A1. The dataset originates from a TiN-coated end mill. The standard deviation is depicted to illustrate the performance variation among annotators 1, 2, and 3.
Sensors 24 04777 g009aSensors 24 04777 g009b
Figure 10. mIoU results of various models for classes of interest: (a) normal wear and (b) abnormal wear. These models were trained using the same dataset but labeled by different annotators. The LR was set to 0.001 and 0.0001, and hyperparameters such as BS and LR varied, as detailed in Table A2. The dataset originates from a TiCN-coated end mill. The standard deviation is depicted to illustrate the performance variation among annotators 1, 2, and 3.
Figure 10. mIoU results of various models for classes of interest: (a) normal wear and (b) abnormal wear. These models were trained using the same dataset but labeled by different annotators. The LR was set to 0.001 and 0.0001, and hyperparameters such as BS and LR varied, as detailed in Table A2. The dataset originates from a TiCN-coated end mill. The standard deviation is depicted to illustrate the performance variation among annotators 1, 2, and 3.
Sensors 24 04777 g010
Figure 11. Prediction results and corresponding masks on test images from a TiN-coated milling tool, predicted by the best-performing models of the three annotators, 1, 2 and 3, as detailed and bold in Table A1. The prediction includes the four classes: normal wear in green, abnormal wear in yellow, background in red, and tool in black. Wrong annotations in the GT Mask are marked red.
Figure 11. Prediction results and corresponding masks on test images from a TiN-coated milling tool, predicted by the best-performing models of the three annotators, 1, 2 and 3, as detailed and bold in Table A1. The prediction includes the four classes: normal wear in green, abnormal wear in yellow, background in red, and tool in black. Wrong annotations in the GT Mask are marked red.
Sensors 24 04777 g011
Figure 12. Prediction results and corresponding masks on test images from a TiN-coated milling tool, predicted by the best-performing models of the three annotators, 1, 2 and 3, as detailed and bold in Table A2. The prediction includes the four classes: normal wear in green, abnormal wear in yellow, background in red, and tool in black. Critical regions, such as wrong predictions or missed wear recognition, are marked in red.
Figure 12. Prediction results and corresponding masks on test images from a TiN-coated milling tool, predicted by the best-performing models of the three annotators, 1, 2 and 3, as detailed and bold in Table A2. The prediction includes the four classes: normal wear in green, abnormal wear in yellow, background in red, and tool in black. Critical regions, such as wrong predictions or missed wear recognition, are marked in red.
Sensors 24 04777 g012
Table 1. Relevant training parameters for the CNN.
Table 1. Relevant training parameters for the CNN.
ParametersValue
Image Size512 × 512 × 3
Image FormatJpeg
BS8, 16
DO0.0, 0.3, 0.5
Epochs70
GPU’s1
Trainable Parameters2,140,740
LossSparse Categorical Cross Entropy
OptimizerRMS Prop
MetricIoU
Train/Valid/Test0.8/0.1/0.1
Table 2. Instance distribution across datasets.
Table 2. Instance distribution across datasets.
Tool CoatingClass BackgroundClass Normal WearClass Abnormal WearClass Tool
TiCN432404806768
TiN432770532768
Table 3. CV for TiN-coated tool.
Table 3. CV for TiN-coated tool.
ModelBackground CV [%]Tool CV [%]Abnormal Wear CV [%]Normal Wear CV [%]wmIoU CV [%]LRBSDO
MTiN 00.130.247.399.618.910.00180
MTiN 10.102.227.4910.058.900.00180.3
MTiN 20.040.372.8210.058.070.00180.5
MTiN 30.1816.789.2718.4912.160.001160
MTiN 40.081.698.072.043.120.001160.3
MTiN 50.030.3911.479.946.170.001160.5
MTiN 60.042.589.9211.2410.660.000180
MTiN 70.011.0910.028.147.430.000180.3
MTiN 80.032.737.327.427.300.000180.5
MTiN 90.030.8120.5017.7418.450.0001160
MTiN 100.031.413.894.434.260.0001160.3
MTiN 110.041.895.413.673.350.0001160.5
Mean CV0.072.758.259.488.61---
Table 4. CV for TiCN-coated tool.
Table 4. CV for TiCN-coated tool.
ModelBackground CV [%]Tool CV [%]Abnormal Wear CV [%]Normal Wear CV [%]wmIoU CV [%]LRBSDO
MTiCN 04.6934.3441.6228.1929.630.00180
MTiCN 10.318.9912.0912.9311.600.00180.3
MTiCN 20.452.1613.6214.0310.320.00180.5
MTiCN 30.3510.3816.8641.6730.340.001160
MTiCN 40.476.1528.0539.3227.510.001160.3
MTiCN 50.546.1716.7822.6717.780.001160.5
MTiCN 60.3010.2627.6733.0123.480.000180
MTiCN 70.312.154.1413.698.540.000180.3
MTiCN 80.426.5215.2823.3417.210.000180.5
MTiCN 90.264.9328.7416.339.640.0001160
MTiCN 100.3616.5117.609.067.860.0001160.3
MTiCN 110.442.8410.028.055.690.0001160.5
Mean CV0.769.9718.5320.2416.32---
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bilal, M.; Podishetti, R.; Koval, L.; Gaafar, M.A.; Grossmann, D.; Bregulla, M. The Effect of Annotation Quality on Wear Semantic Segmentation by CNN. Sensors 2024, 24, 4777. https://doi.org/10.3390/s24154777

AMA Style

Bilal M, Podishetti R, Koval L, Gaafar MA, Grossmann D, Bregulla M. The Effect of Annotation Quality on Wear Semantic Segmentation by CNN. Sensors. 2024; 24(15):4777. https://doi.org/10.3390/s24154777

Chicago/Turabian Style

Bilal, Mühenad, Ranadheer Podishetti, Leonid Koval, Mahmoud A. Gaafar, Daniel Grossmann, and Markus Bregulla. 2024. "The Effect of Annotation Quality on Wear Semantic Segmentation by CNN" Sensors 24, no. 15: 4777. https://doi.org/10.3390/s24154777

APA Style

Bilal, M., Podishetti, R., Koval, L., Gaafar, M. A., Grossmann, D., & Bregulla, M. (2024). The Effect of Annotation Quality on Wear Semantic Segmentation by CNN. Sensors, 24(15), 4777. https://doi.org/10.3390/s24154777

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop