Nothing Special   »   [go: up one dir, main page]

Next Article in Journal
A Multi-Modal Modulation Recognition Method with SNR Segmentation Based on Time Domain Signals and Constellation Diagrams
Next Article in Special Issue
Multi-Modality Tensor Fusion Based Human Fatigue Detection
Previous Article in Journal
Resilient Localization and Coverage in the Internet of Things
Previous Article in Special Issue
An Improved Median Filter Based on YOLOv5 Applied to Electrochemiluminescence Image Denoising
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Underwater Image Color Constancy Calculation with Optimized Deep Extreme Learning Machine Based on Improved Arithmetic Optimization Algorithm

1
School of Mechanical Engineering, Hangzhou Dianzi University, Hangzhou 310018, China
2
School of Electronics and Information Engineering, Hangzhou Dianzi University, Hangzhou 310018, China
3
School of Computer Science and Technology, Zhejiang Sci-Tech University, Hangzhou 310018, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(14), 3174; https://doi.org/10.3390/electronics12143174
Submission received: 4 July 2023 / Revised: 19 July 2023 / Accepted: 19 July 2023 / Published: 21 July 2023
Figure 1
<p>DELM Network structure.</p> ">
Figure 2
<p>Search agent mapping flowchart.</p> ">
Figure 3
<p>IAOA-DELM algorithm flowchart.</p> ">
Figure 4
<p>Experimental facility.</p> ">
Figure 5
<p>IAOA-DELM light correction process diagram.</p> ">
Figure 6
<p>The average chrominance accuracy under different population sizes and number of iterations when alpha = 0.25.</p> ">
Figure 7
<p>The average chrominance accuracy under different population sizes and number of iterations when alpha = 0.5.</p> ">
Figure 8
<p>The average chrominance accuracy under different population sizes and number of iterations when alpha = 0.25.</p> ">
Figure 9
<p>Comparison of average accuracy results between IAOA-DELM algorithm and comparison group algorithm.</p> ">
Figure 10
<p>Algorithm angle error stability analysis box diagram.</p> ">
Figure 11
<p>Correction results of different illumination correction models (<b>A</b>) Underwater image to be corrected under TL83 light source (<b>B</b>) Standard image in air under the same scene under D50 light source (<b>C</b>) IAOA-DELM (<b>D</b>) AOA-DELM (<b>E</b>) GWO-DELM (<b>F</b>) WOA-DELM (<b>G</b>) HHO-DELM (<b>H</b>) ORELM (<b>I</b>) ELM (<b>J</b>) RVFL (<b>K</b>) BP.</p> ">
Figure 12
<p>Comparison with other classical and advanced image correction algorithms (<b>A</b>) Underwater image to be corrected under TL83 light source (<b>B</b>) Standard image in air under the same scene under D50 light source (<b>C</b>) IAOA-DELM (<b>D</b>) Interactive WB Method (<b>E</b>) Data-Driven WB Method (<b>F</b>) WB color augmenter (<b>G</b>) Grey-World (<b>H</b>) Max-RGB (<b>I</b>) Shades of Grey.</p> ">
Review Reports Versions Notes

Abstract

:
To overcome the challenges posed by the underwater environment and restore the true colors of marine objects’ surfaces, a novel underwater image illumination estimation model, termed the iterative chaotic improved arithmetic optimization algorithm for deep extreme learning machines (IAOA-DELM), is proposed. In this study, the gray edge framework is utilized to extract color features from underwater images, which are employed as input vectors. To address the issue of unstable prediction results caused by the random selection of parameters in DELM, the arithmetic optimization algorithm (AOA) is integrated, and the search segment mapping method is optimized by using hidden layer biases and input layer weights. Furthermore, an iterative chaotic mapping initialization strategy is incorporated to provide AOA with a better initial search proxy. The IAOA-DELM model computes illumination information based on the input color vectors. Experimental evaluations conducted on actual underwater images demonstrate that the proposed IAOA-DELM illumination correction model achieves an accuracy of 96.07%. When compared to the ORELM, ELM, RVFL, and BP models, the IAOA-DELM model exhibits improvements of 6.96%, 7.54%, 8.00%, and 8.89%, respectively, making it the most effective among the compared illumination correction models.

1. Introduction

In the field of marine observation, scientists utilize underwater robots to capture photographs for studying marine organisms and mineral resources. For instance, in the research of coral bleaching phenomena and underwater polymetallic nodule deposits, the analysis and recognition of the color characteristics of underwater targets are essential [1,2]. However, in complex underwater environments, various factors, such as lighting conditions, spectral absorption by water media, and backscattering of particles, hinder the direct reflection of the true colors of objects in images [3]. Even with the same light source, the resulting image colors can vary. This instability of colors in underwater images poses a challenge, as the technology to obtain the accurate surface colors of marine objects is not yet mature [4]. Therefore, studying color constancy in underwater images becomes an urgent problem to address.
Early methods for color constancy analysis in underwater images were based on the statistical analysis of pixel values. Among them, the white-patch algorithm [5,6] assumes that a white surface can adequately reflect the illuminant color of the scene by selecting the maximum value among the RGB color channels as the illuminant color for the image. However, this algorithm’s estimation performance is not optimal when the overall scene brightness is low. The gray-world algorithm [7,8] assumes that for color-rich images, the average pixel values of the three RGB color channels tend to be similar. However, this method’s estimation performance is not ideal for images with limited color or a single color, making it challenging to apply in underwater environments. Li et al. [9] proposed a wavelet transform method based on the YUV color model, which significantly improves imaging quality. Yan et al. [10] proposed a new color constancy framework based on the relationship between the reflectance difference and the local normalized reflectance difference. Iqbal et al. [11] introduced the Laplacian transform to minimize artifacts and noise. With the advancement of underwater image processing techniques, Hassan et al. [12] constructed an underwater image processing method supported by Retinex theory and achieved better results. However, these aforementioned algorithms are manually designed and have certain limitations in their application, as they cannot effectively perform image illumination estimation under different lighting conditions and complex environments.
The application of machine learning techniques in color constancy research has provided a new direction for solving complex pattern recognition problems. In recent years, researchers have started to incorporate machine learning methods into color constancy analysis, particularly algorithms based on image feature learning. Different learning methods have been proposed for color constancy, such as Bayesian-based color constancy algorithms [13], backpropagation (BP) neural network-based color constancy algorithms [14], support vector regression (SVR)-based illuminance estimation algorithms [15], and extreme learning machine (ELM)-based illuminance estimation algorithms [16]. Furthermore, deep learning methods with more powerful learning capabilities have been added to the application of color constancy [17]. Deep learning color constancy methods based on convolutional neural networks (CNN) [18,19], transfer learning [20], fast Fourier transform-based color constancy methods [21], and contrastive learning for color Constancy [22] have been proposed. In deep learning algorithms, image features are determined during the training of the network, and the complex deep learning network models significantly increase the computational burden. To combine the powerful computational capabilities of deep learning algorithms with the efficient learning ability of ELM, researchers have extensively investigated the integration of ELM into general deep learning frameworks. Deng et al. [23] presented the DELM by combining ELM with the idea of autoencoders, called the Extreme Learning Machine Autoencoder (ELM-AE). DELM exhibits strong nonlinear modeling and generalization abilities, which can be further expanded by using larger-scale training data and deeper network structures, showing great potential in the application of underwater image color constancy computation. Some studies have found that applying a swarm intelligence algorithm to find the optimal parameters of DELM is very helpful to its performance [24,25,26].
Motivated by the aforementioned research, this study proposes an iterative chaotic-based arithmetic optimization algorithm (IAOA) to optimize the light estimation model of DELM, referred to as IAOA-DELM. For underwater images captured under unknown illuminations, the IAOA-DELM approach first extracts the color features of the images using the gray edge framework [27]. Then, based on IAOA-DELM, it estimates the light source of the images. Finally, the estimated light source is applied to correct the underwater images to standard illuminant values using the Von Kries diagonal model [28]. This correction ensures that the underwater images collected under different color temperature lighting conditions exhibit color-accurate results. The innovation of this paper is primarily demonstrated in the following aspects:
(1)
Constructing the basic model of DELM to compute scene illumination information from the color features of underwater images.
(2)
To address the stability and generalization issues caused by the initial parameters in the orthogonal matrix, AOA is employed to optimize the input layer weights and thresholds of ELM-AE in the DELM structure. The search and development stages of AOA are combined with the nonlinear feature mapping stage of ELM-AE.
(3)
AOA is applied to select the hidden layer nodes’ number and adaptively search for the optimization of effective activation nodes. It simultaneously optimizes hidden layer biases, input weights, and hidden layer nodes’ numbers, obtaining an underwater image illumination estimation model with good predictive performance and stability.
(4)
The overall initial search agents of AOA are generated using iterative chaos mapping to improve the initialization strategy of AOA and obtain IAOA. In the initialization strategy, without prior knowledge, IAOA enhances the initial population’s quality, thereby improving the algorithm’s operation speed and accuracy.
The other chapters can be divided into the following sections: Section 2 provides an introduction to AOA and DELM. Section 3 describes the image dataset built and presents the proposed color constancy model (IAOA-DELM). Section 4 describes the analysis of the experimental results. Finally, Section 5 is the summary of the article.

2. Theoretical Basis

2.1. Arithmetic Optimization Algorithm

AOA is a meta-heuristic optimization algorithm based on the distribution behavior of arithmetic operators (addition, subtraction, multiplication, and division) [28]. It is characterized by the algorithm’s fast convergence speed and high accuracy.
The AOA algorithm determines which search phase to enter according to the relationship between r1 and the value of the mathematical optimizer acceleration function (A). r1 [ 0 , 1 ] . When r1 > A for global exploration; when r1 < A, AOA is developed locally. A is shown as follows:
A ( C i t e r ) = M i n + C i t e r × M a x M i n M i t e r
where A(Citer) represents the value of the current iteration. Citer represents the current iteration. Miter represents the total number of iterations. Min and Max represent the maximum and minimum values of the acceleration function, respectively.
The AOA algorithm uses multiplication and division strategies for global search. r2 is a random number between 0 and 1. When r2 < 0.5, AOA enters the division search stage. When r2 > 0.5, AOA enters the multiplication search stage. The position update formula of the multiplication and division strategy is as follows:
X i , j C i t e r + 1 = b e s t ( X j ) / ( M O P + ε ) × ( ( U B j L B j ) × μ + L B j ) , r 2 < 0.5 b e s t ( X j ) × M O P × ( ( U B j L B j ) × μ + L B j ) , r 2 > 0.5
where, Xi,j(Citer + 1) represents the jth position of the ith solution in the next iteration, and best(Xj) represents the best obtained solution’s jth position so far. ε is a very small integer, UBj and LBj represent the upper and lower bounds at the jth position. μ is a control parameter that adjusts the search process, which is fixed at 0.5 in this research. The mathematical optimizer probability (P) is shown as follows:
P ( C i t e r ) = 1 C i t e r α M i t e r α
where, α defines the mining accuracy in the iterative operation process.
The AOA algorithm uses addition and subtraction strategies for local development. r3 is a random number between 0 and 1. When r3 > 0.5, AOA enters the phase of addition search. When r3 < 0.5, AOA enters the subtraction search stage. The formula for the strategic position update for addition and subtraction is as follows:
X i , j C i t e r + 1 = b e s t ( X j ) M O P × ( ( U B j L B j ) × μ + L B j ) , r 3 < 0.5 b e s t ( X j ) + M O P × ( ( U B j L B j ) × μ + L B j ) , r 3 > 0.5

2.2. Deep Extreme Learning Machine

Suppose there are N arbitrary samples ( X j , t j ), X j = [ x j 1 , x j 2 , . . . , x j n ] R n and t j = [ o j 1 , o j 2 , . . . , o j m ] R m , an ELM [29] with L hidden layer nodes can be shown as:
i = 1 L β i g ( W i X j + b i ) = o j , j = 1 , 2 , , N
where g(x) is the activation function. W i = [ w i 1 , w i 2 , . . . , w i n ] T is the ith hidden layer neuron’s input weight. bi is the ith hidden layer of layer neurons’ bias. β j = [ β j 1 , β j 2 , . . . , β j m ] T is the ith hidden layer neuron’s output weight. represents the inner product.
The output weight matrix is:
β ^ = H + T
where H+ is the Moore-Penrose generalized matrix of the output matrix H.
DELM takes ELM as an autoencoder (ELM-AE) and randomly generates the parameter vector in ELM-AE. DELM employs the ELM-AE technique for training the first-layer network and obtaining the output of the first-layer hidden layer. This output serves as the input for training the second-layer network’s output using a similar process, continuing until the final hidden layer. The output between the last hidden layer and the output layer is computed using the least squares method. The network of DELM is expressed in Figure 1.
DELM uses multiple ELM-AE superpositions for calculation and maps high-dimensional input data to low-dimensional space with low distortion. This method enables DELM to learn more abstract and useful features in the original data and achieve efficient feature representation.

3. Our Contribution

The selection of weights and biases for each hidden layer in ELM-AE is done randomly, and this randomness can have an influence on the training performance and stability of DELM. To address the instability issue in the selection of DELM parameters, the arithmetic optimization algorithm (AOA) was introduced. However, the AOA initialization strategy still suffers from the drawbacks of an uneven and unstable parameter distribution. To overcome this limitation, an iterative chaotic initialization strategy was incorporated to improve the AOA initialization. After optimizing the common parameters of DELM using the improved AOA (IAOA), it is still necessary to conduct repeated experiments to determine the optimal values for the number of search agents, iteration count, hidden layer quantity, and number of neurons. This section explains the construction of the IAOA-DELM color constancy model and provides insights into the setup of the experimental scenarios and the construction of the dataset.

3.1. Search Agent Strategy of DELM Based on AOA

DELM employs an adaptive mechanism to activate effective neurons in the hidden layer based on the characteristics of the training data set. The number of hidden layer neuron nodes is a crucial parameter in the network structure, as it determines the number of relevant nodes in the input layer and the dimensionality of the effective feature parameters. Insufficient hidden layer nodes can impede DELM from accurately capturing the common features of the training set, while an excessive number of nodes may lead to an overemphasis on the specific features of the training set, thereby compromising its generalization ability. Traditional approaches for determining the optimal number of hidden layer nodes include conducting repeated experiments or relying on empirical knowledge to identify the ideal number. Alternatively, a fixed formula derived from statistical experience can be used to calculate the number of nodes in the hidden layer (m) based on the number of nodes in the input layer (n) and the number of nodes in the output layer (s), such as the formulas m = n s and m = log 2 n . In the first method, researchers set the comparison interval and step size subjectively, which had high complexity and low accuracy. The second method only takes into account the number of nodes in the input and output layers as influential factors, neglecting other important considerations. It will miss the effective feature information of the data itself and produce large training errors. To address this issue, this study introduces a search agent strategy for DELM based on AOA optimization, aiming to determine the number of hidden layer neurons. The steps were as follows:
  • To determine the maximum network structure, the number of hidden layers and the upper limit of the number of hidden layer nodes were set for DELM;
  • The relevant parameters of AOA were initialized, and the n input nodes’ number and the s hidden layer nodes’ number were input into AOA as independent parameters for optimization;
  • The fitness value of each individual was calculated to obtain the optimal parameter combination based on the search agent structure, and the node parameter results of the input and output layers were collected;
  • According to the Ceil function, map the result to 0 or 1 (0 means freezing the node, 1 means activating the node), and calculate the number of optimal hidden layer nodes.
The search agent in the research is expressed in Figure 2:
Figure 2 depicts the search agent structure, which consists of four nodes in the input layer and five nodes in the hidden layer. The second and fourth nodes in the input layer are activated, while the first, fourth, and fifth nodes in the hidden layer are activated. Whether it is activated or not indicates that the network selects the effective input features of the second and fourth for training and connects the effective hidden layer nodes 1, 4, and 5. According to the connection relation between the effective nodes of the input layer and the hidden layer, the input layer weight ( W 12 , W 14 , W 42 , W 44 , W 52 , and W 54 ) and the hidden layer bias ( B 1 , B 4 , and B 5 ) are selected, and the output parameter matrix is formed. Thus, the selection and learning of important features in the training set are completed. The optimization of input layer weight, hidden layer bias, and the number of hidden layer nodes are realized simultaneously. The input weight matrix β of ELM-AE was calculated according to formula (6). The output weight matrix of all hidden layers was obtained through training so as to obtain the AOA-DELM.

3.2. Improved Arithmetic Optimization Algorithm Based on Iterative Chaotic Initialization (IAOA)

In the initial stage of the meta-heuristic optimization algorithm, the initial population is selected by means of random generation, which will lead to an uneven distribution of individuals in the population. Iterative chaotic mapping is characterized by strong chaos and pseudo-randomness. Compared with the initialization strategy of randomly generating the initial position, iterative chaotic mapping can make the population more evenly distributed in the search space [30]. If a population has n individuals, then the random initial population X = { x 1 , …, x k , …, x n }, k ∈ [1, n], and the mathematical expression is as follows:
x k + 1 = sin a π x k
where, x k + 1 is the value of introducing iterative chaotic mapping, a ∈ (0,1). In this paper, a is 0.7.
AOA uses mathematical operators as an optimization means to select the population with the lowest loss function from all common populations (candidate schemes). However, the initial strategy of random distribution of AOA will lead to a large number of individuals in the population moving away from the optimal value, which limits the optimization efficiency of the AOA mechanism. If the population distribution is close enough to the optimal scenario, then the exploration and search phases of AOA will be efficient enough. In order to achieve this goal, an iterative chaos algorithm was introduced to form the initial population distribution of AOA into IAOA.
Suppose there are N search agents in IAOA initialization, each search agent vector has M dimensions, and each variable has the same upper boundary ub and the same lower boundary lb. Firstly, the first dimension value was randomly generated for the first search agent vector, X1. The second dimension value X2 was generated using the iterative chaos algorithm, X2 was opposite to the distribution of X1. Generate the first search proxy vector X2 = { x 1 , …, x k , …, x m } based on the idea that x k + 1 = s i n ( 0.7 π / x k ) . Similarly, the population vector of the remaining N-1 search agents was obtained to form the initial search agent population matrix X2 = { x 1 , …, x N } of IAOA. The strategy of boundary absorption was adopted in the subsequent iteration. The latitude value greater than the upper boundary was set as ub, and the latitude value less than the upper boundary was set as lb.
Compared with the completely random generation of search proxy vectors by the original AOA algorithm, the initial population of IAOA is more widely distributed and more uniform in the search interval. The initial population, evenly distributed within the search interval, has a higher likelihood of capturing the correct eigenvalues, and it is more likely to explore and discover the optimal population.

3.3. Color Constancy Algorithm Flow of Underwater Image Based on IAOA-DELM

The IAOA-DELM color constancy algorithm for underwater images was constructed by combining the IAOA-DELM design idea with underwater image feature learning. The detailed steps and complete process were described as follows:
  • Underwater scene images were shot, and a gray edge frame was used to extract color features from the images as an input vector and constitute the input data set;
  • The number of DELM hidden layers, the number of iterations, and the number of search agents were input. A group of excellent initial populations for AOA was generated by using the iterative chaos algorithm;
  • The dataset was randomly divided into training and test sets using ten-fold cross-validation, where nine subsets were used for training and one subset was used for testing;
  • The training data set was input, the chromaticity feature vector was normalized, and the parameters were limited to search the effective interval. The training set was used as input for training, and the effective nodes of DELM were activated. The enhanced AOA algorithm was employed to optimize the input layer weights, hidden layer biases, and hidden layer nodes of DELM;
  • The fitness of the AOA search agent population was calculated and compared with the best fitness in the previous iteration to decide whether to update the population position;
  • The optimal parameters of IAOA-DELM were obtained after reaching the maximum number of iterations, and the input weight matrix β of ELM-AE was calculated. The output weight matrix of DELM was obtained, and the IAOA-DELM illumination estimation model was constructed;
  • The IAOA-DELM illumination estimation model was used to calculate the illumination of the test set images. The color constancy of underwater images is realized by restoring the image to the standard light source based on the diagonal mapping matrix.
The implementation flow chart of the IAOA-DELM lighting correction model is shown in Figure 3.

3.4. Experimental Scene Construction

The experimental environment was a computer with the CPU model ARM [email protected]. All experiments in this paper were completed by MATLAB2017b software in a Windows 11 environment.
The underwater image acquisition system included a PC terminal, pool, camera, object, and light source. The camera was a Mercury II series high precision industrial camera produced by DaHeng Image Company, model MER2-1220-32U3C. The camera has a resolution of 4024 (H) × 3036 (V), a frame rate of 32.3 fps, and a pixel size of 1.85 µm × 1.85 µm. The optical lens has a resolution of 8 mega pixels and a focal length of 8.0 mm. Light source Seven common lighting sources recommended by the International Commission on Illumination (CIE) were used to provide lighting. The light source types are presented in Table 1.
In the layout of the pool environment, the filter was used to filter the impurities injected into the pool water. The method of reflection lighting was selected. The light source was installed on the same side of the camera to obtain the reflected image of the underwater object. The projection angle was guaranteed to remain unchanged after the replacement of the light source. To reduce the impact of bright spots, we painted all sides except the front (where the camera looks in) black to reduce reflections from the air and water interfaces. In the shooting environment, the doors and windows were closed to ensure that there was only one experimental light source in the scene. The straight-line distance between the target and the camera was 1 m. The experimental environment is shown in Figure 4.

3.5. Data Set Acquisition

To enhance algorithm efficiency, it is essential to utilize an efficient and low-dimensional feature vector as the input for subsequent algorithms. Weijer et al. [31] introduced a gray-edge framework that consolidated various traditional unsupervised color constancy algorithms. The mathematical formula is presented as follows:
n f σ ( x ) x n p d x 1 / p = k e n , p , σ = k e i = ( R i , G i , B i )
where, f σ represents the convolution of image f and Gaussian filter G σ .
By adjusting the parameters n, p and σ , different color constancy algorithms can be derived. The gray edge framework can be used as an image feature statistics tool to introduce higher-order derivative information of the image into the input features. In this research, the value range of n was selected as {0, 1, 2}, and the value range of p was selected as {1, 2, 3, …, 10}. The value range of σ was {1, 3, 5, 7, 9}. There were 150 (3 × 10 × 5) cases in parameter combination i. When r and g were extracted, there were a total of 300-dimensional input feature vectors. In this data set, real illumination RGB information was added as a label by extracting color card information. Here, RGB information was converted to r and g chroma information to remove the influence of illumination intensity.
In this research, 300 underwater images were shot with six light sources of different color temperatures, as shown in Table 1.
The color features of the image dataset were extracted using the gray-edge framework. These extracted chromaticity features, along with the corresponding light source information obtained from the underwater color cards, formed the dataset. To evaluate the model, the dataset was divided into a training set and a test set using the ten-fold cross-validation method. The training set consisted of 270 samples, while the test set contained 30 samples. The training set was utilized to train the neural network and obtain the optimal model parameters. The test set was used to check the correction ability of the model and evaluate the performance of the algorithm.

3.6. Evaluation Index

Chroma error and angle error are important evaluation indexes of illumination correction. The true illumination of the image was set as ( r c , g c ) and the illumination predicted by the algorithm as ( r u , g u ). The chroma error ( E d ) and angle error ( E a ) of the algorithm were shown in Equations (10) and (12). The smaller the angle error and chroma error, the better the effect of the algorithm. Chroma accuracy (CR) was used as the fitness of the search proxy population.
( r i , g i ) = R i R i + B i + G i , G i R i + B i + G i
E d = ( r c r u ) 2 + ( g c g u ) 2
CR = 1     E d
E a = cos 1 e c e u | | e c | |   | | e u | | × 180 π
where, | | | | is the Euclidean norm of a vector.
The diagram of the IAOA-DELM light correction process is shown in Figure 5.

4. Experimental Results and Analysis

4.1. Experimental Parameter Setting

4.1.1. DELM Set Network Parameters

The aforementioned explanation provides a detailed account of how the IAOA-DELM algorithm accomplished the selection of hidden layer nodes in DELM using the search agent strategy. However, there are still two important hyperparameters that need to be set: the upper limit of hidden layer nodes and the number of hidden layers. There was no fixed correspondence between the super parameters of DELM and the regression prediction results, so the super parameters of DELM were determined by experimental comparison and cross-validation. For the purpose of fast training, this research discussed the chromaticity error of the algorithm in the six cases where the number of hidden layers of the IAOA-DELM algorithm was 2, 3, 4, 5, 6, and 7 to select the appropriate number of hidden layers. Based on the Kolmogorov theorem [32], the upper limit for the number of hidden layer nodes in this study was determined to be 600. The DELM algorithm employed an infinity regularization coefficient and utilized the sigmoid function as the activation function. To evaluate the performance of the algorithm on the same dataset, average accuracy and time cost were selected as the measurement indices. A total of fifty experiments were conducted using the ten-fold cross-validation method. The appropriate number of hidden layers was determined by analyzing the average accuracy, standard deviation, maximum accuracy, minimum accuracy, median accuracy, and time cost obtained from these fifty experiments. The experimental results are presented in Table 2.
Table 2 reveals that when the number of hidden layers is set to 2, the algorithm achieves optimal performance in terms of average accuracy, maximum accuracy, minimum accuracy, and median accuracy. Consequently, it is assumed that the number of hidden layers should be set to 2 in subsequent experiments. This demonstrates that an increasing number of hidden layers does not necessarily result in better performance. Excessive hidden layers can lead to a higher computational workload, a longer training time for the algorithm, and an overemphasis on the features of the training set, thereby hindering the accurate identification of image features by DELM.
The choice of activation function has a significant impact on the evaluation of light correction experiment results. To enhance the predictive performance of the model, this research analyzes chroma error E d using the ten-fold cross validation method for different activation functions. The table presents E d of the illumination estimation model on the dataset. From Table 3, it is evident that the Sigmoid activation function yields the smallest average E d among the ten-fold experimental results, indicating superior prediction performance. Hence, this paper adopts the Sigmoid activation function for the DELM model.

4.1.2. Parameter Selection of IAOA

The value of the mathematical optimizer acceleration function A in AOA is influenced by the sensitive parameter α , which plays a crucial role in determining the accuracy of the mining process during iterations. Algorithm iteration times and AOA population size also have a decisive influence on the optimization results of the IAOA algorithm. In order to obtain the optimal parameter combination of sensitive parameter alpha, population size, and iteration times, the average chrominance accuracy of the algorithm was taken as the measurement index. The research carried out an in-depth analysis of the influence of changes in these three parameters on the experimental results. Among them, the sensitive parameter alpha has three values of 0.25, 0.5, or 0.75; the population size has six values of 10, 20, 30, 40, 50, or 60; and the number of iterations has six values of 20, 40, 60, 80, 100, or 120, which has 108 different parameter combinations. The average chrominance accuracy under each parameter combination was obtained by means of 50 experiments. The experimental results are presented in Figure 6, Figure 7 and Figure 8. The x-axis represents the population size, while the y-axis represents the number of iterations. The bubble size represents the average chrominance prediction accuracy.
In Figure 6, Figure 7 and Figure 8, the highest average chrominance accuracy is achieved when alpha is set to 0.5 with a population size of 30 and 100 iterations. Due to the comprehensive consideration of computational complexity, operation efficiency, and accuracy, this parameter combination was selected as the preset parameter in subsequent experiments.

4.2. Compare the Experimental Results of the Group

4.2.1. Comparison Group Algorithm Parameter Selection

To assess the performance of the proposed IAOA-DELM illumination correction model, a comparison group consisting of eight neural network models was constructed, including the AOA-DELM, GWO-DELM, WOA-DELM, HHO-DELM, ORELM, ELM, RVFL, and BP models. The experimental results involved comparing the proposed algorithm with a comparison group algorithm and analyzing the impact of algorithm parameters. Table 4 provides the parameters for each algorithm.

4.2.2. Comparison of Chroma Estimates

In order to further show the difference in the degree of fitting accuracy among algorithms, each group of algorithms was tested 50 times to obtain average chrominance accuracy. A histogram was drawn on the basis of the statistical table of experimental results. The abscissa is each algorithm, and the ordinate is the average chrominance accuracy under 50 experiments. The experimental results are shown in Figure 9.
Data in Figure 9 shows that the average prediction accuracy of the IAOA-DELM model is 96.07%, which is the maximum among all comparison models. Compared with AOA-DELM, GWO-DELM, WOA-DELM, and HHO-DELM, the average prediction accuracy of the IAOA-DELM model is increased by 0.63%, 0.99%, 1.27%, and 1.37%, respectively. This indicates that AOA has better searching ability, and improved AOA has higher searching accuracy than AOA. Compared with ORELM, ELM, RVFL, and BP, the average prediction accuracy of the IAOA-DELM model is improved by 6.96%, 7.54%, 8.00%, and 8.89%, respectively. The results indicate a significant improvement in accuracy with the proposed algorithm compared to the traditional single hidden layer network algorithm. Table 5 shows the running time cost of optimizing DELM by various swarm intelligent algorithms. The research findings demonstrate that the IAOA-DELM model proposed in this study achieves high accuracy while consuming less time. Although the IAOA-DELM incurs slightly higher time costs compared to the AOA-DELM, it delivers superior precision. Considering the significance of accuracy improvement in specific applications, compromising on time costs is justifiable when sufficient time is available to achieve enhanced precision.

4.2.3. Stability Analysis

The average angle error was recorded in the ten-fold cross verification, and the experimental results are shown in Figure 10. The more compact the box in the figure, the smaller the gap between the upper and lower quartiles, the more concentrated the distribution of experimental results, and the better the stability of the algorithm. The red line in the figure represents the median angle error. The lower the position of the red line, the smaller the overall prediction error of the algorithm. According to the angle error distribution analysis in Figure 10, the IAOA-DELM algorithm proposed in this research has the smallest prediction error. Its median average angle error is the lowest among all comparison groups. Its stability is higher than ORELM, ELM, RVFL, and BP.

4.2.4. Comparison of Image Correction Effect

To validate the algorithm’s correction ability, a series of images were captured in both air and water environments. One of the underwater images taken by the TL83 light source was imported into each comparison algorithm for illumination estimation and used the diagonal model to restore the image. The recovery results of each algorithm are shown in Figure 11. The corrected image effect of the IAOA-DELM algorithm proposed in this research is the closest to the image taken in the air by the standard light source D50. AOA-DELM, GWO-DELM, WOA-DELM, and HHO-DELM are not as accurate as the algorithms proposed in this research, but they have certain correction effects in the performance of image restoration, which indicates that the DELM model has a good illumination prediction effect. The illuminance estimation abilities of ORELM, ELM, RVFL, and BP are weak, and the suppression effect on the R channel is weak in image restoration. The corrected image retains the low color temperature to varying degrees.
After confirming the application advantages of IAOA-DELM compared with the other eight artificial neural network models, this research selected some classical color constancy algorithms and advanced color constancy algorithms to form a comparison group for comparison with the algorithm proposed in this research. The data set images were imported into each algorithm to obtain the angle errors of each group of algorithms, and the mean, median, upper, and lower quartiles of the angle errors of each group were calculated. The experimental results are presented in Table 6, while Figure 12 displays the image restoration outcomes.
The analysis of Table 6 reveals that the IAOA-DELM algorithm exhibits the smallest angle error in both the mean and lower quartile, the deviation of IAOA-DELM is small, and the stability is strong. Figure 12 clearly illustrates that the IAOA-DELM algorithm outperforms several classical and advanced color constancy algorithms in terms of performance.

5. Conclusions

This research proposes an iterative, chaotic, improved arithmetic optimization algorithm for deep extreme learning machines for the estimation of image illumination. Based on the experimental results, the following conclusions can be drawn:
(1)
By optimizing the initial search agents of the AOA algorithm using the iterative chaotic map, this study achieves a more uniform initialization of the population. Comparing the results of computational accuracy in terms of time cost (shown in Table 5) and average accuracy (depicted in Figure 9), it is evident that this method significantly improves the accuracy of color constancy computation at a minimal time cost;
(2)
By mapping the search agent fragments, the improved AOA algorithm optimizes the input weights and hidden biases of DELM. The stability analysis boxplot of algorithmic angular error, as presented in Figure 10, demonstrates that the improved AOA-DELM model exhibits a smaller angular error and good stability in color constancy;
(3)
Comparison of the correction results with selected classical color constancy algorithms and advanced color constancy algorithms, as illustrated in Figure 12, verifies that the underwater image corrected by the improved AOA-DELM model achieves the best visual effect, resembling images with no color error in the air.
In conclusion, this model can improve the efficiency of illumination correction of underwater images while maintaining a small-time cost, which can be considered an efficient underwater image color constancy model.

Author Contributions

Conceptualization, J.Y. and Q.Y.; methodology, J.Y.; software, Q.Y.; validation, J.Y., Q.Y. and S.C.; formal analysis, S.C.; investigation, J.Y.; resources, J.Y.; data curation, Q.Y.; writing—original draft preparation, Q.Y.; writing—review and editing, J.Y.; visualization, J.Y.; supervision, S.C.; project administration, D.Y.; funding acquisition, J.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research is funded by the National Key Research and Development Program of China (2022YFC2803903).

Data Availability Statement

The data that support the findings of this study are available on request from the corresponding author.

Acknowledgments

We are very grateful to Zhiyu Zhou from Zhejiang Sci-Tech University and Chao Wang from Hangzhou Dianzi University for their guidance on my research methods, and the Marine Center of Hangzhou Dianzi University for providing the experimental site and equipment.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Teague, J.; Willans, J.; Megson-Smith, D.A.; Day, J.C.; Allen, M.J.; Scott, T.B. Using Colour as a Marker for Coral ‘Health’: A Study on Hyperspectral Reflectance and Fluorescence Imaging of Thermally Induced Coral Bleaching. Oceans 2022, 3, 547–556. [Google Scholar] [CrossRef]
  2. Liu, B.; Liu, Z.; Men, S.; Li, Y.; Ding, Z.; He, J.; Zhao, Z. Underwater hyperspectral imaging technology and its applications for detecting and mapping the seafloor: A review. Sensors 2020, 20, 4962. [Google Scholar] [CrossRef] [PubMed]
  3. Hu, H.; Huang, Y.; Li, X. UCRNet: Underwater color image restoration via a polarization-guided convolutional neural network. Front. Mar. Sci. 2022, 9, 1031549. [Google Scholar] [CrossRef]
  4. Berman, D.; Treibitz, T.; Avidan, S. Diving into haze-lines: Color restoration of underwater images. In Proceedings of the British Machine Vision Conference (BMVC), London, UK, 4–7 September 2017; Volume 1, p. 2. [Google Scholar]
  5. Land, E.H. The retinex theory of color vision. Sci. Am. 1977, 237, 108–129. [Google Scholar] [CrossRef]
  6. Banić, N.; Lončarić, S. Improving the white patch method by subsampling. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Paris, France, 27–30 October 2014; pp. 605–609. [Google Scholar]
  7. Buchsbaum, G. A spatial processor model for object colour perception. J. Frankl. Inst. 1980, 310, 1–26. [Google Scholar] [CrossRef]
  8. Xu, L.L.; Jia, B.X. Automatic white balance based on gray world method and retinex. Appl. Mech. Mater. 2014, 462, 837–840. [Google Scholar] [CrossRef]
  9. Li, Q.Z.; Li, C.S.; Wang, Z.Q. Restoration algorithm for degraded underwater image based on wavelet transform. Comput. Eng. 2011, 37, 202–203. [Google Scholar]
  10. Yan, M.; Hu, Y.; Zhang, H. Color Constancy Based on Local Reflectance Differences. Electronics 2023, 12, 1396. [Google Scholar] [CrossRef]
  11. Iqbal, M.; Riaz, M.M.; Ali, S.S. Underwater image enhancement using laplae decomposition. IEEE Geosci. Remote Sens. Lett. 2020, 99, 1–5. [Google Scholar]
  12. Hassan, N.; Ullah, S.; Bhatti, N. The Retinex based improved underwater image enhancement. Multimed. Tools Appl. 2021, 80, 1839–1857. [Google Scholar] [CrossRef]
  13. Brainard, D.H.; Freeman, W.T. Bayesian color constancy. J. Opt. Soc. Am. A 1997, 14, 1393–1411. [Google Scholar] [CrossRef]
  14. Cardei, V.C.; Funt, B.; Barnard, K. Estimating the scene illumination chromaticity by using a neural network. J. Opt. Soc. Am. A 2002, 19, 2374–2386. [Google Scholar] [CrossRef] [Green Version]
  15. Xiong, W.; Funt, B. Estimating illumination chromaticity via support vector regression. J. Imaging Sci. Technol. 2006, 50, 341–348. [Google Scholar] [CrossRef] [Green Version]
  16. Li, B.; Xiong, W.; Xu, D. A supervised combination strategy for illumination chromaticity estimation. ACM Trans. Appl. Percept. (TAP) 2010, 8, 1–17. [Google Scholar] [CrossRef]
  17. Lou, Z.; Gevers, T.; Hu, N. Color Constancy by Deep Learning. In Proceedings of the British Machine Vision Conference (BMVC), Swansea, UK, 7–10 September 2015; pp. 71–76. [Google Scholar]
  18. Bianco, S.; Cusano, C.; Schettini, R. Color constancy using CNNs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Boston, MA, USA, 8–10 June 2015; pp. 81–89. [Google Scholar]
  19. Bianco, S.; Cusano, C.; Schettini, R. Single and multiple illuminant estimation using convolutional neural networks. IEEE Trans. Image Process. 2017, 26, 4347–4362. [Google Scholar] [CrossRef] [Green Version]
  20. Tang, Y.; Kang, X.; Li, C. Transfer learning for color constancy via statistic perspective. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 22 February–1 March 2022; Volume 36, pp. 2361–2369. [Google Scholar]
  21. Barron, J.T.; Tsai, Y.T. Fast fourier color constancy. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 886–894. [Google Scholar]
  22. Lo, Y.C.; Chang, C.C.; Chiu, H.C. Clcc: Contrastive learning for color constancy. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 8053–8063. [Google Scholar]
  23. Deng, C.W.; Huang, G.B.; Xu, J. Extreme learning machines: New trends and applications. Sci. China Inf. Sci. 2015, 2, 1–16. [Google Scholar] [CrossRef] [Green Version]
  24. Jiang, S.; Liu, H.; Lian, M. Rock slope displacement prediction based on multi-source information fusion and SSA-DELM model. Front. Environ. Sci. 2022, 10, 982069. [Google Scholar] [CrossRef]
  25. Qiao, L.; Jia, Z.; Cui, Y. Shear Sonic Prediction Based on DELM Optimized by Improved Sparrow Search Algorithm. Appl. Sci. 2022, 12, 8260. [Google Scholar] [CrossRef]
  26. Wang, X. User variable load forecasting based on FPA-DELM. J. Electrotechnol. Electr. Eng. Manag. 2022, 5, 48–55. [Google Scholar]
  27. Chong, H.Y.; Gortler, S.J.; Zickler, T. The von Kries hypothesis and a basis for color constancy. In Proceedings of the IEEE 11th International Conference on Computer Vision, Rio de Janeiro, Brazil, 14–21 October 2007; pp. 1–8. [Google Scholar]
  28. Abualigah, L.; Diabat, A.; Mirjalili, S. The arithmetic optimization algorithm. Comput. Methods Appl. Mech. Eng. 2021, 376, 113609. [Google Scholar] [CrossRef]
  29. Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: A new learning scheme of feedforward neural networks. IEEE Int. Jt. Conf. Neural Netw. 2004, 2, 985–990. [Google Scholar]
  30. Wazi, M.T.; Ali, D.S.; Al-Saidi, N.M.G. A secure image cryptosystem via multiple chaotic maps. Discret. Math. Algorithms Appl. 2022, 14, 2150141. [Google Scholar] [CrossRef]
  31. Weijer, J.; Gevers, T.; Gijsenij, A. Edge-based color constancy. IEEE Trans. Image Process. 2007, 16, 2207–2214. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Vereshchagin, N.K.; Vitányi, P.M.B. Kolmogorov’s structure functions and model selection. IEEE Trans. Inf. Theory 2004, 50, 3265–3290. [Google Scholar] [CrossRef]
  33. Zhang, K.; Luo, M. Outlier-robust extreme learning machine for regression problems. Neurocomputing 2015, 151, 1519–1527. [Google Scholar] [CrossRef]
  34. Majumder, I.; Dash, P.K.; Bisoi, R. Short-term solar power prediction using multi-kernel-based random vector functional link with water cycle algorithm-based parameter optimization. Neural Comput. Appl. 2020, 32, 8011–8029. [Google Scholar] [CrossRef]
  35. Afifi, M.; Brown, M.S. Interactive White Balancing for Camera-Rendered Images. In Proceedings of the IS&T 28th Color and Imaging Conference, Online, 4–19 November 2020; pp. 136–141. [Google Scholar]
  36. Afifi, M.; Price, B.; Cohen, S. When color constancy goes wrong: Correcting improperly white-balanced images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1535–1544. [Google Scholar]
  37. Afifi, M.; Brown, M.S. What else can fool deep learning? Addressing color constancy errors on deep neural network performance. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 243–252. [Google Scholar]
  38. Finlayson, G.D.; Trezzi, E. Shades of gray and colour constancy. Color and Imaging Conference. Soc. Imaging Sci. Technol. 2004, 1, 37–41. [Google Scholar]
Figure 1. DELM Network structure.
Figure 1. DELM Network structure.
Electronics 12 03174 g001
Figure 2. Search agent mapping flowchart.
Figure 2. Search agent mapping flowchart.
Electronics 12 03174 g002
Figure 3. IAOA-DELM algorithm flowchart.
Figure 3. IAOA-DELM algorithm flowchart.
Electronics 12 03174 g003
Figure 4. Experimental facility.
Figure 4. Experimental facility.
Electronics 12 03174 g004
Figure 5. IAOA-DELM light correction process diagram.
Figure 5. IAOA-DELM light correction process diagram.
Electronics 12 03174 g005
Figure 6. The average chrominance accuracy under different population sizes and number of iterations when alpha = 0.25.
Figure 6. The average chrominance accuracy under different population sizes and number of iterations when alpha = 0.25.
Electronics 12 03174 g006
Figure 7. The average chrominance accuracy under different population sizes and number of iterations when alpha = 0.5.
Figure 7. The average chrominance accuracy under different population sizes and number of iterations when alpha = 0.5.
Electronics 12 03174 g007
Figure 8. The average chrominance accuracy under different population sizes and number of iterations when alpha = 0.25.
Figure 8. The average chrominance accuracy under different population sizes and number of iterations when alpha = 0.25.
Electronics 12 03174 g008
Figure 9. Comparison of average accuracy results between IAOA-DELM algorithm and comparison group algorithm.
Figure 9. Comparison of average accuracy results between IAOA-DELM algorithm and comparison group algorithm.
Electronics 12 03174 g009
Figure 10. Algorithm angle error stability analysis box diagram.
Figure 10. Algorithm angle error stability analysis box diagram.
Electronics 12 03174 g010
Figure 11. Correction results of different illumination correction models (A) Underwater image to be corrected under TL83 light source (B) Standard image in air under the same scene under D50 light source (C) IAOA-DELM (D) AOA-DELM (E) GWO-DELM (F) WOA-DELM (G) HHO-DELM (H) ORELM (I) ELM (J) RVFL (K) BP.
Figure 11. Correction results of different illumination correction models (A) Underwater image to be corrected under TL83 light source (B) Standard image in air under the same scene under D50 light source (C) IAOA-DELM (D) AOA-DELM (E) GWO-DELM (F) WOA-DELM (G) HHO-DELM (H) ORELM (I) ELM (J) RVFL (K) BP.
Electronics 12 03174 g011
Figure 12. Comparison with other classical and advanced image correction algorithms (A) Underwater image to be corrected under TL83 light source (B) Standard image in air under the same scene under D50 light source (C) IAOA-DELM (D) Interactive WB Method (E) Data-Driven WB Method (F) WB color augmenter (G) Grey-World (H) Max-RGB (I) Shades of Grey.
Figure 12. Comparison with other classical and advanced image correction algorithms (A) Underwater image to be corrected under TL83 light source (B) Standard image in air under the same scene under D50 light source (C) IAOA-DELM (D) Interactive WB Method (E) Data-Driven WB Method (F) WB color augmenter (G) Grey-World (H) Max-RGB (I) Shades of Grey.
Electronics 12 03174 g012
Table 1. The light source used in this article.
Table 1. The light source used in this article.
Light Source NameColor TemperatureColor IndexType of Light Source
TL833000 K85European commercial fluorescence
U353500 K85American commercial fluorescence
TL844000 K85European commercial fluorescence
CWF4150 K62American commercial fluorescence
D505000 K95+Filtered tungsten lamp
D656500 K95+Filtered tungsten lamp
Table 2. The influence of the number of different hidden layers in DELM on the experimental results (The best values are shown in bold).
Table 2. The influence of the number of different hidden layers in DELM on the experimental results (The best values are shown in bold).
Number of Hidden LayersAverage AccuracyStandard DeviationMaximum AccuracyMinimum AccuracyMedian AccuracyTime Cost (s)
194.47%0.031896.82%88.91%94.47%0.1513
295.08%0.011896.69%91.56%95.31%0.2624
389.71%0.008392.49%88.21%89.54%0.3699
489.32%0.007191.21%88.03%89.23%0.4801
589.21%0.005990.44%87.95%89.31%0.5796
689.27%0.005690.54%88.01%89.24%0.6649
788.68%0.007390.52%87.03%88.62%0.7578
Table 3. E d of different activation functions.
Table 3. E d of different activation functions.
Activation FunctionSigmoidSineHardlimSignTribas
E d 0.03870.04650.04150.04040.0541
Table 4. Parameter setting of extreme learning machine and its optimization algorithm.
Table 4. Parameter setting of extreme learning machine and its optimization algorithm.
AlgorithmParameterValue
AOA-DELMNumber of populations30
I t e r M a x 100
P M a x 1
P M i n 0.2
Alpha5
GWO-DELMNumber of populations30
I t e r M a x 100
WOA-DELMNumber of populations30
I t e r M a x 100
HHO-DELMNumber of populations30
I t e r M a x 100
ORELM [33]C230
ELMNumberofHiddenNeurons600
RVFL [34]Option.mode2
Option.Scale1
Option.ActivationFunction“sig”
Option.Scalemode3
BPNet.trainParam.epochs100
Net.trainParam.Ir0.1
Net.trainParam.goal0.001
Abbreviations: AOA, Arithmetic Optimization Algorithm; GWO, Grey Wolf Optimization Algorithm; WOA, Whale Optimization Algorithm; HHO, Harris Hawks Optimization Algorithm; ORELM, Extreme Learning Machine with Outlier Robustness; ELM, Extreme Learning Machine; RVFL, Random Vector Function Link Network; BP, Back Propagation.
Table 5. The running time cost of each algorithm.
Table 5. The running time cost of each algorithm.
IAOA-DELMAOA-DELMGWO-DELMWOA-DELMHHO-DELM
Time cost (s)18.3117.9728.4629.7341.45
Table 6. Correction results of different illumination correction models (Minimum values are shown in bold).
Table 6. Correction results of different illumination correction models (Minimum values are shown in bold).
AlgorithmMeanMedBest 25%Worst 25%
Interactive WB Method [35]3.633.460.946.94
Data-Driven WB Method [36]3.493.040.875.86
WB color augmenter [37]3.183.530.796.37
Grey-World [7]7.817.341.629.78
White-Patch [5]6.145.761.039.31
Shades of Grey [38]5.124.861.129.76
IAOA-DELM (ours)2.813.120.914.32
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, J.; Yu, Q.; Chen, S.; Yang, D. Underwater Image Color Constancy Calculation with Optimized Deep Extreme Learning Machine Based on Improved Arithmetic Optimization Algorithm. Electronics 2023, 12, 3174. https://doi.org/10.3390/electronics12143174

AMA Style

Yang J, Yu Q, Chen S, Yang D. Underwater Image Color Constancy Calculation with Optimized Deep Extreme Learning Machine Based on Improved Arithmetic Optimization Algorithm. Electronics. 2023; 12(14):3174. https://doi.org/10.3390/electronics12143174

Chicago/Turabian Style

Yang, Junyi, Qichao Yu, Sheng Chen, and Donghe Yang. 2023. "Underwater Image Color Constancy Calculation with Optimized Deep Extreme Learning Machine Based on Improved Arithmetic Optimization Algorithm" Electronics 12, no. 14: 3174. https://doi.org/10.3390/electronics12143174

APA Style

Yang, J., Yu, Q., Chen, S., & Yang, D. (2023). Underwater Image Color Constancy Calculation with Optimized Deep Extreme Learning Machine Based on Improved Arithmetic Optimization Algorithm. Electronics, 12(14), 3174. https://doi.org/10.3390/electronics12143174

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop