-
We used the super-spheroid as a simplified model to mimic atmospheric particles because it has been demonstrated as a promising model for the optical properties of dust and sea salt aerosols (Bi et al., 2018; Lin et al., 2018), although realistic particle shapes could be much more complex than super-spheroids. The super-spheroidal equation is defined as follows (Barr, 1981; Wriedt, 2002):
where
$a$ and$c$ are the length of the horizontal and vertical axes in a Cartesian coordinate system, and n is the roundness parameter, which determines whether a particle is sharp or smooth. This model can generate a large number of particles by changing the aspect ratio and roundness parameter. Figure 1 demonstrates twenty-five typical shapes of super-spheroids. Importantly, the shape variation is continuous so that the shape parameters can be directly used as training parameters in a DNN.Figure 1. Twenty-five representative super-spheroid shapes including oblate (
$a/c > 1.0$ ), prolate ($a/c < 1.0$ ), concave ($n > 2.0$ ), and convex particles ($n < 2.0$ ).The optical properties include the extinction efficiency factor
${Q_{{\text{ext}}}}$ , the single scattering albedo SSA, the asymmetry factor g, and the phase matrix element${P_{ij}}$ . The extinction efficiency factor${Q_{{\text{ext}}}}$ can then be defined as:where
${C_{{\text{ext}}}}$ is the extinction cross-section,${A_{{\text{proj}}}}$ is the projected cross-sectional area of the super-spheroidal particle, xsize represents the size parameter of the particle, and factor is determined by the aspect ratio$a/c$ and the roundness parameter n, which are computed by a Monte-Carlo ray-tracing technique. Specifically, for each particle orientation, the projected area was determined by counting the proportion of "photons" falling within the projected shadow. The averaged projected area was obtained by taking the average of projected areas associated with several orientations. SSA is defined as:in which
${C_{{\text{sca}}}}$ is the scattering cross-section. The asymmetry factor g represents a measure of the angular distribution of the scattering intensity. The scattering matrix${P_{ij}}$ determines the relationship between the Stokes vectors of the incident beam$\left( {{{\boldsymbol{I}}_0},{{\boldsymbol{Q}}_0},{{\boldsymbol{U}}_0},{{\boldsymbol{V}}_0}} \right)$ and that of the scattered beam$\left( {{{\boldsymbol{I}}_{\text{s}}},{{\boldsymbol{Q}}_{\text{s}}},{{\boldsymbol{U}}_{\text{s}}},{{\boldsymbol{V}}_{\text{s}}}} \right)$ , which is given by (Bohren and Huffman, 1983):where
$\theta $ is the scattering angle, ranging from 0° to 180°, k is the corrected wave number, and R is the distance between the scatter and the detector. The intensity of the light beam is given by${\boldsymbol{I}}$ , while${\boldsymbol{Q}}$ ,${\boldsymbol{U}}$ , and${\boldsymbol{V}}$ describe the linear and circular polarization of the beam. For randomly oriented super-spheroids (note, the particles have a plane of symmetry),${P_{12}}{\text{ = }}{P_{21}}$ , and$ {P_{34}} = - {P_{43}} $ (Van De Hulst, 1981; Mishchenko and Yurkin, 2017). -
Deep learning is one of the latest trends in Machine Learning (ML) and Artificial Intelligence (AI) research. A deep neural network has good self-adaptation capabilities. It is trained through a batch of corresponding input and output data, allowing for the intrinsic relationship between the input and output to be resolved in the final form of a complex non-linear system function (Schmidhuber, 2017). In most situations, a DNN architecture is a multi-level, non-linear repetition of a simple neuron that obtains highly intricate functions from the input. This kind of network is composed of simple neurons in a complex system that processes information through the relationship between the internal nodes. In this study, the procedure mentioned above in the context of artificial neural networks is essentially a non-linear statistical regression. The hidden layers containing some neurons constitute a deep neural network. The output results
${{\boldsymbol{o}}^{(l)}}$ of neurons in the layer l was calculated from the output of the previous layers${{\boldsymbol{o}}^{(l - 1)}}$ :where
$\varphi$ is a non-linear activation function,${N^{(l - 1)}}$ is the sum of nodes in layer$l - 1$ ,${\boldsymbol{w}}_{i,j}^{(l)}$ and${\boldsymbol{b}}_i^{(l)}$ are the weights and biases in layer l, respectively. The node i in layer l must satisfy the relation$1 \leqslant i \leqslant {N^{(l)}}$ .The parametric scheme mentioned in section 2.1 makes a DL scheme feasible. All of the parameters (shape parameters, refractive indices, and size parameters) are continuous and smooth. Compared to particle shape models without continuous particle parameters, super-spheroid models in conjunction with DL schemes are very suitable for establishing relationships between the input (particle shapes, refractive indices, and particle sizes) and the output (the optical properties). Once the relationships are established through DL schemes, the optical properties of particles in the atmosphere can be directly obtained by the neural networks.
The neural network method is an optimal choice to compress the database. In principle, if we want to preserve a neural network, we can do it by saving the weights and biases of the neurons in each of the hidden layers. As the number of parameters increases, so does the storage size of the model file. Then, we can just use the trained and preserved DNN models to predict the optical properties of specific particles. To use the DNN model instead of the original database, we must achieve the following goals: (1) obtain a file size of a saved DNN model that is much smaller than that of the database size, and (2) design a model that can predict particle properties with high accuracy. The value of the learning rate in the Adam algorithm and the patience parameter of early stopping should be tuned in the training procedure (details in section 2.4). In addition, various weight initialization methods and batch sizes of input data also need to be adjusted.
-
The database chosen in this study contains abundant non-spherical particles based on the super-spheroidal model. The optical properties of super-spheroids are computed by using the IITM for all combinations of shape parameters, the complex refractive index, and the size parameter. The aspect ratio
$a/c$ ranges from 0.5 to 2.0, and the roundness parameter n ranges from 1.2 to 3.0. A total of 110 super-spheroid shapes are considered.The refractive index is complex, and its real part
${m_{\text{r}}}$ ranges from 1.30 to 1.80, while the imaginary part${m_{\text{i}}}$ ranges from 10–7 to 0.1. The dataset is computed using 11 real parts and 9 imaginary parts (99 complex refractive indices in all). In the single-scattering computation, the size parameter xsize ranges from 0.1 to 50 with 180 discrete values, and the scattering angle$\theta $ ranges from 0° to 180° with 721 discrete values. For convenience, the parameters of the database are explicitly given in Table 1.Parameters Values Optical Properties Real Part of the Refractive Index ${m_{\text{r}}}$ 1.30, 1.35, 1.40, 1.45, 1.50, 1.55, 1.60, 1.65, 1.70, 1.75, 1.80 ${Q_{{\text{ext}}}}$, ${\rm{SSA}}$, g, ${P_{11}}$, ${P_{12}}$, ${P_{22}}$, ${P_{33}}$, ${P_{43}}$, and ${P_{44}}$ Imaginary Part of the Refractive Index ${m_{\text{i}}}$ 10−7, 10−6, 10−5, 0.0001, 0.001, 0.005, 0.01, 0.05, 0.1 Aspect Ratio $a/c$ 0.5, 0.6, 0.7, 0.8, 0.9, 1.0, 1.2, 1.4, 1.6, 1.8, 2.0 Roundness n 1.2, 1.4, 1.6, 1.8, 2.0, 2.2, 2.4, 2.6, 2.8, 3.0 Size Parameter xsize 0.1, 0.2, 0.3, …, 9.8, 9.9, 10 (step = 0.1);
10.2, 10.4, 10.6, …, 19.6, 19.8, 20 (step = 0.2);
21, 22, 23, …, 48, 49, 50 (step = 1);Scattering Angle $\theta $(°) 0, 0.25, 0.5, 0.75, 1.0, 1.25, 1.5, 1.75, 2.0, 2.25, 2.5, 2.75, 3…179, 179.25, 179.5, 179.75, 180 (step = 0.25) Table 1. Super-spheroid optical properties database. The real part of the refractive index
${m_{\text{r}}}$ ranges from 1.30 to 1.80, while the imaginary part${m_{\text{i}}}$ ranges from 10–7 to 0.1. The aspect ratio$a/c$ ranges from 0.5 to 2.0, and the roundness parameter n ranges from 1.2 to 3.0. The size parameter xsize ranges from 0.1 to 50, and the scattering angle$\theta $ ranges from 0° to 180°. The optical properties include the extinction efficiency factor${Q_{{\text{ext}}}}$ , single scattering albedo SSA, asymmetry factor g, and all phase matrix elements${P_{ij}}$ .The optical properties include the extinction efficiency factor
${Q_{{\text{ext}}}}$ , single scattering albedo SSA, asymmetry factor g, and all phase matrix elements${P_{ij}}$ . The results in the database computed from the IITM are used as truth values when training and testing the DNN models. The total amount of data in the set is 1 960 200 for the optical properties${Q_{{\text{ext}}}}$ , SSA, and g. For the phase matrix element${P_{ij}}$ , the dataset contains 1 413 304 200 combinations because the scattering angle$\theta $ is involved.The optical data of non-spherical particles can be obtained immediately from the database, so it is unnecessary to solve Maxwell's equations, and computation speed is no longer a problem. However, as more and more particles or parameters are considered, the size of the database increases substantially. The parameters given in Table 1 would take up almost 133 377 877 kilobytes (~127 GB) of hard disk storage space for saving the data in ASCII format. Thus, it is inconvenient to apply the dataset in scientific research. In addition, if the shape parameters or refractive indices are unavailable, we have to read the data from the database and then perform necessary interpolations.
-
DNN Models are used to predict the extinction efficiency factor
${Q_{{\text{ext}}}}$ , single scattering albedo SSA, asymmetry factor g, and phase matrix element${P_{ij}}$ . These models consist of several fully connected layers. The input layer has five inputs (${m_{\text{r}}}$ ,${m_{\text{i}}}$ ,$a/c$ ,$n$ , xsize) for the${Q_{{\text{ext}}}}$ , SSA, and g and six inputs (${m_{\text{r}}}$ ,${m_{\text{i}}}$ ,$a/c$ ,$n$ , xsize,$\theta $ ) for the phase matrix elements${P_{ij}}$ .Figure 3. Deep Neural Network (DNN) architectures for
${Q_{{\text{ext}}}}$ , SSA, g, and${P_{ij}}$ . The number of layers is 5, 5, 6, and 7 for${Q_{{\text{ext}}}}$ , SSA, g, and${P_{ij}}$ , respectively. The number of neuron nodes in the layer is included in the parentheses. For example,$\left( {{Q_{{\text{ext}}}},{\text{ }}128} \right)$ indicates 128 neuron nodes in a layer for the prediction of${Q_{{\text{ext}}}}$ .The number of hidden layers and neurons varies for different optical properties. For example, for the extinction efficiency (
${Q_{{\text{ext}}}}$ ), five layers were designed. For convenience, the number of neuron nodes in each layer was included in parenthesis. For example,$\left( {{Q_{{\text{ext}}}},{\text{ }}128} \right)$ indicates that 128 neuron nodes in this layer were used for the prediction of${Q_{{\text{ext}}}}$ . Based on several tests (see section 3.1), we found that accurate prediction of SSA, g, and${P_{ij}}$ requires five, six, and seven layers, respectively. Among the nodes of hidden layers, every activation function was a rectified linear unit (ReLU) function, namely,$f(x) = \max (0,{\text{ }}x)$ . The chosen activation functions could build a sparse neural network and prevent over-fitting (Glorot et al., 2011). Moreover, a higher computational efficiency was achieved because no complex processes, such as exponential calculations, were involved in the ReLU function. However, we applied a linear mapping on the last layer instead of the ReLU function.The networks were trained with the Adam algorithm, a stochastic gradient descent method based on an adaptive estimation of first-order and second-order moments (Kingma and Ba, 2015). The initial learning rate was set to 0.001, the exponential decay rates for the 1st- and 2nd-moment estimates were set to 0.9 and 0.999, respectively, and the epsilon parameter (a constant for controlling the numerical stability) was set to 10–7. All weighting coefficients were initialized by the Glorot uniform initializer (Glorot and Bengio, 2010).
We implemented the DNN networks using the Keras API running on the TensorFlow2 ML platform (Abadi et al., 2016). The key tool is the parallel computing framework CUDA, which makes the training process efficient. The calculations were performed on an NVIDIA GeForce RTX 3070 GPU with 16 GB of memory. We saved the trained DNN model parameters in HDF5 file format. The optimal weight matrices (kernel) and bias vectors corresponding to each layer were saved in these files.
All data were split into a training set, a validation set, and a test set. The training set is used to fit the samples and change parameters between the adjacent units. The validation set is used to tune the hyperparameters of the models, save the optimal parameters of nodes in every epoch, and determine an early stopping point when the monitored metric stops improving. The test set is only used to evaluate the error and accuracy of the resulting model at the end. However, the amount of data in each set varies among different evaluation schemes.
We have designed three schemes for specific purposes (the details are in section 3). We used 70% of the data for training, 10% for validation, and 20% for testing. We selected 100 000 points or specific particles from the test set to verify the different targets. Before training, all input parameters were normalized to between 0 and 1.
To quantify the accuracy of the DNN predicted results, we used the root mean square error (RMSE) as a regression metric, which is defined as follows:
where,
${y_i}$ and${f_i}$ represent the i-th true value and the predicted value in the data samples, respectively, and M is the sample size. Once the validation loss decreases, the whole model is saved after every epoch. Thus, the model or weights can be loaded afterward to continue the training from the saved state. Training can be stopped early when the loss has stopped decreasing. We set the patience parameter to 20, which means that the training process is stopped if the validation loss does not decrease after 20 epochs. The patience parameter is also a sensible choice to avoid overfitting. -
To achieve maximum compression for the database, we designed several DNN models, which resulted in different file sizes. The model size depends on the weights and bias in the hidden layers, so we controlled the network parameters by adjusting the numbers of nodes and layers. Table 2 shows 35 network architectures that were established for the prediction of optical properties. The total parameters of every network range from 102 to 106, and the model file sizes range from 25 kB to 104 kB.
The nodes of layers Total network parameters Model file size (KB) 5, 16, 1 113 25 5, 32, 1 225 27 5, 64, 1 449 29 5, 8, 8, 1 129 32 5, 128, 1 897 35 5, 16, 16, 1 385 36 5, 16, 16, 16, 1 657 45 5, 32, 32, 1 1281 46 5, 256, 1 1793 47 5, 16, 16, 16, 16, 1 929 55 5, 64, 32, 1 2497 61 5, 32, 32, 32, 1 2337 66 5, 32, 32, 32, 32, 1 3393 84 5, 64, 64, 1 4609 86 5, 64, 64, 16, 1 5601 103 5, 64, 64, 32, 1 6657 116 5, 32, 32, 32, 32, 32, 1 5505 124 5, 128, 64, 1 9089 139 5, 64, 64, 64, 1 8769 142 5, 128, 64, 64, 1 13249 193 5, 128, 128, 1 17409 236 5, 64, 128, 64, 64, 1 21185 294 5, 128, 128, 64, 1 25601 338 5, 128, 128, 128, 1 33921 436 5, 64, 128, 128, 128, 64, 1 50049 640 5, 128, 256, 128, 1 66817 821 5, 256, 256, 128, 1 100353 1215 5, 128, 256, 256, 128, 1 132609 1600 5, 256, 256, 256, 1 133377 1602 5, 256, 256, 256, 256,1 199169 2380 5, 256, 256, 256, 256, 128, 1 231937 2771 5, 256, 256, 256, 256, 256, 1 264961 3159 5, 256, 256, 256, 256, 256, 256, 1 330753 3936 5, 128, 256, 256, 256, 256, 256, 256, 256, 256, 128, 1 527361 6272 5, 256, 256, 256, 256, 256, 512, 512, 256, 256, 256, 256, 256, 1 1053697 12456 Table 2. Network structures that are established for the prediction of optical properties. The number of network parameters ranges from ~102 to ~106, and the model file size ranges from 25 KB to ~104 KB.
To assess the performance of the networks, we used the coefficient of determination R2 to evaluate the fitness between predictions and true values. The coefficient of determination is defined as follows:
$\overline y $ is the average of all true values.$\mathrm{}$ The closer R2 is to 1, the higher the performance of the network is.We tested all of the established models, and Fig. 2 shows the RMSE and R2 of each network. The test data was fixed to 100 000 sets, which were chosen randomly from the testing dataset. The figure shows that the RMSE and R2 from the testing sets reached a limit when the network parameters increased. The DNN models attained appreciable performance for predicting SSA, and g when the network parameters were of the order 104 to 105. At the same time, performance was optimized for
${P_{11}}$ when the parameters were of the order 105 to 106. Benefits were no longer realized with further increases in the number of total parameters. In the testing process, we paid more attention to the total number of network parameters and did not specifically consider the influence of hidden layers and nodes of each layer on the network.Figure 2. Scatter plot of RMSE (orange dots) and the coefficient of determination
${R^2}$ (green dots) from several established DNN models for (a)${Q_{{\text{ext}}}}$ , (b) SSA, (c) g, and (d)${P_{11}}$ . In each subfigure, the left y-axis represents the RMSE, and the right y-axis represents${R^2}$ . As the network parameters increase, the DNN models have an upper limit of the prediction accuracy for each optical property.To replace the database with an optimized DNN model, we must consider both the model's file size and accuracy. Therefore, we chose a DNN model with minimal network parameters for each optical property and ensured that R2 is greater than 0.999. The optimal model’s storage size was different for the various optical properties. The size of the DNN files was 193 KB for
${Q_{{\text{ext}}}}$ , 142 KB for SSA, 294 KB for g, and 3159 KB for the phase matrix elements${P_{ij}}$ . The architecture of selected networks is illustrated in Fig. 3. -
To verify the optimized DNN model, we evaluated its performance in optical property prediction. We trained the DNN for target particle optical properties separately, as described in section 2.4. The verification was done with testing on 1) 100 000 sets that were chosen from the testing dataset randomly and 2) specific particles for which the optical properties are available or unavailable in the database. First, we randomly chose many discrete points from the testing dataset and then examined the fitness between the DNN predictions and the true values. This procedure can give a general picture of the model's overall performance. The number of data points is 100 000 for each optical property.
In Fig. 4, the top row shows that the R2 values of
${Q_{{\text{ext}}}}$ , SSA, and g, are 0.99991, 0.99988, and 0.99985, and the RMSEs are 0.0091, 0.0024, and 0.0029, respectively. The performance of the optimized models for the phase matrix elements${P_{{ij}}}$ is also illustrated in the remaining two rows in Fig. 4. The coefficient of determination, R2, is greater than 0.999 for the phase function${P_{11}}$ and 0.99 for the other phase matrix elements. Therefore, the chosen DNN model has good learning capability. The original large database can be compressed into several small neural network models without losing much accuracy, and the compression ratio is approximately 6810 (from 133 377 877 kB to 19 583 kB).Figure 4. Scatter plot of (a)
${Q_{{\text{ext}}}}$ , (b) SSA, (c) g, (d)${P_{11}}$ , and (e) – (i)${P_{ij}}/{P_{11}}$ by DNN predictions against the true values.Next, we compared the optical properties predicted from the DNN and the database values. Figure 5 shows the absolute errors of
${Q_{{\text{ext}}}}$ calculated by the neural network and the database values. The results for SSA and g are shown in Figs. 6 and 7, respectively. Figures 5–7 present a broad picture of the model performance in reproducing the optical properties for particles available in the database. By identifying the upper and lower limits of errors, we can assess the model's ability to predict optical properties.Figure 5. Absolute errors of
${Q_{{\text{ext}}}}$ predicted by the DNNs and the true values from the database.Table 3 lists the particle parameters chosen from the lighter colored areas (smaller errors) and darker colored areas (larger errors) in Figs. 5–7 to better understand the model's accuracy. Figures 8 (a–c) shows
${Q_{{\text{ext}}}}$ , SSA, and g as functions of the size and shape parameters chosen from the lighter (smaller error) regions. The DNN and IITM results match almost perfectly, and the mean absolute errors (MAEs) are 0.003456, 0.001317, and 0.000876 for${Q_{{\text{ext}}}}$ , SSA, and g, respectively.Optical properties Particle parameters of lighter regions Particle parameters of darker regions ${Q_{{\text{ext}}}}$ ${m_{\text{r}}}$=1.40, ${m_{\text{i}}}$=10–6, $a/c$ =1.2, and $n$=2.6 ${m_{\text{r}}}$=1.30, ${m_{\text{i}}}$=10–6, $a/c$ =0.9, and $n$=1.4 SSA ${m_{\text{r}}}$=1.70, ${m_{\text{i}}}$=0.1, $a/c$ =1.8, and $n$=2.0 ${m_{\text{r}}}$=1.50, ${m_{\text{i}}}$=0.005, $a/c$ =1.0, and $n$=2.0 g ${m_{\text{r}}}$=1.45, ${m_{\text{i}}}$=0.005, $a/c$ =1.6, and $n$=1.6 ${m_{\text{r}}}$=1.60, ${m_{\text{i}}}$=0.005, $a/c$ =1.0, and $n$=2.4 Figure 8. Comparison of
${Q_{{\text{ext}}}}$ , SSA, and g predicted by the DNNs against the true values in the database. The upper row (a–c) in the figure shows the results for parameters in light-colored areas, and the lower row (d–f) shows the results for parameters in dark-colored areas (Figs. 5–7). The mean absolute errors (MAEs) are also included in each subplot.Particle parameters Optical properties Case 1 ${m_{\text{r}}}$=1.45, ${m_{\text{i}}}$=10-6, $a/c$=0.6, $n$=2.4, and xsize=29.5 ${P_{11}}$, ${P_{12}}/{P_{11}}$, ${P_{22}}/{P_{11}}$, ${P_{33}}/{P_{11}}$, ${P_{43}}/{P_{11}}$, and ${P_{44}}/{P_{11}}$ Case 2 ${m_{\text{r}}}$=1.33, ${m_{\text{i}}}$=0.008, $a/c$ =0.6, and $n$=2.4, and xsize=38 ${P_{11}}$, ${P_{12}}/{P_{11}}$, ${P_{22}}/{P_{11}}$, ${P_{33}}/{P_{11}}$, ${P_{43}}/{P_{11}}$, and ${P_{44}}/{P_{11}}$ Case 3 ${m_{\text{r}}}$=1.33, ${m_{\text{i}}}$=0.008, $a/c$ =0.6, and $n$=2.4, and xsize=29.5 ${P_{11}}$, ${P_{12}}/{P_{11}}$, ${P_{22}}/{P_{11}}$, ${P_{33}}/{P_{11}}$, ${P_{43}}/{P_{11}}$, and ${P_{44}}/{P_{11}}$ Table 4. The test parameters of unavailable particles in the database. In case 1, the size parameter xsize is fixed at 29.5. In case 2, the real part of the refractive index
${m_{\text{r}}}$ is fixed at 1.33, and the imaginary part${m_{\text{i}}}$ is 0.008, and both of them are unavailable in the database. Case 3 is a “triple interpolation” scheme, and${m_{\text{r}}}$ ,${m_{\text{i}}}$ , and xsize are not contained in the database.The lower panels of Figs. 8 (d–f) are similar to the upper panels but show the results from the darker (larger error) regions, including the darkest dot. The MAEs for
${Q_{{\text{ext}}}}$ , SSA, and g are 0.006313, 0.001523, and 0.000945, respectively. Although relatively large errors were identified in darker regions, the predicted curves are still extremely consistent with the true values. The performance shown in Figs. 5–8 indicates that the DNN models can accurately calculate the optical properties of the particles contained in the whole database.Next, we examined the learning capability of the selected DL scheme for cases with some parameters that are not in the datasets. Such an examination shows the robustness of the models by using what it has already learned to predict a new situation. In the database (as shown in Table 1), the real part of the refractive index
${m_{\text{r}}}$ has values of 1.30 to 1.80 with a step size of 0.05, the imaginary part of the refractive index${m_{\text{i}}}$ has values above 0.001 to 0.01 with a step size of 0.005 or 0.05. The size parameter xsize has integer values of 20 to 50 with a step size of 1. The unknown particle parameters that we tested are listed in Table 4.To examine the accuracy of the predicted optical properties from the DNN, we also used the IITM to calculate the optical properties of these particles as reference values. The phase matrix element
${P_{ij}}/{P_{11}}$ oscillates with respect to the scattering angle (${P_{12}}/{P_{11}}$ and${P_{43}}/{P_{11}}$ in particular). Nevertheless, a good agreement can still be obtained between the DNN and IITM results for all phase matrix elements. The numerical errors are expected to be further reduced if a bulk-scattering matrix is computed.Figure 9 shows a comparison of the DNN prediction and the IITM results for case 1. The real part of the refractive index
${m_{\text{r}}}$ is 1.45, the imaginary part${m_{\text{i}}}$ is 10–6, the aspect ratio$a/c$ is 0.6, the roundness parameter n is 2.4, and the size parameter xsize is 29.5. Yet again, there is excellent agreement between the DNNs and IITM, suggesting that the selected DL scheme is capable of predicting patterns given the values of parameters in the interpolation region. The most encouraging part is the remarkable increase in computing speed. Consider that when the size parameter (xsize) is fixed at 29.5, our models calculate all phase matrix elements in near-real-time (the average time is 1.2674 seconds), duly noting that the cost is more than 5 minutes if applying IITM to calculate them, especially if 24 central processing units are used.Figure 9. Comparison of (a)
${P_{11}}$ , (b)${P_{12}}/{P_{11}}$ , (c)${P_{22}}/{P_{11}}$ , (d)${P_{33}}/{P_{11}}$ , (e)${P_{43}}/{P_{11}}$ , and (f)${P_{44}}/{P_{11}}$ predicted by the DNNs against the true values. The size parameter xsize is unavailable in the database.Figure 10 shows the results for case 2, where the refractive index is also not in the training or testing grids (both the real part
${m_{\text{r}}}$ and imaginary part${m_{\text{i}}}$ ). In this computation, the real part of the refractive index${m_{\text{r}}}$ is 1.33, the imaginary part is 0.008, the aspect ratio$a/c$ is 0.6, the roundness parameter$n$ is 2.4, and the size parameter xsize is 38. These results imply that DNN models can predict the properties of a particle when the refractive index is not included in the database. Therefore, we can conclude that the DL scheme not only can significantly compress the whole database but it can also compute the optical properties of super-spheroids with reasonable accuracy for parameters that are not covered by the database.Figure 10. Comparison of (a)
${P_{11}}$ , (b)${P_{12}}/{P_{11}}$ , (c)${P_{22}}/{P_{11}}$ , (d)${P_{33}}/{P_{11}}$ , (e)${P_{43}}/{P_{11}}$ , and (f)${P_{44}}/{P_{11}}$ predicted from the DNNs against the true values. The refractive index is unavailable in the database.Case 3 represents a “triple interpolation” scheme (for
${m_{\text{r}}}$ ,${m_{\text{i}}}$ , and xsize). In this case, the real part of the refractive index${m_{\text{r}}}$ is 1.33, the imaginary part is 0.008, the aspect ratio$a/c$ is 0.6, the roundness parameter n is 2.4, and the size parameter xsize is 29.5. The results are shown in Fig. 11. Good agreement can still be obtained between the DNN and IITM results, even when three inputs of parameters are not available in the dataset. This finding indicates that the DNN models are able to predict the particle optical properties with reasonable accuracy. Furthermore, it is unnecessary to increase the resolution of parameters in the database.Figure 11. Comparison of (a)
${P_{11}}$ , (b)${P_{12}}/{P_{11}}$ , (c)${P_{22}}/{P_{11}}$ , (d)${P_{33}}/{P_{11}}$ , (e)${P_{43}}/{P_{11}}$ , and (f)${P_{44}}/{P_{11}}$ predicted from the DNNs against the true values. Both the size parameter xsize and refractive index are unavailable in the database.
Parameters | Values | Optical Properties |
Real Part of the Refractive Index ${m_{\text{r}}}$ | 1.30, 1.35, 1.40, 1.45, 1.50, 1.55, 1.60, 1.65, 1.70, 1.75, 1.80 | ${Q_{{\text{ext}}}}$, ${\rm{SSA}}$, g, ${P_{11}}$, ${P_{12}}$, ${P_{22}}$, ${P_{33}}$, ${P_{43}}$, and ${P_{44}}$ |
Imaginary Part of the Refractive Index ${m_{\text{i}}}$ | 10−7, 10−6, 10−5, 0.0001, 0.001, 0.005, 0.01, 0.05, 0.1 | |
Aspect Ratio $a/c$ | 0.5, 0.6, 0.7, 0.8, 0.9, 1.0, 1.2, 1.4, 1.6, 1.8, 2.0 | |
Roundness n | 1.2, 1.4, 1.6, 1.8, 2.0, 2.2, 2.4, 2.6, 2.8, 3.0 | |
Size Parameter xsize | 0.1, 0.2, 0.3, …, 9.8, 9.9, 10 (step = 0.1); 10.2, 10.4, 10.6, …, 19.6, 19.8, 20 (step = 0.2); 21, 22, 23, …, 48, 49, 50 (step = 1); | |
Scattering Angle $\theta $(°) | 0, 0.25, 0.5, 0.75, 1.0, 1.25, 1.5, 1.75, 2.0, 2.25, 2.5, 2.75, 3…179, 179.25, 179.5, 179.75, 180 (step = 0.25) |