-
The process used to optimize and predict the quadrupole deformation parameters in our study is illustrated in Fig. 1. Theoretical and experimental datasets are utilized in this study, and key features include the proton number Z and neutron number N. These features are standardized to ensure uniform scaling, which is crucial for neural network training. The datasets are strategically partitioned into training, validation, and testing sets. Specifically, the theoretical data are allocated into two distinct sets: 60% for training and the remaining 40% for validation purposes. For the experimental dataset, a more nuanced division is employed. Initially, 80% of the dataset is designated for a retraining phase, of which 40% is subsequently earmarked for validation. The residual 20% of the overall dataset is exclusively reserved for final testing and evaluation, ensuring a robust assessment of the model's predictive capabilities.
The neural network architecture is specifically designed to address the characteristics of nuclear quadrupole deformation data, with its configuration and experimental details outlined in Table 1. It comprises an input layer, two hidden layers, and an output layer, each serving a distinct purpose in the data processing flow. The input layer is configured to receive the standardized proton number Z and neutron number N. Following this, there are two hidden layers, each consisting of 64 neurons. These layers play a pivotal role in feature extraction and executing non-linear transformations of the data. Both hidden layers utilize rectified linear units (ReLUs) as the activation function, enhancing the network's non-linear processing capabilities and mitigating gradient vanishing issues. The architecture concludes with an output layer housing a single neuron employing the Sigmoid activation function. In this study, the neuron number in the hidden layer is determined after several trials. The final value gives the best results.
Parameter/Setting Value/Description Input layer size 2 (Z, N) Hidden layers 2 Neurons per hidden layer 64 Output layer size 1 Activation function (hidden) ReLU Activation function (output) Sigmoid Loss function MSE Optimizer Adam Learning rate 0.001 Epochs (theoretical data) 5000 Epochs (experimental data) 5000 Train-validation split (theory) 60%−40% Train-validation-test split (exp) 48%−32%−20% Random seed 0 (fixed) Transfer learning layers frozen Two hidden layers Table 1. Neural network configuration and experiment details.
Our approach consists of two primary stages: Initially, we conduct pre-training of the neural network on the theoretical dataset, adjusting its weights by utilizing this data. The training epoch is set to 5000, with a learning rate (LR) of 0.001. The comparison between the expected output and the neural network's output is facilitated using a loss function, employing the mean squared error (MSE). Subsequently, we fine-tune the model on a limited experimental dataset. The weights of the input and hidden layers of the neural network model are frozen during this phase of transfer learning. This implies that these weights remain unchanged throughout the transfer learning process. This approach is chosen because these layers have already learned the fundamental features of quadrupole deformation from the pre-training phase. We exclusively retrain the output layer and update its weights based on the experimental dataset, thereby fine-tuning the pre-trained network. Finally, to test the model, performance metrics such as the MSE, mean absolute error (MAE) and root MSE (RMSE) can be employed to evaluate the results estimated by the neural network model and determine the performance level. The calculation methods for these performance metrics are as follows:
$ {\rm{MSE}} = \frac{1}{n} \sum\limits_{i=1}^{n} (E_i - P_i)^2, $
(1) $ {\rm{MAE}} = \frac{1}{n} \sum\limits_{i=1}^{n} |E_i - P_i|, $
(2) $ {\rm{RMSE}} = \sqrt{\frac{1}{n} \sum\limits_{i=1}^{n} (E_i - P_i)^2}, $
(3) where n is the total number of data, and
$ E_i $ and$ P_i $ denote the experimental and predicted values of the ith sample. -
Next, we elucidate the calculation of the capture cross-section in heavy-ion fusion reactions, highlighting the pivotal role of quadrupole deformation parameters in these computations. The empirical coupled channel model is utilized for the computation of this capture cross-section [37, 38]. The capture cross-section is expressed as [39]
$ \sigma_{\rm cap} = \frac{\pi \hbar^2}{2\mu E_{\rm c.m.}}\sum\limits_J (2J + 1) T(E_{\rm c.m.}, J), $
(4) where
$E_{\rm c.m.}$ is the center-of-mass incident energy, and the transmission probability$T(E_{\rm c.m.}, J)$ is calculated using the Hill-Wheeler formula [40]. Integrating the effect of coupling channels through the potential barrier distribution function, the transmission probability is$ \begin{aligned}[b] T(E_{\text{c.m.}}, J) =\; &\int f(B) \bigg[ 1 + \exp \Bigg( -\frac{2\pi}{\hbar \omega(J)} \\ &\times\left[ E_{\text{c.m.}} - B - \frac{\hbar^2}{2\mu R_B^2(J) } J(J + 1) \right] \Bigg) \bigg]^{-1} {\rm d}B, \end{aligned} $
(5) where
$ \hbar\omega(J) $ is the width of the parabolic form at the position of the barrier$ R_B(J) $ . The barrier distribution function$ f(B) $ takes an asymmetric Gaussian shape,$ f(B) = \left\{\begin{array}{*{20}{l}} {\dfrac{1}{N} \exp \left[ -\left( \dfrac{B - B_m}{\Delta_1} \right)^2 \right], }& {B < B_m} \\ {\dfrac{1}{N} \exp \left[ -\left( \dfrac{B - B_m}{\Delta_2} \right)^2 \right],} & {B > B_m} \end{array} \right. $
(6) where
$ B_m = \dfrac{B_s + B_0}{2} $ , with$ B_0 $ as the Coulomb barrier height at waist-to-waist orientation, and$ B_s $ as the minimal height influenced by the dynamical deformation parameters$ \beta_1 $ and$ \beta_2 $ . N is the normalization constant,$ \Delta_2 = (B_0 - B_s)/2 $ , and$ \Delta_1 $ is typically 2−4 MeV less than$ \Delta_2 $ [41]. Incorporating quadrupole deformation, the nucleus-nucleus interaction potential is formulated as$ \begin{aligned}[b] V\left( {r,{\beta _1},{\beta _2},{\theta _1},{\theta _2}} \right) =\;& {V_C}\left( {r,{\beta _1},{\beta _2},{\theta _1},{\theta _2}} \right)\\ &+ {V_N}\left( {r,{\beta _1},{\beta _2},{\theta _1},{\theta _2}} \right)\\ &+ \frac{1}{2}{C_1}{\left( {{\beta _1} - \beta _1^0} \right)^2} \\ &+ \frac{1}{2}{C_2}{\left( {{\beta _2} - \beta _2^0} \right)^2}, \end{aligned} $
(7) where
$ \beta_1 (\beta_2) $ is the dynamical quadrupole deformation parameter for the projectile (target), and$ \beta_1^0 (\beta_2^0) $ is the static deformation parameter.$ \theta_1 (\theta_2) $ represents the angle between the radius vector and the symmetry axes of the statically deformed projectile (target). The stiffness parameters$ C_{1,2} $ are derived using the liquid drop model [42]. The Coulomb and nuclear potentials,$ V_C $ and$ V_N $ , are as specified in Ref. [39]. Therefore, in the calculations of barrier heights and capture cross-section in heavy-ion fusion reactions, the quadrupole deformation parameter emerges as an indispensable parameter that cannot be overlooked.
Transfer learning and neural networks in predicting quadrupole deformation
- Received Date: 2024-01-22
- Available Online: 2024-06-15
Abstract: Accurately determining the quadrupole deformation parameters of atomic nuclei is crucial for understanding their structural and dynamic properties. This study introduces an innovative approach that combines transfer learning techniques with neural networks to predict the quadrupole deformation parameters of even-even nuclei. With the application of this innovative technique, the quadrupole deformation parameters of 2331 even-even nuclei are successfully predicted within the nuclear region defined by proton numbers