Page 38 - Read Online
P. 38

Bao et al. Complex Eng Syst 2022;2:16  I http://dx.doi.org/10.20517/ces.2022.30   Page 7 of 10


















                                         Figure 5. Training loss curve on the HTRU-Medlat dataset.

                                         Table 2. HTRU-Medlat dataset partitioning after expansion

                                  Sample size                    of real samples  of fake samples
                                  Total number of samples in the original dataset  1,196  89,996
                                  IDS training set size              696     10,000
                                  BDS training set size             10,000   10,000
                                  Test set size                      500     500


               To verify that the generative model can alleviate the problem of sample imbalance, we divide the dataset into
               two cases, one is an imbalanced data scenario (IDS) with 696 real samples and 10,000 fake samples, and the
               other is a balanced data scenario (BDS) with 10,000 real samples and 10,000 fake samples. For the HTRU-
               Medlat dataset, the detailed partitioning scenarios are shown in Table 2.


               This paper uses a CNN [7]  model for comparison, which has a similar structure to the LeNet network struc-
               ture [22] ,butwithsomeadaptationsforthepulsarcandidateidentificationtask. Forthehyperparametersettings
               of the residual network model, this paper uses a mini-batch size of 128, a learning rate of 0.001, and a size of
               0.00001fortheL2regularisationtermused. WealsousedastandardGaussiantoinitialisetheparametersofthe
               model and a ReLU [23]  activation function for all layers except the last layer of the model, which uses a sigmoid
               activation function. The objective function for optimisation is cross-entropy, and the Adam [24]  optimiser is
               used.

               For the hyperparameter settings of the generative adversarial network, the learning rate is 0.001, the L2 regular-
               ization weight is 0.0005, the number of training rounds is 200, the optimizer is Adam, the size of the minibatch
               is 128, the parameters areinitialized usingKaiming initialization [25] , the discriminator is trained with 5 rounds
               for each batch, the discriminator weights range from [-0.005, 0.005], and LeakyReLU employed a slope of 0.1.


               The simulated samples are generated using the generative model based on the generative adversarial networks
               designedinthis paper. Duringthetrainingprocess, real samplesfromtheIDS trainingset areusedfor training.
               The trained model is then used to generate a series of simulated samples that are used to augment the real
               sample data. In addition, the simulated samples are filtered, i.e., the generated simulated samples need to
               be identified as positive by the corresponding residual network model proposed in this paper. The filtered
               simulated samples are used to expand the IDS training set to obtain the BDS training set.

               The results of the automatic identification of pulsar candidates for each method on the HTRU-Medlat dataset
               are shown in Table 3. ”Subints” indicates that the temporal phase images were input, and ”Subbands” indicates
               that the frequency phase images were input. The method with ”IDS” indicates that the experiment was tested
               with small data, while ”BDS” indicates that a large dataset consisting of images generated by a generative
               adversarial network was incorporated. In the IDS data set scenario, the F1 value of the CNN model method
               reached approximately 95%, while the F1 value of the ResNet method proposed in this paper reached 97.3%
   33   34   35   36   37   38   39   40   41   42   43