Page 91 - Read Online
P. 91

Page 79                           Shu et al. Intell Robot 2024;4(1):74-86  I http://dx.doi.org/10.20517/ir.2024.05










































                                        Figure 4. Feature extraction neural network structure diagram.



               2.3 Feature extraction
               Patients with PD typically exhibited decreased movement amplitude and velocity during continuous limb
               movements, which can be captured in the spatio-temporal domain of motion capture data. This paper uses
               the one-dimensional CNN to extract the spatial-temporal features of motion capture signals and characterize
               the movement differences in PD patients with varying degrees of bradykinesia and healthy controls. The mo-
               tion capture data for the finger and toe tapping tasks have dimensions of six and three, respectively. Given the
               similarity between these two datasets, we have established a unified feature extraction network structure [Fig-
               ure 4]. To capture the interactive features between the multivariate dimensions, the data were reorganized into
               three groups, each formed by randomly selecting two dimensions. Each of these groups is then fed into one-
               dimensional convolutional blocks. The global average pooling layer is used to extract different dimensional
               spatio-temporal features. Finally, the outputs of the pooling layer are concatenated, and the latent features are
               derived through a fully connected layer. The architecture of each one-dimensional convolutional block com-
               prises a one-dimensional convolution layer, a batch normalization layer, and a rectified linear unit (ReLU).
               The last two convolutional blocks employ a parameter sharing mechanism to reduce the number of network
               parameters.




               For the feature extraction neural network, the learning rate is set to 5 × 10 , and the training is terminated
                                                                              −6
                                                  −9
               when the change in loss is less than 1 × 10 . In addition, the dropout and regularization methods are added
               to reduce the effect of overfitting. In this paper, the dropout parameter is set to 0.8, and the parameter of L2
               regularization is set to 0.5.
   86   87   88   89   90   91   92   93   94   95   96