Page 85 - Read Online
P. 85

Page 269                          Liu et al. Intell Robot 2024;4(3):256-75  I http://dx.doi.org/10.20517/ir.2024.17














































               Figure 8. The runtime comparison of feature extraction with different graphics card models. (A) RTX 2080Ti; (B) GTX 1050Ti. CPU: Central
               processing unit; GPU: graphic processing unit.


               between GPU and CPU takes a lot of time, which indirectly shows that the time of parallel feature matching
               algorithm is mainly spent on data transmission. By comparing the runtime of the two graphics cards, it can be
               seen that the data transmission efficiency and computing performance of GTX 1050Ti are far lower than that
               of RTX 2080Ti.


               Through comparisons on accuracy and runtime, it can be seen that the parallel feature matching algorithm
               based on CUDA proposed in the paper is roughly the same as the serial OpenCV algorithm in terms of match-
               ing accuracy, but has significantly improved in terms of running efficiency.

               4.2 Bundle adjustment
               In this section, we compare the parallel optimization algorithm proposed in the paper with the g2o algorithm
               running on CPU using different linear solvers, such as Eigen, Csparse and PCG. The test data for the experi-
               ments is obtained from the map generated by ORB-SLAM running on Kitti 00 sequence and saved in the form
               of a graph which includes keyframe-pose vertices, map point vertices, corresponding binary edges and ternary
               edges. The data are categorized as different scales, and the runtime for ten iterations in optimization can be
               seen in Table 3.


               To analyze the results more intuitively, we draw the runtime curves of the two algorithms for different data
               sizes, as shown in Figure 9A, where the horizontal axis represents the number of poses and the vertical axis
               represents the optimization time with ten iterations. As seen in Figure 9A, the parallel algorithm runs faster
   80   81   82   83   84   85   86   87   88   89   90