Page 65 - Read Online
P. 65

Kimbowa et al. Art Int Surg 2024;4:149-69  https://dx.doi.org/10.20517/ais.2024.20                                                      Page 159

               Table 1. A summary of the reviewed deep learning-based methods in descending order by date of publication, and their
               corresponding results
                                                                          Tip error Orientation error  Processing
                Author                          Mode Dataset  Target organ
                                                                          (mm)   (degrees)     time (fps)
                     [87]
                Che et al.     2D   Phantom (bovine, carrageenan)  Liver  1.17
                Wang et al. [84]  2D+t  Ex vivo (porcine)  Liver, heart, kidney,   0.92        10
                                                          abdomen
                      [88]
                Zade et al.    2D+t  Phantom (lumbar epidural   Epidural space,   2.12  2.08   14
                                    simulator, femoral Vascular Access  vascular structures
                                    Ezono)
                Chen et al. [78]  2D  Ex vivo (bovine, porcine)           0.45   0.42          14
                       [77]
                Wijata et al.  2D   In vivo (core needle biopsies)                             46
                Mwikirize et al. [66]  2D+t  Ex vivo (bovine, porcine, chicken)  Epidural space  0.52  16
                     [76]
                Gao et al.     2D   Ex vivo (bovine, porcine, chicken)    0.29   0.29          67
                Zhang et al. [79]  2D  In vivo (HDR prostate   Prostate   0.33
                                    brachytherapy)
                Andersén et al. [81]  2D  In vivo (HDR prostate   Prostate       1
                                    brachytherapy)
                       [89]
                Gillies et al.  2D  Phantom (tissue)      Tissue, prostate,   4.4  1.4
                                    In vivo (prostate, gynecologic, liver,  gynecologic, liver,
                                    kidney)               kidney
                Zhang et al. [79]  3D  In vivo (HDR prostate   Prostate   0.44
                                    brachytherapy)
                Zhang et al. [75]  3D  In vivo (HDR prostate   Prostate   0.442
                                    brachytherapy)
                Lee et al. [74]  2D  In vivo (biopsy)     Kidney                 13.3
                         [65]
                Mwikirize et al.  2D+t  Ex vivo (bovine, porcine, chicken)  Epidural space     1
                Pourtaherian et al. [90]  3D  Ex vivo (porcine leg)
                     [73]
                Arif et al.    3D+t  Phantom (abdominal, liver)   Liver, abdomen
                                    In vivo (liver biopsy)
                        [72]
                Pourtaherian   3D   Ex vivo (chicken breast, porcine      0.7
                                    leg)
                Mwikirize et al. [83]  2D  Ex vivo (porcine, bovine)  Epidural space  0.23  0.82
                         [60]
                Mwikirize et al.  3D  Ex vivo (bovine, porcine)  Epidural space  0.44
                Pourtaherian et al. [71]  3D  Ex vivo (chicken breast)    0.5    2
                      [69]
                Beigi et al.   2D+t  In vivo (pig)        Biceps femoris muscle 0.82  1.68
                Beigi et al. [55]  2D+t  In vivo (pig)    Biceps femoris muscle 1.69  2.12
                Geraldes and   2D   Phantom (ballistic gelatin)           5
                Rocha [70]
               Mode 2D+t indicates methods that incorporate temporal information, that is, they take as input multiple consecutive ultrasound images before
               making a prediction. Almost all methods were trained and evaluated only on phantom datasets, with the exception of those developed for HDR
               prostate brachytherapy. Note that blanks in the table indicate that the entry was not reported in the paper. HDR: High dose rate.


               One observation from many of the proposed learning-based methods performing needle segmentation is
               the use of metrics, such as the Dice coefficient and intersection over union (IOU), that do not reflect clinical
               needs for needle localization. For instance, both DICE and IOU only measure the overlap between the
               prediction and the ground truth; they are inherently biased to focus more on the needle shaft, which may
               not be as informative as the needle trajectory error (degrees) or tip localization error (mm). The trend
               toward such metrics has mainly increased with machine learning and deep learning-based methods and
               could be a result of adopting metrics commonly used in other domains where they directly reflect the
               domain needs. Caution should thus be taken when selecting metrics to optimize and evaluate proposed
               software-based methods [93,94] . Maier-Hein et al. developed a comprehensive framework to act as a guideline
                                                                                                      [93]
               in selecting problem-aware metrics for assessing medical image processing machine learning algorithms .
   60   61   62   63   64   65   66   67   68   69   70