Page 66 - Read Online
P. 66

Page 160                                                                                                                                                      Kimbowa et al. Art Int Surg 2024;4:149-69  https://dx.doi.org/10.20517/ais.2024.20



                          Table 2. A summary of the deep reviewed learning-based methods listed in descending order of date published

                          Date     Author                     Mode Approach                                           Strengths                                Limitations
                                           [87]
                          2024-    Che et al.         2D     - Calibrate ultrasound probe and needle to optical tracker   - Combines both hardware and software-  - Requires extra hardware: optical tracker and localizers
                          01-12                              - Capture needle tip using tracker                       based approaches                         - Calibration may be erratic if the tip strays from the imaging
                                                             - Represent tip in the ultrasound image coordinate system                                         plane
                                                                                                                                                               - Involves a two-person workflow and separate algorithms for
                                                                                                                                                               needle tracking and detection
                                           [91]
                          2023-    Yan et al.         2D+t   Consists of 3 modules                                    - Incorporates temporal information      - Fails if there is sustained disappearance of the needle tip
                          05-17                              (1) Visual tracking module                               - Thoroughly evaluated for multiple users   - Tracking fails when the needle tip is too small and
                                                             - Pass sampled patches from consecutive frames through the 2- and insertion motions               indistinguishable from background right at insertion
                                                             branch Siamese network
                                                             - One branch dynamically extracts features from template
                                                             patches and the other extracts features from current patch
                                                             (2) Motion prediction module
                                                             - Needle velocity is estimated using displacement of the tip
                                                             between consecutive frames
                                                             (3) Data fusion
                                                             - Use MLP to independently fuse data from visual tracking
                                                             module and motion prediction module
                                             [84]
                          2023-    Wang et al.        2D+t   (1) Enhanced ultrasound image by subtracting current frame   - Accounts for motion                - Only in plane
                          05-09                              from previous frame and fusing the difference with the current   - Continous needle detection even when   - Detects only needle tip
                                                             frame                                                    static between frames                    - In plane
                                                             (2) Extract spatial constraint (bounding box around shaft) if   - Real time                       - Mentions that different detectors could be used, but this is
                                                             present                                                  - Evaluated on human data- Evaluated on   not true as specific detectors (that yield bounding boxes)
                                                             (3) Extract temporal constraint (bounding box around tip) if   CPU                                should be used
                                                             present                                                  - Independent of detector                - Doesn’t visualize shaft
                                                             (4) Combine the constraints to localize needle tip
                                            [88]
                          2023-    Zade et al.        2D+t   - Extract motion field’s amplitude and phase from successive   - Evaluated approach on 2 categories of   - No ablation study to show relevance of the different
                          02-06                              frames                                                   images (needle aligned correctly, and needle  components of the proposed assisted excitation module
                                                             - Extract spatiotemporal features from the motion fields   imperceptible)                         - No comparison with SOTA to show how they fail to model
                                                             - Pass features through detector that outputs a vector of size                                    speckle dynamics
                                                             values including pixelwise line parameters (tx,ty,theta), and
                                                             probabilities for needle shaft and tip
                                            [78]
                          2021-    Chen et al.        2D     - 2 frameworks                                           - Segments shaft, localizes needle       - Does not justify how approach accounts for time (inputs are
                          10-22                              (1) Predict segmentation from two adjacent images        - Doesn’t require prior knowledge of     two consecutive images but with no time information
                                                             (2) 2 models - one segments shaft, the other segments needle   insertion side/orientation         encoded)
                                                             tip from ROIs extracted from the predictions of (1)      - Fully automatic                        - Did not compare with state of the art methods
                                                             - Use segmentations to find
                                             [77]
                          2021-    Wijata et al.      2D     - Pass image through Unet like architecture              - Evaluated on in vivo data              - Not robust to high intensity artifacts
                          06-28                              - Architecture uses large kernels (11 × 11, 9 × 9, 7 × 7) to detect
                                                             large objects
                                                             - Apply Radon transform to determine needle trajectory
                                                             - back-transform trajectory back to binary image
                                            [85]
                          2021-    Rubin et al.       3D     - Pass k previous frames to 3D CNN (including current frame)   - Efficient (real-time, can run on low-cost   - Only detects needle with a bounding box and does not
                          05-11                              - Repeatedly apply 3D convolution to obtain a single feature   computing hardware)                provide metrics on needle tip localization
                                                             map for temporal information                             - Evaluated on challenging cases
                                                             - Pass the feature map through a 2D detector
                                                             - 2D detector outputs bounding box of needle in current frame
   61   62   63   64   65   66   67   68   69   70   71