Page 22 - Read Online
P. 22

Page 227                                                    Dababneh et al. Art Int Surg 2024;4:214-32  https://dx.doi.org/10.20517/ais.2024.50

               instability is crucial to avoid deterioration of this condition. Nevertheless, signs of carpal instability are often
                                                [86]
               unnoticed on conventional radiographs . In response to this diagnostic challenge, Hendrix et al. developed
                                                                            [87]
               an AI model to identify and assess signs of carpal instability on X-rays . The model demonstrated mean
               absolute errors (MAE) of 0.65 mm in measuring SL distances, 7.9 degrees for SL angles, and 5.9 degrees for
               capitolunate angles. Furthermore, the algorithm achieved an AUC of 0.87 for carpal arc interruption
               detection, outperforming four out of five experts.

               Peripheral nerve injuries
               Gu et al. conducted a study focusing on the remote screening of peripheral nerve injuries . Three gestures,
                                                                                          [88]
               each corresponding to a specific nerve, were developed by an expert in the field to detect functional
               abnormalities caused by radial, ulnar, or median nerve injury. The authors trained multiple algorithms, all
               of    which  achieved  an  accuracy  rate  above  95%  for  all  three  gestures,  demonstrating  their  efficacy  in
               detecting abnormalities in the radial, ulnar, and median nerves.


               Prolonged postoperative opioid use prediction
               It is well known that opioid use is common after hand surgery. To address this, Baxter et al. conducted a
                                                                                            [89]
               study exploring the potential of AI in predicting prolonged opioid use post-hand surgery . Their results
               indicate that AI, with further training, can potentially be used to identify patients at risk of prolonged opioid
               use, with one of the models achieving an AUC of 0.84.

               AI LIMITATIONS
               Despite the rapid advancement in the field of AI and its promising performance in controlled settings,
               several challenges must be addressed before its full integration into hand surgery.  One significant limitation
               of many studies examining AI’s role in fracture detection is the lack of validation using external datasets,
               often due to the small size and homogeneity of the samples. This limitation arises from data privacy and the
               absence of large, labeled datasets across multiple institutions, as well as the need for expert labeling in
                                [33]
               supervised learning . Additionally, methodological variations, concerns about applicability, risks of bias,
               and differences in diagnostic protocols between centers further complicate the integration of AI into clinical
               practice, highlighting the necessity of standardized guidelines to ensure the quality and reliability of AI-
               driven models and provide structured and consistent methodologies . Moreover, most existing studies are
                                                                         [46]
               retrospective, which, while useful for demonstrating proof of concept, fall short of establishing the robust
               evidence required for clinical application. Therefore, prospective studies are needed to confirm the
               performance metrics and to demonstrate AI’s potential to enhance patient management and outcomes.

               Moreover, while LLMs have shown their potential in accessing, interpreting, and synthesizing extensive
               amounts of information, they still struggle with complex cases that require the integration of nuanced
               clinical contexts. Despite advancements in newer versions, these models are not yet fully equipped to meet
               the clinical needs of patients and healthcare providers. Similarly, to ensure the effective integration of AI
               models into academic contexts, there is a need for further training to enhance the validity of AI-generated
               content and to improve the transparency of the sources from which this information is derived.

               Finally, it is also essential to acknowledge that at this stage of its evolution, AI does not possess the
               capability to replace humans. Numerous ethical and liability concerns must be thoroughly examined before
               such a possibility can even be considered. As AI technology continues to advance, it is crucial to develop
               clear guidelines and establish a robust regulatory framework in parallel, ensuring that these innovations are
               integrated responsibly and ethically.
   17   18   19   20   21   22   23   24   25   26   27