Chest CT scans are often the preferred imaging modality for patients with chest trauma because they can identify many injuries that may be missed by chest X-rays, including pulmonary contusions, hemothorax, pneumothorax, and rib fractures. Moreover, rib fractures are considered indicators of severe trauma [26]. Traditional detection methods require physicians to meticulously evaluate the entire CT scan, which is a time-consuming and error-prone process, particularly for less experienced radiologists or thoracic surgeons. Additionally, the number and displacement of rib fractures are related to the follow-up treatment plan[27; 28]. Therefore, thoracic surgeons should prioritize different types of rib fractures. Chronic rib fractures, characterized by the presence of a mature callus or an invisible fracture line, do not require clinical intervention and do not affect treatment decisions or patient prognosis, making their detection unnecessary. In contrast, timely identification and localization of fresh rib fractures are crucial, especially in emergency settings, as severely displaced rib fractures may be associated with internal organ injuries, significantly impacting patient outcomes[29].
To address this issue, we developed a deep learning-based intelligent diagnostic model for the classification and detection of fresh rib fractures, which was validated on both internal multicenter datasets and external public datasets. The model outperformed experienced thoracic surgeons and demonstrated exceptional performance.
Recently, Yao et al.[30] developed a deep learning-based rib fracture detection system that achieved high-performance detection and diagnosis of rib fractures on chest CT images, significantly reducing physician workload and minimizing misdiagnoses. However, this study only addressed the binary classification of fracture presence. Zhou et al.[31] developed a convolutional neural network (CNN) model that classified fractures into fresh, healing, and old fractures but did not perform a graded diagnosis for precise fracture diagnosis. Zhou et al.[21] reported the use of clinical information in their CNN model, which similarly improved diagnostic efficiency and reduced diagnosis time. Xiong et al.[32] found that the performance of radiologists on night shifts was inferior to that on day shifts, and the use of a deep learning-based computer-aided diagnosis (CAD) system for rib fractures helped night shift radiologists achieve performance levels comparable to their daytime performance. Unlike previous studies, our research focused on fresh rib fractures, as their rapid localization and diagnosis are crucial components of intelligent diagnosis and treatment of acute chest trauma. Chronic rib fractures do not affect treatment decisions or patient outcomes, making the focus on fresh rib fractures more aligned with real clinical scenarios. Our model can also intelligently grade fractures based on their severity, aiding in treatment decisions and prognosis evaluation for posttraumatic rib fractures.
Our model's performance is also comparable to that of physicians, making it suitable as the "first reader." This approach can help improve diagnostic accuracy, reduce diagnosis time, and reduce the workload of physicians; additionally, it can aid in building medical resources in under resourced areas.
In our model improvements, we optimized the Backbone, Neck, and Head networks to enhance feature extraction, fusion, and detection capabilities. The Backbone, designed to extract deep features from medical images, alternates between CBS modules (Convolution, Batch Normalization, and SiLU activation) and C2f_EMA modules, which include convolution layers and parallel Bottleneck_EMA branches for richer feature representation. It concludes with the SPPF module, which combines multiple Maxpool and Concat operations to capture critical global information. In the Neck, we focused on refining and consolidating multi-scale features using an Upsample operation and the C2f_EMA module to merge and process features across different levels, improving detection accuracy and robustness. Finally, the Head network, composed of additional C2f_EMA modules, integrates and processes refined features before passing them to the Detect layer, where bounding boxes and class predictions are generated. This comprehensive multi-level integration allows our model to achieve high-precision detection in medical images.
However, our study had several limitations. One limitation of the current model is its inability to identify specific anatomical structures, such as which rib is fractured, which may reduce its utility in acute settings requiring rapid, precise localization. Additionally, as AI is increasingly integrated into clinical workflows, ethical concerns around data security, patient privacy, and clinician reliance on automated systems must be addressed. Our model has demonstrated robustness across both internal and external datasets. However, future studies should focus on validating the model in a variety of clinical settings, particularly in hospitals with differing patient demographics, trauma protocols, and imaging equipment to ensure broad applicability and our sample size could be further expanded. Finally, prospective studies remain relatively scarce. We aim to collect more data and conduct prospective studies to further validate and optimize the model.