Figures
Abstract
Objective
The Thyroid Imaging Reporting and Data System (TI-RADS) is an essential tool for assessing thyroid nodules, primarily used by radiologists. This study aimed to compare the agreement of TI-RADS scores between sonographers and radiologists and to assess the diagnostic performance of these scores against histological findings in suspicious thyroid nodules.
Methods
In a retrospective analysis, 168 patients with suspicious thyroid nodules classified as TR3 and above by the radiologists were included. Both sonographers and radiologists independently assigned the American College of Radiologists (ACR) TI-RADS scores, which were then compared for inter-reader agreement using Cohen’s Kappa statistic. The scores were also evaluated for diagnostic performance against histological results based on the Bethesda system.
Results
The study revealed a moderate overall agreement between sonographers and radiologists in TI-RADS scoring (κ = 0.504; 95% CI: 0.409–0.599), with poor agreement noted specifically for nodule margin scores (κ = 0.102; 95% CI: -1.430–0.301). In terms of diagnostic performance against histological outcomes, sonographers’ TI-RADS scores showed a sensitivity of 100% and a specificity of 44.6%, while radiologists’ scores showed a sensitivity of 100% but a lower specificity of 29.3%.
Conclusion
The findings indicate moderate agreement in TI-RADS scoring between sonographers and radiologists, with reproducibility challenges especially in scoring nodule margins. The marginally superior diagnostic performance of sonographers’ scores suggests potential efficiency benefits in involving sonographers in preliminary assessments. Future research should aim to encompass a wider range of TI-RADS categories and focus on minimizing scoring variability to enhance the system’s clinical utility.
Citation: Alfuraih AM, Alotaibi AM, Alshammari AK, Alrashied BF, Mashhor YM, Mahmoud M, et al. (2024) Evaluating diagnostic accuracy and agreement of TI-RADS scoring in thyroid nodules: A comparative analysis between sonographers and radiologists. PLoS ONE 19(10): e0312121. https://doi.org/10.1371/journal.pone.0312121
Editor: Milad Khorasani, Neyshabur University of Medical Sciences, ISLAMIC REPUBLIC OF IRAN
Received: August 1, 2024; Accepted: October 1, 2024; Published: October 11, 2024
Copyright: © 2024 Alfuraih et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: All relevant data are within the manuscript and its Supporting Information files.
Funding: This research was funded by PRINCE SATTAM BIN ABDULAZIZ UNIVERSITY, grant number PSAU/2024/01/78904.
Competing interests: The authors have declared that no competing interests exist.
Introduction
Thyroid nodules are a common clinical finding, with a prevalence of 19–68% in the general population when assessed by ultrasound [1]. The majority of these nodules are benign, and only a small percentage are malignant [2]. Accurate identification of high-risk nodules is crucial for timely intervention and optimal patient management. The Thyroid Imaging Reporting and Data System (TI-RADS) is a widely used classification system for the risk stratification of thyroid nodules based on ultrasound features [3,4]. This system aims to standardize the reporting of thyroid ultrasound findings and improve the diagnostic accuracy of thyroid nodule evaluation [5,6].
The rationale behind this study is rooted in the clinical need for accurate scoring of thyroid nodules to guide patient management. High-risk nodules may require fine needle aspiration (FNA) to rule out malignancy, whereas low-risk nodules can be managed conservatively without invasive procedures [7,8]. Accurate TI-RADS scoring is essential for identifying which patients require FNA and sparing those with low-risk nodules from unnecessary invasive procedures.
Multiple previous studies have demonstrated a high accuracy of sonographers’ diagnostic opinion compared to radiologists in abdominal [9–16], obstetric [17], musculoskeletal [18], and emergency ultrasound [19]. However, no previous studies have directly compared the agreement of TI-RADS scores between sonographers and radiologists. Albeit previous studies have examined the agreement of TI-RADS scores between radiologists, with varying results [20–25]. Given the increasing demand for ultrasound examinations and the limited availability of radiologists, it is essential to determine if sonographers can accurately assess thyroid nodules using the TI-RADS scoring system [26,27]. Moreover, if sonographers can reliably and accurately score thyroid nodules, radiologists may be able to focus on more complicated tasks, improving overall efficiency and resource allocation within the radiology department [28,29].
Despite the potential benefits of using TI-RADS in clinical practice, some concerns have been raised regarding the reproducibility and reliability of the classification system [30]. Inconsistencies in the application of TI-RADS criteria and interobserver variability may affect the accuracy of the scoring system, warranting further investigation to ensure its validity and reliability [31].
The main objectives of this study are twofold. The first is to compare agreement of TI-RADS scores assigned by sonographers and radiologists. The second objective is to analyze the diagnostic accuracy of the TI-RADS scoring system. The underlying hypothesis is that the scores assigned by sonographers agree with those assigned by radiologists. If confirmed, this could have considerable implications for the allocation of resources within radiology.
Materials and methods
Study design and case selection
The study employed a retrospective design, covering the period from January 2021 to January 2022. The research data were accessed from the radiology department at King Fahad Medical City in Riyadh, Saudi Arabia, between July 2022 and January 2023 for research purposes. To minimize selection bias, we included all consecutive adult patients who met the inclusion criteria during the study period, ensuring a representative sample. The study was approved by the hospital’s IRB committee and a consent waiver was granted (NIH registration number: IRB00010471) to retrospectively collect this study data.
During the study period, all patients with thyroid nodules received a TI-RADS score [according to the TI-RADS system developed by the American College of Radiologists (ACR) [8]] by the scanning sonographer and the reporting radiologist in real-time during the patient visits, which was documented on the radiology information system (RIS). Each case was scored based on two-plane grey-scale images of the thyroid nodule. When a patient has more than one nodule, the most suspicious one was scored and documented. The patients were considered eligible if they were an adult (18 years or older) and had a documented Bethesda score following an FNA biopsy done after the ultrasound scan. Patients who did not undergo thyroid sonography prior to the FNA or the duration between the scan and FNA was more than two weeks were excluded.
TI-RADS scoring
The thyroid ultrasound scans were performed using a linear array transducer (5–15 MHz) on a LOGIQ E9 ultrasound system (GE Healthcare, Milwaukee, WI, USA) by one of 11 accredited sonographers with ultrasound experiences ranging from 2 to 18 years. All sonographers were trained to interpret thyroid US images and assign ACR TI-RADS scores after receiving a focused training course by the chief consultant radiologist (12 years of experience). The training was done in July 2020, which included official educational materials developed by the ACR including the TI-RADS atlas and user guide [32]. This was followed by regular monitoring and feedback over 3 months. The retrospective data collected for this study was set after approximately 6-months of the sonographers’ training and practice. Hence, all sonographers had the same level of experience (6-months) in TI-RADS scoring. Following the sonographers TI-RADS scores, the images were reviewed by one of 5 accredited radiologists with experiences ranging from 5 to 12 years and a TI-RADS score was documented on the RIS system. The radiologists included in the study consisted of three neuroradiologists, one cardiothoracic radiologist, and one abdominal radiologist. While their primary subspecialties are varied, all radiologists are board-certified and routinely interpret thyroid ultrasound examinations as part of their clinical duties.
Bethesda scoring
After the ultrasound examinations, fine-needle aspiration (FNA) biopsies were performed on suspicious thyroid nodules (≥ TR3 score by the radiologist). The cytological results were reported according to the Bethesda System for Reporting Thyroid Cytopathology (TBSRTC) [33]. This standardized system categorizes thyroid FNA results into six diagnostic categories: Category I: Non-diagnostic or Unsatisfactory; Category II: Benign; Category III: Atypia of Undetermined Significance or Follicular Lesion of Undetermined Significance; Category IV: Follicular Neoplasm or Suspicious for a Follicular Neoplasm; Category V: Suspicious for Malignancy; Category VI: Malignant.
FNA procedures were conducted by experienced interventional radiologists under ultrasound guidance using a 25-gauge needle. The obtained samples were processed and evaluated by board-certified cytopathologists following the TBSRTC guidelines. For our analysis, we grouped the Bethesda categories as follows: Benign: Category II; Malignant or Suspicious for Malignancy: Categories V and VI; Indeterminate: Categories I, III, and IV [25,33].
Statistical analysis
All categorical variables are presented as frequencies and percentages while continuous variables are presented as mean ± SD. The Kolmogorov-Smirnov test was used to confirm the assumption of normal distribution. The Bethesda scores were used to categorize the nodules into benign (Bethesda II), suspicious or malignant (Bethesda V and VI), or indeterminate (Bethesda I, III, and IV) [25,33].
Inter-reader reliability for TI-RADS scores were assessed using Cohen’s Kappa statistic. The Kappa scale was interpreted as follows: <0, poor agreement; 0.01–0.20, slight agreement; 0.21–0.40, fair agreement; 0.41–0.60, moderate agreement; 0.61–0.80, substantial agreement; and 0.81–1.00; almost perfect agreement. Diagnostic accuracy analysis was performed to determine the sensitivity and specificity analysis of TI-RADS by sonographer and radiologist. To simplify the analysis, TR1 and TR2 were categorized as “low suspicion of malignancy” and TR3, TR4, TR5 categorized as “high suspicion of malignancy” [34]. The indeterminate (Bethesda I, III, and IV) scores were excluded from the diagnostic accuracy analysis. Receiver Operating Characteristic (ROC) curve analyses were conducted to evaluate and compare the diagnostic performance of the sonographers and radiologists. The area under the curve (AUC) was calculated for each group to quantify their overall diagnostic accuracy. The association between TI-RADS scores and the histological outcomes was evaluated using the Chi-Square test. A two-sided p-value less than 0.05 was considered statistically significant. All data was entered and analyzed using the SPSS 25 Statistics Package (SPSS Inc., Chicago, Illinois, USA) and MEDCALC version 18.11.6 (Acacialaan 22 8400 Ostend Belgium).
Results
A total of 168 patients were eligible (adults with available TI-RADS and Bethesda scores) and included in the study, of which 18 (11.1%) were male and 150 (88.9%) were female. The mean age was 50.3 ± 13.7 years. The histological results of the nodules were as follows: malignant or suspicious nodule (n = 11), benign nodule (n = 92), and indeterminate nodule (n = 65). The distribution of TI-RADS scores and Bethesda scores is presented in Table 1 and Fig 1 respectively. The 9 patients scored by the sonographers as TR1 or TR2 had benign follicular nodules (Bethesda II). The detailed distribution of the TI-RADS features based on the Bethesda subcategories is available in the supplementary materials (S1 Table). All 11 patients with FNA results indicating malignancy (n = 8) or suspicion for malignancy (n = 3) underwent surgical excision and histopathological examination. The histopathological analysis confirmed malignancy in all cases, showing concordance with the FNA cytology results.
Agreement results
Table 2 cross-tabulates the scores by the radiologists and sonographers’ readers per case. The scores are also visually presented in Fig 2. The results demonstrate good agreement overall, and few instances of substantial discrepancies in scoring between readers. For example, two cases where scored TR1 by the sonographer compared to TR3 by the radiologist.
The Bubble labels represent the instances of matched scores.
Inter-reader reliability was assessed using the Kappa statistic in Table 3. Moderate agreement was found between the sonographer and radiologist readers for TI-RADS (κ = 0.504), composition score (κ = 0.463), echogenicity score (κ = 0.584), shape score (Kappa = 0.462), and echogenic foci score (κ = 0.548) (Table 3). All p-values were significant at p < 0.001, indicating significant agreement among readers. The margin score, however, was not significant (p = 0.153) and the Kappa statistic of 0.102 indicated slight agreement.
Diagnostic accuracy results
In evaluating the diagnostic accuracy of TI-RADS scores in Tables 4 and 5, in which the indeterminate nodules were excluded, both sonographers and radiologists demonstrated high sensitivity but varied in specificity and positive likelihood ratios. The sonographers achieved a sensitivity of 100% (95% CI: 71.5–100.0), indicating accurate identification of all positive (suspicious or malignant) cases (11 out of 11). However, specificity was moderate at 44.6% (95% CI: 34.2–53.3), suggesting a higher rate of false-positive diagnoses compared to true negatives. The positive likelihood ratio was 1.80 (95% CI: 1.5–2.2), reinforcing the reliability of a positive test result in diagnosing the disease. The positive predictive value (PPV) was 17.7% (95% CI: 15.2–20.6), and the overall accuracy of the sonographer’s TI-RADS scoring was 50.5% (95% CI: 40.5–60.5).
Similarly, the radiologist’s sensitivity was 100% (95% CI: 71.5–100.0). However, specificity was lower at 29.3% (95% CI: 20.3–39.8), indicating a greater tendency for false positives with only 27 of 92 non-malignant cases correctly identified. The positive likelihood ratio of 1.4 (95% CI: 1.2–1.6) was also lower compared to the sonographer, suggesting a slightly reduced diagnostic accuracy in positive test interpretation. The PPV was 14.5% (95% CI: 12.9–16.2), and the overall accuracy was 36.9% (95% CI: 27.6–47.0). The negative predictive value (NPV) was 100% for both sonographers and radiologists, indicating that all nodules with negative or low suspicion TI-RADS scores were confirmed to be benign on histology when tested.
The ROC curves for sonographers and radiologists are presented in Fig 3. The AUC for sonographers was 0.723 (95% CI: 0.684–0.762), indicating acceptable diagnostic accuracy. For radiologists, the AUC was 0.662 (95% CI: 0.626–0.699). The difference between the two AUCs was statistically significant (p < 0.001), suggesting that sonographers had a slightly better diagnostic performance compared to radiologists.
Discussion
This study is the first, to our knowledge, that compared the agreement of TI-RADS scores in suspicious thyroid nodules between sonographers and radiologists. Our results highlighted an overall moderate agreement between the two groups of readers. This agrees with the published results on inter-reader reproducibility between radiologists in other studies. For example, Liu et al. [23] conducted a meta-analysis of seven studies investigating the agreement of TI-RADS scores between various groups of radiologists. They found an overall moderate agreement across the seven studies (κ = 0.54; 95% CI: 0.49−0.58). More recently, Li et al [35] conducted a similar meta-analysis, which yielded a total of 13 studies. The pooled inter-reader agreement for overall ACR TI-RADS system was also moderate (k = 0.51, 95% CI: 0.42–0.59). They reported no significant difference between experienced and inexperienced readers (p = 0.55). They also noted that agreements were usually higher in inexperienced readers, which could be explained by their likelihood to follow the ACR scoring lexicon more strictly than experienced readers. Indeed, a recent study by Daniels et al [36] reported the worst intra-reader agreement in the most experienced radiologists.
Chung et al. [24] investigated the inter-reader agreement between three experienced and three inexperienced radiologists. They reported a moderate inter-reader agreement amongst experienced radiologists (κ = 0.47, 95% CI: 0.41–0.54) and inexperienced radiologists (κ = 0.53, 95% CI: 0.47–0.58), and between them (κ = 0.47, 95% CI: 0.44–0.49). More recently, Alyami et al. [37] conducted a similar study between two radiologists on a small sample of 39 patients and reported fair to substantial agreements (κ = 0.21–0.78) using an arbitrary scoring system of five characteristics (composition, margin, shape, calcification, and vasculitis).
A recent study by de Carlos et al. investigated interobserver variability in thyroid ultrasound assessments among four endocrinologists evaluating 100 thyroid nodules [38]. They assessed ultrasound features, risk stratification using various classification systems (ATA, EU-TIRADS, K-TIRADS, and ACR-TIRADS), and the indication for fine needle aspiration (FNA). Their findings showed acceptable interobserver agreement for identifying nodules requiring cytological study but limited concordance in risk stratification and ultrasonographic characteristics, especially for margins. The Krippendorff’s alpha coefficient for margins was 0.41, indicating moderate agreement, which aligns with our finding of poor agreement in nodule margin scores between sonographers and radiologists (κ = 0.102).
Previously, only one study was found that evaluated the sonographers TI-RADS scores [39]. Fifteen sonographers were assigned to retrospectively score 100 nodules. They found fair to moderate agreement across most scores (κ = 0.21–0.52) but only slight agreement on margin (κ = 0.18, 95% CI: 0.16–0.20). The above studies underscore the inherent challenges in achieving uniformity in TI-RADS scoring, a factor that our study further elucidates by comparing different professional groups. Notably, while previous research focused on the variability among radiologists with various levels of experience and training, our study expands this understanding by including sonographers, highlighting the broader applicability of the TI-RADS system across different medical qualifications. The ROC analysis demonstrated that sonographers had a slightly higher AUC than radiologists, indicating marginally better diagnostic performance in distinguishing between benign and malignant nodules. This finding aligns with our earlier results showing higher specificity for sonographers.
Despite the good agreement noted between the sonographers and radiologists, the agreement for the nodule margin scores was poor. This finding also corresponds with the challenges noted in the literature with this score category between radiologists [23,35]. This is critical due to the importance of this category for predicting malignancy (PPV = 94%) [40]. Our study’s findings, supported by existing literature, indicate that distinguishing between ill-defined and irregular/lobulated margins can be challenging. Indeed, the ACR guidelines lack specific criteria regarding the number and extent of protrusions required to classify margins as lobulated or spiculated. Updating the ACR whitepaper [41], published in 2015, for the TI-RADS reporting lexicon of nodule characteristics to clarify any ambiguous definitions is warranted [42]. This may potentially improve the overall agreement of the TI-RADS score. Moreover, to improve the overall reproducibility within and between readers we recommend conducting practical training on the ACR TI-RADS user guide and lexicon to eliminate personal interpretation biases. This has demonstrated its effectiveness in previous studies [24,43].
Unlike most previous agreement studies, which lack histological confirmation, the current study investigated the diagnostic performance of ACR TI-RADS scores between sonographers and radiologists. While the sonographers appear to have a better diagnostic performance, this should be cautiously interpreted for two reasons. Firstly, a key limitation of our study is that it only included nodules classified as TR3 and above by radiologists, as we focused on samples with available histological Bethesda data. This means that our findings might not fully represent the diagnostic performance across the entire spectrum of thyroid nodules, particularly those in the lower-risk categories (TR1 and TR2), which are often evaluated in clinical settings. Moreover, it is essential to note that TI-RADS is primarily a system for risk-stratification for further investigation, such as fine needle aspiration (FNA), rather than a direct diagnostic tool for thyroid malignancy. This context is crucial because it frames the system’s role in guiding clinical decisions about the need for more invasive procedures. Another limitation of our study is its retrospective design and relatively small sample size, which may introduce selection bias. By including all consecutive eligible patients and strictly defining our criteria, we attempted to mitigate this risk. Future prospective studies with larger cohorts are recommended.
Our study adds to the understanding of how TI-RADS is scored by sonographers and radiologists. The use of TI-RADS scoring by sonographers may offer several advantages, such as reducing the need for radiologist intervention in routine cases, optimizing the triage of patients requiring further evaluation, and potentially shortening the time to diagnosis and treatment [44]. Moreover, accurate TI-RADS scoring by sonographers could contribute to more cost-effective healthcare delivery, as well as reducing the overall burden on radiology departments [45]. Empowering sonographers to perform TI-RADS interpretation could enhance their job satisfaction by recognizing their expertise and expanding their professional roles. This practice also improves clinical care efficiency by streamlining workflows, reducing radiologists’ workload, and allowing for more timely patient management. Additionally, advances in artificial intelligence (AI) and deep learning have the potential to impact TI-RADS scoring. Recent studies suggest that AI algorithms can enhance diagnostic accuracy and consistency in thyroid imaging by providing objective assessments and reducing interobserver variability [46]. Incorporating AI into TI-RADS evaluation could further optimize workflow efficiency and support clinicians in decision-making.
Future research should aim to include a wider range of TI-RADS categories and work on reducing the variability in interpretation to improve the system’s effectiveness in clinical decision-making. Finally, we recommend investigating the benefits and drawbacks of enabling conclusive sonographers’ TI-RADS scores in future research, which can yield valuable insights of the outcomes.
Conclusions
Comparing ACR TI-RADS scoring between sonographers and radiologists revealed moderate agreement, highlighting some challenges in achieving inter-reader reproducibility in thyroid nodule assessment. The poor agreement in nodule margin scores emphasizes the need for clearer lexicon definitions and training. The diagnostic performance of the sonographers’ TI-RADS scores was marginally superior to that of the radiologists when evaluated against histological outcomes. Our research, centered on suspicious nodules, highlights the importance of sonographer involvement in preliminary assessments, potentially enhancing efficiency in clinical settings. Future studies should aim to encompass a broader range of TI-RADS categories and focus on minimizing scoring variability.
Supporting information
S1 Table. Distribution of the TI-RADS features based on the Bethesda subcategories.
https://doi.org/10.1371/journal.pone.0312121.s001
(DOCX)
References
- 1. Gharib H, Papini E, Paschke R, Duick DS, Valcavi R, Hegedüs L, et al. American Association of Clinical Endocrinologists, Associazione Medici Endocrinologi, and European Thyroid Association medical guidelines for clinical practice for the diagnosis and management of thyroid nodules: Executive Summary of recommendations. J Endocrinol Invest. 2010;33: 287–291. pmid:20479572
- 2. Hegedüs L. The thyroid nodule. New England Journal of Medicine. 2004;351: 1764–1771.
- 3. Horvath E, Majlis S, Rossi R, Franco C, Niedmann JP, Castro A, et al. An ultrasonogram reporting system for thyroid nodules stratifying cancer risk for clinical management. J Clin Endocrinol Metab. 2009;94: 1748–1751. pmid:19276237
- 4. Kwak JY, Han KH, Yoon JH, Moon HJ, Son EJ, Park SH, et al. Thyroid imaging reporting and data system for US features of nodules: a step in establishing better stratification of cancer risk. Radiology. 2011;260: 892–899. pmid:21771959
- 5. Park J-Y, Lee HJ, Jang HW, Kim HK, Yi JH, Lee W, et al. A proposal for a thyroid imaging reporting and data system for ultrasound features of thyroid carcinoma. Thyroid. 2009;19: 1257–1264. pmid:19754280
- 6. Shin JH, Baek JH, Chung J, Ha EJ, Kim J, Lee YH, et al. Ultrasonography diagnosis and imaging-based management of thyroid nodules: revised Korean Society of Thyroid Radiology consensus statement and recommendations. Korean J Radiol. 2016;17: 370–395. pmid:27134526
- 7. Russ G, Bonnema SJ, Erdogan MF, Durante C, Ngu R, Leenhardt L. European Thyroid Association guidelines for ultrasound malignancy risk stratification of thyroid nodules in adults: the EU-TIRADS. Eur Thyroid J. 2017;6: 225–237. pmid:29167761
- 8. Tessler FN, Middleton WD, Grant EG, Hoang JK, Berland LL, Teefey SA, et al. ACR thyroid imaging, reporting and data system (TI-RADS): white paper of the ACR TI-RADS committee. Journal of the American college of radiology. 2017;14: 587–595.
- 9. Necas M, Shen Y, Ong QH, Prout K, Wackrow W. Do radiologists need to review abdominal ultrasound examinations reported as ‘normal’by the sonographer? Australas J Ultrasound Med. 2020;23: 167–175. pmid:34760596
- 10. Gibbs V, Edwards H, Harrison G. Independent reporting sonographers-could other countries follow the UK’s lead 2017. Imaging & Therapy Practice. 2017; 25–29.
- 11. Lo RH, Chan PP, Chan LP, Wilde CC, Pant R. Routine abdominal and pelvic ultrasound examinations: an audit comparing radiographers and radiologists. Ann Acad Med Singap. 2003;32: 126–128. pmid:12625111
- 12. Gerhardsen KV, Hofmann B. The accuracy of sonographers carrying out upper abdominal ultrasound examinations. Dieurope. 2014;30: 12–15.
- 13. Riley SJ, Groves CJ, Chandramohan M. Musculoskeletal ultrasound: audit of sonographer reporting. Ultrasound. 2010;18: 36–40.
- 14. Dawkins A, George N, Ganesh H, Ayoob A, Lee J, Nair R, et al. Radiologist and sonographer interpretation discrepancies for biliary sonographic findings: our experience. Ultrasound Q. 2017;33: 261–264. pmid:28399070
- 15. Leslie A, Lockyer H, Virjee JP. Who should be performing routine abdominal ultrasound? A prospective double-blind study comparing the accuracy of radiologist and radiographer. Clin Radiol. 2000;55: 606–609. pmid:10964731
- 16. Dongola NA, Guy RL, Giles JA, Ward S. Can sonographers offer an accurate upper abdominal ultrasound service in a district general hospital? Radiography. 2003;9: 29–33.
- 17. Persutte WH, Drose J, Spitz JL, Cyr D, Sansoucie DA, West FW, et al. Advanced-practice sonography in obstetrics and gynecology: a pilot study investigating the efficacy of the ultrasound practitioner. J Allied Health. 1999;28: 71–79. pmid:10389050
- 18. Riley SJ, Groves CJ, Chandramohan M. Musculoskeletal ultrasound: audit of sonographer reporting. Ultrasound. 2010;18: 36–40.
- 19. Schneider M, Bloesch J, Lombardo P. Abdominal ultrasound referred by the Emergency department–Can sonographer findings help guide timely patient management? Radiography. 2014;20: 4–7.
- 20. Castellana M, Castellana C, Treglia G, Giorgino F, Giovanella L, Russ G, et al. Performance of Five Ultrasound Risk Stratification Systems in Selecting Thyroid Nodules for FNA. J Clin Endocrinol Metab. 2020;105: 1659–1669. pmid:31690937
- 21. Haugen BR, Alexander EK, Bible KC, Doherty GM, Mandel SJ, Nikiforov YE, et al. 2015 American Thyroid Association management guidelines for adult patients with thyroid nodules and differentiated thyroid cancer: the American Thyroid Association guidelines task force on thyroid nodules and differentiated thyroid cancer. Thyroid. 2016;26: 1–133. pmid:26462967
- 22. Grani G, Lamartina L, Ascoli V, Bosco D, Biffoni M, Giacomelli L, et al. Reducing the number of unnecessary thyroid biopsies while improving diagnostic accuracy: toward the “right” TIRADS. J Clin Endocrinol Metab. 2019;104: 95–102. pmid:30299457
- 23. Liu H, Ma A-L, Zhou Y-S, Yang D-H, Ruan J-L, Liu X-D, et al. Variability in the interpretation of grey-scale ultrasound features in assessing thyroid nodules: A systematic review and meta-analysis. Eur J Radiol. 2020;129: 109050. pmid:32447147
- 24. Chung R, Rosenkrantz AB, Bennett GL, Dane B, Jacobs JE, Slywotzky C, et al. Interreader concordance of the TI-RADS: impact of radiologist experience. American Journal of Roentgenology. 2020;214: 1152–1157. pmid:32097031
- 25. Macedo BM de Izquierdo RF, Golbert L Meyer ELS. Reliability of Thyroid Imaging Reporting and Data System (TI-RADS), and ultrasonographic classification of the American Thyroid Association (ATA) in differentiating benign from malignant thyroid nodules. Arch Endocrinol Metab. 2018;62: 131–138. pmid:29641731
- 26. Frates MC, Benson CB, Charboneau JW, Cibas ES, Clark OH, Coleman BG, et al. Management of thyroid nodules detected at US: Society of Radiologists in Ultrasound consensus conference statement. Radiology. 2005;237: 794–800. pmid:16304103
- 27. Wolinski K, Szkudlarek M, Szczepanek-Parulska E, Ruchala M. Usefulness of different ultrasound features of malignancy in predicting the type of thyroid lesions: a meta-analysis of prospective studies. Pol Arch Med Wewn. 2014;124: 97–104. pmid:24473342
- 28. Middleton WD, Teefey SA, Reading CC, Langer JE, Beland MD, Szabunio MM, et al. Comparison of performance characteristics of american college of radiology TI-RADS, Korean Society of thyroid radiology TIRADS, and American Thyroid Association guidelines. American Journal of Roentgenology. 2018;210: 1148–1154. pmid:29629797
- 29. Remonti LR, Kramer CK, Leitao CB, Pinto LCF, Gross JL. Thyroid ultrasound features and risk of carcinoma: a systematic review and meta-analysis of observational studies. Thyroid. 2015;25: 538–550. pmid:25747526
- 30. Yoon JH, Lee HS, Kim E-K, Moon HJ, Kwak JY. Malignancy risk stratification of thyroid nodules: comparison between the thyroid imaging reporting and data system and the 2014 American Thyroid Association management guidelines. Radiology. 2016;278: 917–924. pmid:26348102
- 31. Choi SH, Kim E-K, Kwak JY, Kim MJ, Son EJ. Interobserver and intraobserver variations in ultrasound assessment of thyroid nodules. Thyroid. 2010;20: 167–172. pmid:19725777
- 32. Tessler FN, Middleton WD, Grant EG. Thyroid Imaging Reporting and Data System (TI-RADS): A User’s Guide. Radiology. 2018;287: 29–36. pmid:29558300
- 33. Cibas ES, Ali SZ. The 2017 Bethesda system for reporting thyroid cytopathology. Thyroid. 2017;27: 1341–1346. pmid:29091573
- 34. Anwar K, Mohammad AY, Khan S. The sensitivity of TIRADS scoring on ultrasonography in the management of thyroid nodules. Pak J Med Sci. 2023;39: 870. pmid:37250572
- 35. Li W, Sun Y, Xu H, Shang W, Dong A. Systematic Review and Meta-Analysis of American College of Radiology TI-RADS Inter-Reader Reliability for Risk Stratification of Thyroid Nodules. Front Oncol. 2022;12. Available: https://www.frontiersin.org/articles/10.3389/fonc.2022.840516. pmid:35646667
- 36. Daniels KE, Xu J, Liu J-B, Chen X, Huang K, Patel J, et al. Diagnostic value of TI-RADS classification system and next generation genetic sequencing in indeterminate thyroid nodules. Acad Radiol. 2021;28: 1685–1691. pmid:32839097
- 37. Alyami J, Almutairi FF, Aldoassary S, Albeshry A, Almontashri A, Abounassif M, et al. Interobserver variability in ultrasound assessment of thyroid nodules. Medicine. 2022;101: e31106. pmid:36254067
- 38. de Carlos J, Garcia J, Basterra FJ, Pineda JJ, Dolores Ollero M, Toni M, et al. Interobserver variability in thyroid ultrasound. Endocrine. 2024;85: 730–736. pmid:38372907
- 39. Wildman-Tobriner B, Ahmed S, Erkanli A, Mazurowski MA, Hoang JK. Using the American College of Radiology Thyroid Imaging Reporting and Data System at the point of care: sonographer performance and interobserver variability. Ultrasound Med Biol. 2020;46: 1928–1933. pmid:32507343
- 40. Srinivas MNS, Amogh VN, Gautam MS, Prathyusha IS, Vikram NR, Retnam MK, et al. A prospective study to evaluate the reliability of thyroid imaging reporting and data system in differentiation between benign and malignant thyroid lesions. J Clin Imaging Sci. 2016;6.
- 41. Grant EG, Tessler FN, Hoang JK, Langer JE, Beland MD, Berland LL, et al. Thyroid Ultrasound Reporting Lexicon: White Paper of the ACR Thyroid Imaging, Reporting and Data System (TIRADS) Committee. Journal of the American College of Radiology. 2015;12: 1272–1279. pmid:26419308
- 42. Solymosi T, Hegedűs L, Bonnema SJ, Frasoldati A, Jambor L, Karanyi Z, et al. Considerable interobserver variation calls for unambiguous definitions of thyroid nodule ultrasound characteristics. Eur Thyroid J. 2023;12. pmid:36692389
- 43. Grani G, Lamartina L, Cantisani V, Maranghi M, Lucia P, Durante C. Interobserver agreement of various thyroid imaging reporting and data systems. Endocr Connect. 2018;7: 1–7. pmid:29196301
- 44. Trimboli P, Ngu R, Royer B, Giovanella L, Bigorgne C, Simo R, et al. A multicentre validation study for the EU‐TIRADS using histological diagnosis as a gold standard. Clin Endocrinol (Oxf). 2019;91: 340–347. pmid:31002419
- 45. Kim PH, Suh CH, Baek JH, Chung SR, Choi YJ, Lee JH. Diagnostic performance of four ultrasound risk stratification systems: a systematic review and meta-analysis. Thyroid. 2020;30: 1159–1168. pmid:32303153
- 46. Yang WT, Ma BY, Chen Y. A narrative review of deep learning in thyroid imaging: current progress and future prospects. Quantitative Imaging in Medicine and Surgery. AME Publishing Company; 2024. pp. 2069–2088. pmid:38415152