Interrater reliability of quantitative ultrasound using force feedback among examiners with varied levels of experience

View article
PeerJ

Main article text

 

Rehabilitative ultrasound imaging (RUSI) is an approach to diagnostic sonography that incorporates both quantitative and qualitative assessment techniques to characterize musculoskeletal tissue and aid the implementation of therapeutic interventions (Harris-Love et al., 2014). RUSI applications are typically used to quantify post-intervention changes in tissue morphology, obtain joint space measures, provide visual biofeedback during therapeutic exercise, and further elucidate the contributions of muscle structure and neuromuscular activity to physical performance (Teyhen, 2007; Blazevich et al., 2007; Whittaker & Stokes, 2011).

Musculoskeletal assessment involving RUSI may feature quantitative techniques that differ from other uses of sonography (Harris-Love et al., 2016). Quantitative ultrasound imaging techniques are emerging as a non-invasive approach for describing muscle morphology in patients with neuromuscular disease and age-related dysfunction (Janssen et al., 2014; Ismail et al., 2015). However, this application of ultrasound is dependent on a specific set of examiner psychomotor skills, such as force, which may affect key measures of tissue dimensions and image echogenicity. Other investigators have demonstrated that variations in examiner scanning force may yield errors in the measurement of muscle tissue thickness (Ishida & Watanabe, 2012; Harris-Love et al., 2014). These previous investigations describing the impact of examiner performance on ultrasound image acquisition and quantitative assessment have involved both human subjects and ultrasound phantoms.

The concept that both normal and pathologic tissue can be simulated, thus allowing learners to acquire procedural skills while minimizing patient burden, has resulted in the development of an array of ultrasound phantoms and simulators. While the initial use of phantoms in sonography was driven by the need to calibrate ultrasound devices (Woo, 2002), tissue-mimicking ultrasound phantoms are now frequently used to train clinicians. Practitioner training experiences featuring tissue-mimicking ultrasound phantoms may be used to instruct ultrasound-guided invasive procedures or assist investigators in the development and validation of new ultrasound applications.

Given the potential examiner dependency associated with quantitative ultrasound techniques, we propose that the use of force-feedback scanning will foster the acquisition of reliable morphometry measures under a variety of scanning conditions. In this study, we determine the reliability of force-feedback image acquisition and morphometry over a range of examiner-generated forces using a muscle tissue-mimicking ultrasound phantom. We obtained reliability estimates for feedback-enhanced sonography using two methodological approaches. First, interrater reliability among the six examiners was determined based on material thickness measures obtained using manual force-feedback scanning and a series of applied force targets. Second, the criterion-based reliability of material thickness measures was determined by comparing the values obtained from each examiner using manual force-feedback scanning with the values obtained with robot-assisted force feedback scanning. We hypothesized that the manual force-feedback image acquisition method would yield reliable morphometry measures among the examiners, and in comparison to the criterion values obtained using robot-assisted image acquisition. In addition, we posited that the examiners would exhibit similar criterion-based reliability for morphometry measures, independent of experience level.

Materials and Methods

Ultrasound phantom

Examiners

Force feedback ultrasound materials and approach

Manual force feedback image acquisition

Robot-assisted image acquisition

Data analysis

Results

Measurement error from images acquired using force feedback scanning

Examiner interrater reliability across examiners and relative to the criterion-based reference measurements

Discussion

Additional Information and Declarations

Competing Interests

The authors declare there are no competing interests.

Author Contributions

Michael O. Harris-Love conceived and designed the experiments, performed the experiments, analyzed the data, contributed reagents/materials/analysis tools, wrote the paper, prepared figures and/or tables, reviewed drafts of the paper.

Catheeja Ismail and Reza Monfaredi conceived and designed the experiments, performed the experiments, contributed reagents/materials/analysis tools, wrote the paper, reviewed drafts of the paper.

Haniel J. Hernandez performed the experiments, wrote the paper, reviewed drafts of the paper.

Donte Pennington wrote the paper, reviewed drafts of the paper.

Paula Woletz, Valerie McIntosh and Bernadette Adams performed the experiments, reviewed drafts of the paper.

Marc R. Blackman analyzed the data, reviewed drafts of the paper.

Ethics

The following information was supplied relating to ethical approvals (i.e., approving body and any reference numbers):

The Washington DC VAMC Institutional Review Board (IRB; #01671) and Research and Development Committee.

Data Availability

The following information was supplied regarding data availability:

The data is owned by the US Department of Veterans Affairs and is therefore restricted.

Funding

This publication was partially supported by Award Number UL1TR000075 and UL1TR000101 from the NIH National Center for Advancing Translational Sciences (NCATS), National Institutes of Health (NIH), through the Clinical and Translational Science Awards Program (CTSA), and a VISN 5 Pilot Research Grant (VISN 5; VA Station: 688)—VHA/VA Capitol Health Care Network. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

17 Citations 3,465 Views 569 Downloads