Reporting Guidelines for Music-Based Interventions: An update and Validation Study☆
Article information
Abstract
Background
Detailed intervention reporting is essential to interpretation, replication, and translation of music-based interventions (MBIs). The 2011 Reporting Guidelines for Music-Based Interventions were developed to improve transparency and reporting quality of published research; however, problems with reporting quality persist.
Methods
The purpose of this study was to update and validate the 2011 reporting guidelines using rigorous Delphi approach that involved an interdisciplinary group of MBI researchers; and to develop an explanation and elaboration guidance statement to support dissemination and usage. We followed the methodological framework for developing reporting guidelines recommended by the EQUATOR Network and guidance recommendations for developing health research reporting guidelines. Our three-stage process included: (1) an initial field scan, (2) a consensus process using Delphi surveys (two rounds) and Expert Panel meetings, and (3) development and dissemination of an explanation and elaboration document.
Results
First-Round survey findings revealed that the original checklist items were capturing content that investigators deemed essential to MBI reporting; however, it also revealed problems with item wording and terminology. Subsequent Expert Panel meetings and the Second-Round survey centered on reaching consensus for item language. The revised RG-MBI checklist has a total of 12-items that pertain to eight different components of MBI interventions including name, theory/scientific rationale, content, interventionist, individual/group, setting, delivery schedule, and treatment fidelity.
Conclusion
We recommend that authors, journal editors, and reviewers use the RG-MBI guidelines, in conjunction with methods-based guidelines (e.g., CONSORT) to accelerate and improve the scientific rigor of MBI research.
Introduction
Detailed intervention reporting is essential to interpretation, replication, and eventual translation of music-based interventions (MBIs) into practice. Persistent problems with the reporting quality of MBIs represent a significant barrier to advances in scientific research and translation of findings to clinical practice and community settings [1–4]. Interest in the quality of published research reports emerged in the 1980s due to growing awareness about deficiencies in reports of clinical trials at the time [5,6]. For example, several studies at this time found that an increasing number of randomized controlled trials (RCTs) had missing or inaccurate information, such as whether the assessment of outcomes was masked, a primary endpoint specified, or how sample size was determined [5,7,8]. As a result, the use of reporting guidelines was recommended.
Reporting guidelines are a simple, structured tool for health researchers to use while writing manuscripts, which provides a minimum list of information needed to ensure a published manuscript can be understood by a reader, replicated by a researcher, used to inform clinical decisions, and included in systematic reviews [9]. The Consolidated Standards for Reporting Trials (CONSORT) and Transparent Reporting of Evaluations with Non-randomized Designs (TREND) guidelines were developed to improve the quality and transparency of published research [10,11]. Subsequent publications centered on complexities related to the reporting of behavioral and non-pharmacological interventions, noting that CONSORT and TREND, which have only one item dedicated to intervention reporting, were inadequate [12–15]. This led to the development of supplemental guidelines specific to intervention reporting, including elaborated CONSORT guidelines for reporting non-pharmacological interventions [12,13] and the Template for Intervention Description and Replication (TIDieR) checklist [16].
Music-based interventions are especially difficult to fully describe, due to the complexity of music stimuli (e.g., rhythm, pitch, tempo, harmonic structure, timbre), variety of music experiences (e.g., active music making, music listening) and other factors unique to MBIs. To determine whether intervention reporting guidelines were necessary, Robb and Carpenter [17] examined how authors described music interventions and found significant gaps in reporting that hinder cross-study comparisons, generalization, and integration of findings into practice. Subsequently, Robb, Burns, and Carpenter [18] developed Reporting Guidelines for Music-Based Interventions (RG-MBI), which specified components of music interventions that publishing authors were encouraged to report and discuss [18]. The checklist included 11 items organized across seven component areas including intervention content (five items), theory, delivery schedule, interventionist, treatment fidelity, setting, and unit of delivery (one item each).
The 2011 RG-MBIs are available through the EQUATOR Network [19] and have been referenced by authors in > 430 publications. However, recent reviews reveal sustained problems with reporting quality [20–33]. In their 2018 review of MBI reporting quality, Robb et al [4] found overall reporting quality was poor with fewer than 50% of authors reporting information for four of the seven checklist components (theory, interventionist qualifications, treatment fidelity, setting). Reporting of intervention content was also poor; again, fewer than 50% of authors reported information about the music used, decibel levels/controls, or materials [4].
Sustained problems with reporting quality suggest limited uptake by authors and journal editors of the 2011 music reporting guidelines. This may be due to limited awareness of those guidelines, problems with perceived relevance or clarity of checklist items, and/or the absence of an explanation and elaboration document to provide practical examples across diverse areas of MBI intervention research. Thus, to ensure validity of current checklist items and improve uptake of the reporting guidelines, we completed a rigorous process to update the current guidelines and to establish a process by which to disseminate the resulting validated checklist and guidance statements.
Methods
We followed the methodological framework for developing reporting guidelines recommended by the EQUATOR Network [34] and recommendations for developing health research reporting guidelines [35]. The lead author convened a nine-member advisory group that included leaders from the National Institutes of Health (NIH) Music and Health initiative, music intervention researchers, and policy advocates (see acknowledgements). The advisory group worked with authors SR and SS to develop the study protocol and registered the RG-MBI update with the EQUATOR network [36]. Here we report methods and findings from our three-stage process: (1) field scan, (2) consensus process including Delphi survey and Expert Panel, and (3) resulting modified checklist and planned explanation and elaboration (E&E) guidance statement. This study did not meet criteria for Human Subjects Research and was exempt from Institutional Review Board approval.
1. Stage 1: Field scan
In 2018, based on items specified in the RG-MBI, Robb et al [4] examined reporting quality of published music intervention studies [4]. Overall, reporting quality was determined to be poor, and the terminology used to describe interventions was varied and inconsistent. Golden et al [3] found similar problems with reporting, and recommended the generation and uptake of reporting guidelines [37].
Building on these two reviews, and as our first step, authors SR and JM conducted a field scan of systematic reviews of MBIs published between 2018 and 2022. The purpose of the field scan was to examine and elucidate gaps in reporting quality to inform our Delphi survey and processes. Specifically, we examined whether authors of the systematic reviews discussed reporting quality and, if so, whether they identified additional problems not captured in the current guidelines. We identified 33 systematic reviews, 48% (n = 16) of which discussed specific problems with reporting quality. Notably, all the identified problems had been captured by the 2011 MBI reporting guidelines, suggesting limited awareness or uptake of those early guidelines. As such, the field scan findings supported the use of the 2011 RG-MBI checklist as the starting point for a subsequent Delphi Survey process; it also indicated the need to engage stakeholders and interdisciplinary experts to improve content, item clarity, and usage of the guidelines (see Supplement 1) [38].
2. Stage 2: Item revision and consensus (Delphi survey and expert panel)
The purpose of Stage 2 was to invite music intervention researchers to evaluate content of the 2011 MBI checklist; specifically, they were asked to determine the importance of each item, identify gaps in content, identify problems with wording, and to reach consensus regarding recommended changes to the checklist. Our Delphi process, based on methods described by Sinha and colleagues [37], included two survey rounds to reach item consensus, with the plan to add additional rounds as needed. Following each survey round, an Expert Panel reviewed all survey data and made final consensus decisions concerning checklist items. In this section we provide details about the Expert Panel, survey participants, and methods for reaching consensus.
2.1. Participants
2.1.1. Interdisciplinary expert panel
The Advisory Group worked with lead authors SR and SS to identify expert panelists with varied expertise and who represent different stakeholder groups engaged in the design, conduct, and dissemination of music and health research. Selection criteria were to identify investigators conducting research: (1) along the translational science continuum, (2) across various domains (sociological, psychological, clinical, community health), (3) with varied methodological expertise, and (4) from a variety of disciplinary backgrounds. This eleven-member panel (authors EE, TG, JJ, DB, MB, JB, CG, AH, JI, ML, and SP) included individuals with expertise in the design, conduct, dissemination, and publication of music and behavioral intervention research from the United States, Europe, and Canada. The group included authors of the original MBI reporting guidelines, journal editors, and researchers with expertise in music cognition and neuroscience, music therapy, intervention research, biostatistics, and community music interventions.
2.1.2. Survey participants
Individuals invited to participate in the Delphi survey included United States-based and international experts in music and music-based intervention research. The target sample was comprised of Cochrane review authors, NIH MBI Toolkit panelists, journal editors, and established authors/investigators (including NIH funded Music and Health grant recipients and authors of systematic reviews identified through our initial field scan), and representatives from patient advocacy and arts organizations. Professional backgrounds included behavioral health, neuroscience, nursing, medicine, music therapy, social work, psychology, and public health. The target sample included 106 experts for Round One and 103 experts for Round Two. Accepting the invitation to complete the survey constituted participants’ consent to participate.
2.2. Round one survey
The survey opened with a brief overview of the survey purpose, defined key terms, provided an estimated time commitment (including number of rounds), and emphasized the importance of completing each round. Each reporting item from the original guidelines (12 items total), was assigned an identification number to facilitate random ordering. Participants were asked to rate the importance of each item on a four-point Likert Scale (1 = item has limited importance and not required for reporting; 2 = item has moderate importance; 3 = item has high importance; 4 = item has very high importance and essential to reporting). For each item, participants could also provide additional comments or edits to improve the reporting criterion. For items that received a rating of “1= limited importance” or “4 = very high importance,” we asked participants to provide their rationale for selecting that value and to include any references to support their rationale, if possible. The final two survey items asked participants for additional criteria they believed should be reported in published music intervention research (Question 13) and any additional comments they wanted to share about their responses or the survey (Question 14). See Supplement 2 for survey [38].
2.2.1. Round one data collection and sample
To ensure confidentiality, the Indiana University Center for Survey Research (CSR) distributed and managed survey data using a Qualtrics web survey and recruitment via e-mail. Potential participants were sent an email and invitation; non-respondents and respondents who did not answer all 12 of the first 12 questions received up to two e-mail reminders. To bolster responses, the first author personally e-mailed non-respondents to request their response before the third and final CSR reminder. Additionally, a special reminder with a separate survey link for Questions 13 and 14 only was sent to respondents who partially completed the survey but had not made it to these questions. The first-round survey opened November 3, 2022, and closed January 17, 2023.
The Round One survey was sent to 103 experts for completion after removing three who self-identified as ineligible. The final sample for Round One analysis involved 65 respondents (including partial and complete responses) for a response rate of 63%. Median time to complete the survey was 14 min and an IQR of 23.2 min excluding outliers (> 70 min). Outliers included 5 respondents with survey times between 107.90 and 341.42 min, and 6 respondents over 1160 min. We excluded these cases because they represent individuals who filled out the survey but never submitted it or selected submit after some time with it open in their browser.
2.2.2. Round one analysis and expert panel meeting consensus
All data from completed surveys were downloaded to an Excel spreadsheet for descriptive analysis. Likert scores were grouped based on the four response categories: Limited importance (1); Moderate Importance (2); High importance (3); Very high importance (4). For each item, we calculated descriptive statistics for each response category (frequency, percent, valid percent, cumulative valid percent). Consensus criteria for retaining an item was defined as ≥80% of survey respondents rating an item as having “High” or “Very High” importance. Items scoring lower than this threshold were reviewed by the Expert Panel to determine inclusion, removal, and/or refinement of the item for the second-round survey. In addition, comments provided in open-response fields for all items, as well as any suggested additional items (Question 13), were downloaded verbatim for analysis. Two independent reviewers (SR, SS) identified common themes, and then discussed independent findings to reach agreement. In advance of the first Expert Panel meeting, panelist received numeric data, common themes, and representative statements for each item, along with a synthesized list of any new items and related comments.
Expert Meeting panelists were charged with discussing and reaching consensus about: (1) item retention/removal based on numeric and narrative survey data, (2) item level revision based on narrative data, and (3) inclusion of any newly identified items. During meetings, a meeting facilitator invited each panelist to share a unique comment or insight, with an option to pass or affirm another’s comment. Once each panelist had the opportunity to comment, the group assessed whether they had reached consensus. Originally, the authors had identified nominal group technique as its planned approach to reach consensus; however, the panel did not require voting or ranking to achieve consensus on each item.
2.2.3. Round one survey and expert panel meeting results
Findings from the Round One survey are available in Supplement 2 [38]. Three items did not reach the threshold score for consensus (≥80% of survey respondents rating the item as having “High” or “Very High” importance); these included Q4: Music (78%), Q6: Intervention Materials (64%), and Q11: Setting (75%). Associated comments pointed to the need for revised language (Q4; Q6), with some suggestions that Q11 could be removed and captured in methods-specific checklists. Consensus from the Expert Panel was that current checklist items were adequate, important, and relevant (no items removed or added). However, there was also consensus that wording/language for all checklist items needed revision, and that the revision process should be the focus of the Round Two Survey. To inform revised item language for the second survey, we used discussion notes from the Expert Panel meeting, and gave panelists time after the meeting to submit more detailed edits. Lead authors (SR, SS) then synthesized these recommendations to create revised item language for the second survey.
2.3. Round two survey
All eligible participants from the first-round survey (n = 102; one person removed by request) were invited to complete the second-round survey which provided a side-by-side comparison of checklist items (original vs. revised). For each item (12 items total), participants were asked to indicate one of three options: (1) prefer original checklist wording; (2) prefer revised checklist wording; (3) a suggested edit (with open text box to provide revised wording). See Supplement 3 for survey [38].
2.3.1. Round two data collection and sample
Invitation and reminder e-mails followed the same structure and frequency as Round One. The survey opened May 31, 2023, and closed July 18, 2023. The final sample for Round Two analysis involved 61 respondents (including partial and complete responses) for a response rate of 60%. Median time to complete the survey was seven minutes and an IQR of 5.3 min excluding outliers (> 70 min). Outliers included 9 respondents with survey times between 70.5 and 965.5 min, and 9 respondents over 3,273 min. We excluded these cases because they represent individuals who filled out the survey but never submitted it or selected submit after some time with it open in their browser.
2.3.2. Round two analysis
To determine whether there was consensus for original or revised items we calculated frequency, mean, and percent scores for each item. Consensus was defined as items that were selected by ≥ 80% of survey respondents. In addition, the panel used open-ended comments from survey respondents to determine if an item required further revision. The Expert Panel’s charge was to review items that did not reach consensus using discussion as well as survey respondents’ open-ended comments to inform final changes to item wording, sentence structure, or organization.
2.3.3. Results round two survey and expert panel meetings
Findings from the Round Two survey are available in Supplement 3 [38]. Survey respondents preferred revised language for all items; however, three items did not reach the threshold for consensus (≥ 80% of respondents preferred revised item language): Q2: Person Selecting the Music (63%); Q3: Music (74%); Q9: Treatment Fidelity (52%). For all items, we received suggestions on how we could improve item language. The Expert Panel had two subsequent meetings in which they discussed survey respondent recommendations, terminology, whether to include embedded examples, and the ultimate order of checklist items (including alignment with TIDieR and CONSORT Non-Pharmacological checklists). All Expert Panel decisions were made using our a priori consensus threshold of ≥ 80% agreement.
Results
Discussion
The 2011 Reporting Guidelines for Music-Based Interventions were developed to improve transparency and reporting quality of published research. Despite an increased number of publications citing the guidelines, recent reviews indicate persistent problems with reporting quality. Incomplete and inconsistent reporting of MBIs impedes cross-study comparisons, interpretation, replication, and application of findings to clinical practice and community-based programming.
To improve uptake of the RG-MBIs by a larger and more diverse group of MBI researchers, we convened a team of experts from diverse disciplines to engage in a rigorous Delphi study process. This process revealed that the original checklist items were indeed capturing content that investigators deemed essential to MBI reporting; however, it also identified important problems with existing items that may have been affecting its uptake and effective usage. In particular, findings indicted changes in wording and terminology that would allow checklist items to be inclusive of a wide range of music experiences (e.g., music as a sound stimulus and creating music/musicking) and approaches (e.g., social, psychological, physical, neurological, and biological). The illumination of these issues resulted in robust discussion among Expert Panelists and several rounds of revisions to item language in the guidelines. By engaging an international and diverse group of experts to revise item language, our expectation is that the revised checklist will be clearer, easier to apply, and of greater relevance for a diverse group of MBI investigators.
To further facilitate usage, items were re-ordered to align with the TIDieR checklist including the addition of item one from the TIDieR checklist [16]. Expert panel members also co-authored an Explanation and Elaboration (E&E) guidance document to companion the revised RG-MBI [39]. This document includes a rationale for each item, concrete instructions for optimally reporting each item, and annotated examples from published manuscripts. Our expectation is that the revised RG-MBI will be of greater utility to investigators across a wider range of disciplines and that the E&E document will support greater adoption of the RG-MBI by authors and journal editors.
A primary limitation of this validation study was limited representation of investigators and stakeholders from countries outside the United States. Reliance on systematic reviews, Cochrane Reviews, journal editors, and US-based research initiatives to generate our survey sample did not ensure representation of music and health researchers, clinicians, and advocates at a global level. Second, we did not obtain information about survey respondents’ professional background and country, limiting our ability to assess representation. Finally, we did not conduct a formal study to investigate researchers’ awareness of the 2011 RG-MBIs to gain further insight into specific barriers to adoption.
We recommend that authors, journal editors, and reviewers use the RG-MBI guidelines, in conjunction with methods-based guidelines like CONSORT and TREND, to accelerate and improve the scientific rigor of MBI research. We also recommend a review of MBI reporting quality in five years to evaluate the impact of the revised guidelines and subsequent international studies centered on RG-MBI utility, along with barriers and facilitators to their adoption.
Supplementary Materials
Supplementary materials are available at doi:https://doi.org/10.56986/pim.2025.10.009.
Notes
Acknowledgements
Special thanks to our Advisory Panel members: Wen G. Chen, PhD; Emmeline Edwards, PhD; Tasha Golden, PhD; Julene Johnson, PhD; Susan Magsamen, MAS; Coryse St. Hillaire-Clarke, PhD; Dana Greene-Schloesser, PhD; Stacey Springs, PhD. Special thanks to Elizabeth Harman, PhD, MT-BC for assistance with reference management and formatting.
Author Contributions
Conceptualization, Data curation, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Supervision, Validation, Visualization, Writing - original draft, Writing - review & editing: SLR. Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Supervision, Validation, Visualization, Writing - original draft, Writing - review & editing: SS. Conceptualization, Investigation, Validation, Writing - original draft, Writing - review & editing: EE. Conceptualization, Investigation, Validation, Writing - original draft, Writing - review & editing: TLG. Conceptualization, Investigation, Validation, Writing - original draft, Writing - review & editing: JKJ. Investigation, Validation, Writing - original draft, Writing - review & editing: DSB. Investigation, Validation, Writing - original draft, Writing - review & editing: MB. Investigation, Validation, Writing - original draft, Writing - review & editing: JB. Investigation, Validation, Writing - original draft, Writing - review & editing: CG. Investigation, Validation, Writing - original draft, Writing - review & editing: AH. Investigation, Validation, Writing - original draft, Writing - review & editing: JRI. Investigation, Validation, Writing - original draft, Writing - review & editing: ML. Investigation, Validation, Writing - original draft, Writing - review & editing: JAM. Investigation, Validation, Writing - original draft, Writing - review & editing: SMP.
Conflicts of Interest
The authors declare that they have no conflicts of interest.
Author Use of AI Tools Statement
The authors declare that no Generative AI was used in the creation of the manuscript.
Funding
The author(s) declare that financial support was received for the research and/or publication of this article. This study was funded by the Walther Cancer Foundation through Dr. Sheri Robb’s endowed professorship.
Ethical Statement
The Human Research Protection Program of the Indiana University waived the need for ethics approval and oversight for the collection, analysis, and publication of anonymized data for this non-interventional study. Invited survey participants were provided details about the study (purpose, duration, and procedures) and that individual responses would be kept confidential. Accepting the invitation to complete the survey constituted participants’ consent to participate.
Data Availability
De-identified data supporting the conclusions of this article are included in online supplementary material (see original article: https://www.frontiersin.org/journals/psychology/articles/10.3389/fpsyg.2025.1551920/full#SM1). Further inquiries can be directed to the corresponding author.
