Skip to main content

A cross-cultural investigation into the dimensional structure and stability of the Barriers to Research and Utilization Scale (BARRIERS Scale)



It is important that scales exhibit strong measurement properties including those related to the investigation of issues that impact evidence-based practice. The validity of the Barriers to Research Utilization Scale (BARRIERS Scale) has recently been questioned in a systematic review. This study investigated the dimensional structure and stability of the 28 item BARRIERS Scale when completed by three groups of participants from three different cross-cultural environments.


Data from the BARRIERS Scale completed by 696 occupational therapists from Australia (n = 137), Taiwan (n = 413), and the United Kingdom (n = 144) were analysed using principal components analysis, followed by Procrustes Transformation. Poorly fitting items were identified by low communalities, cross-loading, and theoretically inconsistent primary loadings, and were systematically removed until good fit was achieved. The cross-cultural stability of the component structure of the BARRIERS Scale was examined.


A four component, 19 item version of the BARRIERS Scale emerged that demonstrated an improved dimensional fit and stability across the three participant groups. The resulting four components were consistent with the BARRIERS Scale as originally conceptualised.


Findings from the study suggest that the four component, 19 item version of the BARRIERS Scale is a robust and valid measure for identifying barriers to research utilization for occupational therapists in paediatric health care settings across Australia, United Kingdom, and Taiwan. The four component 19 item version of the BARRIERS Scale exhibited good dimensional structure, internal consistency, and stability.


Evidence-based practice (EBP) is considered to be a combination of recent best quality evidence, clinical expertise and analysis as well as the values, expectations and views of the patient [1]. EBP is a significant factor in assuring that practice is systematic and guided by current research and that its implementation is associated with enhanced clinical outcomes in patients [2]. Research utilization is directly linked to EBP, and involves searching for, sourcing, critiquing, integrating, translating, and applying best evidence into daily professional practice.

Health professionals value EBP and research utilization, however, its implementation is often obstructed by factors such as time, skill, funding and access limitations. This trend has been reported in several health care fields including speech-language pathology, dietetics, physiotherapy [3], nursing [4] and occupational therapy [5, 6]. Parahoo [7] identified that increased EBP and continual developments in health care tertiary education were correlated with recognizing barriers to research utilization. This recognition of barriers was said to inform the most appropriate interventions, which subsequently would increase the implementation and integration of research findings into clinical practice [8], and result in enhanced patient outcomes [9]. Differences in barriers perceived in different clinical settings, cultural contexts, and health care professions further supports the importance of distinguishing barriers to research utilization [3]. Therefore having access to a valid scale that explores barriers to EBP and research utilization is timely and significant.

Development of the BARRIERS Scale

The Barriers to Research and Utilization Scale, commonly known as the BARRIERS Scale, was developed in 1991 by Funk, Champagne, Wiese and Tornquist to measure barriers to research utilization as experienced by nurses [10]. Its development was focused on nurses’ perspectives as previously emphasis of identifying barriers was restricted mainly to the observations of educators and administrators [11].

The BARRIERS Scale is a self-report questionnaire, which was constructed initially in the United States with 1989 nurse participants, with credentials comprising of doctorates, masters, bachelor’s, diplomas and associate degrees [12]. The development of the scale was informed by literature, unofficial testimonies by nurses, as well as the Conduct and Utilization of Research in Nursing (CURN) questionnaire [13]. A total of 29 items were included in the initial version of the BARRIERS Scale, however the item ‘the amount of research information is overwhelming’ was found incompatible with scale’s factor structure after a factor analysis was completed and consequently this item was deleted from the scale [14].

The BARRIERS Scale is comprised of 28 items that quantify a respondent’s perceived barriers that negatively impact research utilization and EBP. Respondents rate these barriers using a four-point Likert scale and an additional “no opinion” option. Confirmatory factor analysis and Roger’s Model of Diffusion of Innovations [15] were used to identify and inform four barrier factors: (1) characteristics of the adopter; competency, ideals, awareness and receptiveness to research (eight items), (2) characteristics of the organization; practice setting (eight items), (3) characteristics of the innovation/research; attributes of the research (six items), and (4) characteristics of communication; ease of application, accessibility, and comprehension of the research (six items) [16].

Administration and application of the BARRIERS Scale has been broad across the nursing profession. A variety of different sample groups in the nursing sector have been surveyed including Pediatric Nurses, Registered Nurses, Breast Cancer Nurses, Clinical Administrators, and a multitude of practice contexts have been incorporated including palliative care, community settings, hospitals, surgical and medical practice and mental health [17]. Other specialised nursing fields have also been administered the scale, including peri-anaesthesia nurses [18] and nurses working in intensive or critical care domains [19]. Educators have completed the scale in both acute and intensive care nursing settings [20]. The BARRIERS Scale has also been completed by other health professionals including occupational therapists, physiotherapists, dieticians, and speech language pathologists [21, 22].

Construct validity and dimensionality of the BARRIERS Scale

Much of the research involving the BARRIERS Scale has relied on rank ordering of individual items to determine the barriers to research utilization and allow for comparison across settings and professions [12]. Underlying this approach is the assumption of equivalence of single item performance across samples (both professions and cultures), despite the evidence to suggest that single items measure have lower reliability, are more vulnerable to the effects of error, and are comparatively poor predictors compared to multiple item constructs [23]. Perhaps some of the use of rank ordering of individual items is understandable given that the BARRIERS Scale has demonstrated some difficulties in replicating the theorized factor structure across samples and professions.

During the initial development, Funk et al. tested the stability of the BARRIERS Scales by dividing a sample group into two before performing a factor analysis for each [10]. Equivalent results were achieved after comparing the two factor analyses with the analysis of the entire sample: the loadings remained the same, as did the identified four factors, minimal changes in explained variance percentages were evident (sample 1; 43.4 %, sample 2; 44.9 %), a fourth factor item did not match factors three and four, but this was resolved through the final analyses. The authors of the scale have endorsed its stability in a variety of nursing contexts, and across time, despite differences occurring in numbers ascribed to individual barriers [7].

The BARRIERS Scale’s internal consistency was examined by the scale’s authors. Cronbach’s alpha coefficients ranging between 0.65 and 0.80 were reported [10]. Factors 1–3 of the scale demonstrated reasonable levels of internal consistency with Cronbach alpha coefficients of 0.75, 0.74 and 0.70 reported. Factor 4 was documented at 0.54 [24] suggesting improvements to this factor’s internal consistency could be made.

Face and content validity has been examined by others and has been found to be appropriate given that the development of the scale utilized reports from nurses’, the CURN questionnaire and literature investigating utilization of research, additionally specialists in nursing research, nursing practice, research utilization, and an expert in quantitative measurement [14]. Bayik et al. considered the scale’s validity and reliability by evaluating culturally diverse studies that used the scale [25].

Dimensional stability plays an important role in construct validity, and as previously stated, the replicability of the BARRIERS Scale factor structure has been mixed when explored in published studies. For example, Kajermo et al. completed a systematic review of studies using the BARRIERS scale [17]. Of the 63 articles examined in the review, there were a number of issues were raised surrounding the construct validity of the scale. The dimensionality of the BARRIERS scales ranged between three and eight factor solutions using both factor analysis and confirmatory factor analysis [17]. These findings suggest further psychometric investigation needs to be undertaken, including cross-cultural appraisals.

The validity of the BARRIERS Scale has been challenged when considering its wide use across clinical practice settings and cross-cultural settings. Kajermo et al. argued that changes in health care, the discipline of nursing and patient involvement have resulted in transformations in research utilization [17]. But in spite of this, the barrier factors identified by the scale remain unchanged; the theories underpinning the scale are reflective of the late 1980′s and early 1990’s health care settings, and its initial design is the one most utilized in studies. The BARRIERS Innovation subscale was also argued to be inadequate as it was suggested to have been simplified [17] and not truly reflective of Rogers’ Diffusion of Innovation Model [15]. One factor identified as absent from the BARRIERS Scale was related to the speed of adoption of innovation, which comprises of “relative advantage, compatibility, complexity, trialability, and observability” [26].

A study conducted in the United Kingdom employed the BARRIERS Scale to determine whether barriers to research utilization encountered by nurses working in British contexts were captured by the factors of the BARRIERS Scale [27]. The factor model was found to be unsuited to the United Kingdom data and comparatively to results obtained by the authors in the initial studies of the scale, with poorer validity, and lower reliability. Cross-cultural language was considered to impact this, and slight revisions were made to the wording of 18 scale items, however, only slight improvements in reliability were achieved [9].

Using a principal components analysis (PCA) with Varimax Rotation, Retsas and Nolan obtained a 26 item, three-factor model that accounted for 38.9 % of the variance [28]. Using the same methodology, Retsas generated a 29 item, four factor model accounting for 46.5 % of the variance [11]. In an Australian study, Hutchinson and Johnston obtained a 27 item, four factor model that accounted for 39.2 % of the variance using PCA [24]. In an American study, Ashley obtained a 29 item, four factor solution [29]. The factor analytic studies completed by Ashley [29] and Hutchinson and Johnston [24] were similar to the factor model originally reported by Funk et al. [10] These studies shed some light on the dimensional structure of the BARRIERS Scale.

Closs and Bryar investigated the appropriateness of the BARRIERS Scale for use within the context of the United Kingdom health care system [12]. The BARRIERS Scale was sent to 4,501 nurses, with a 44.6 % response rate. Using PCA with Varimax Rotation, Closs and Bryar found a 22 item, four factor solution which explained 47.5 % of the variance [12]. The four factors were labelled benefits; quality of research; accessibility of research; and resources for implementation. These were similar, but not identical, to factors identified in the original United States study. Closs and Bryar concluded that the four factor solution they obtained was similar in principle to the United States version of the BARRIERS Scale, although fewer items were retained (22 rather than 28) [12]. This indicates that there appears to be some stability in the BARRIERS Scale dimensional structure, however, the variability in number of items retained suggests that problematic items remain.

Given the mixed results of the psychometric properties of the BARRIERS Scale among nursing cohorts, further investigation of the dimensionality is warranted with other health professional participant groups from different cross-cultural contexts. Therefore the aim of this study was to investigate the dimensional structure and stability of the 28 item BARRIERS Scale when completed by three occupational therapy cohorts from three different cross-cultural environments.



A mailed survey was administered to participants in Australia, the United Kingdom, and Taiwan. Including participants from three countries provided insights into the dimensional structure and stability of the BARRIERS Scale.


The membership data bases from OT Australia, College of Occupational Therapists in the United Kingdom, and the Taiwan Occupational Therapy Association (TOTA), were used to invite occupational therapists who indicated that pediatrics was their clinical specialty to participate in the study. The inclusion criteria for this study were: (1) consenting to participate in the study; and (2) being a qualified occupational therapist in Australia, the United Kingdom, or Taiwan. Ethics Committee approval was obtained from Monash University, Australia; the British Association of Occupational Therapists; and the Taiwan National University Hospital prior to the commencement of the study in each respective country. The justification for including participants from three different countries was that it provided an opportunity for the dimensional structure and stability of the BARRIERS Scale to be examined when completed by respondents from three different cross-cultural contexts but who were from the same health care profession.


The BARRIERS Scale was used to measure participants’ perceived barriers to research utilization. This scale has been widely used with nurses [4, 30] and to a lesser extent, allied health professionals [3, 12]. The BARRIERS Scale contains four subscales derived from confirmatory factor analysis: (1) the Adopter (values, skills, and awareness); (2) the Organization (setting); (3) the Innovation (qualities of the research); and (4) Communication (presentation and accessibility of the research). The 28 items are rated according to the degree to which the respondent perceives the item to be a research barrier, rated from 1 (‘to no extent’), to 4 (‘to a great extent’). The “no opinion” response option was not provided in the current study (see discussion for further comments). The authors reported good internal consistencies of the first three factors (Cronbach’s α of 0.72–0.80), lower internal consistency for the fourth factor (Cronbach’s α of 0.65), and preliminary evidence of test–retest reliability with Pearson correlations ranging from 0.68 to 0.72 over a one week interval [10]. The BARRIERS Scale can be accessed at or from the Funk et al. article [10]. The BARRIERS Scale has been widely used in many studies investigating barriers to research utilization [16] and has been translated previously into Turkish, German, Thai, Korean, and French. Recently the BARRIERS Scale was used to investigate barriers to EBP and research utilization with occupational therapists in Sweden, although it should be noted that no attempt at validating the structure of the instrument was undertaken [21].

For the version of the BARRIERS Scale used in this study, the word ‘nurse’ was replaced with ‘clinician’ as the participant group were occupational therapists. This did not alter the meaning or relevance of the BARRIERS Scale items. The version of the BARRIERS Scale completed by the participants in Taiwan was in Mandarin Chinese. It was translated from English to Mandarin Chinese by a qualified translator. The Mandarin Chinese version of the BARRIERS Scale was then reviewed by a panel of three bilingual (Mandarin Chinese–English) occupational therapists for phrasing, diction, understanding, and coherence. No changes to the scale were suggested by the panel. The Mandarin Chinese version of the BARRIERS Scale was then back-translated into English by another qualified translator. The back-translated English version of the BARRIERS Scale was then compared to the original version, in accordance with the translation procedures described by Cha et al. [31] and Wang et al. [32]. The items from the two versions of the BARRIERS Scale (English version and Taiwanese version) were similar in meaning, content, and wording.

Data collection

Survey packages were mailed to prospective participants who met the inclusion criteria, and a reminder letter was sent 2 weeks following the initial survey distribution, as a method to facilitate an increased response rate. Survey responses were recorded anonymously since participants were not asked to report any personally identifying information and consent was inferred through completion of the questionnaire. Entry of the data obtained from respondents in Taiwan was completed in Taiwan while data entry for the Australian and United Kingdom data was completed in Australia. All surveys that were returned were complete, so missing data or incomplete questionnaires were not an issue.

Statistical analysis

The BARRIERS scale was subjected to principal components analysis with Varimax rotation using Statistical Package for Social Sciences (SPSS), IBM version 20. To evaluate the dimensional structure and stability across samples, a Procrustes transformation was employed [33] using Orthosim version 2.01 [34]. Items were systematically removed based on the following criteria: (1) low communality (<0.3); (2) significant cross-loading (difference between primary and secondary loadings <0.3) [35]; and (3) primary loading on a factor inconsistent with the original theorized model. Less than ideal item Procrustes congruence values across all three participant groups was used to evaluate the cross-cultural equivalence, and the model fit comparisons were evaluated using overall congruence coefficient values.


Participant demographics

Questionnaires were mailed to a total of 1230 participants (300 in Australia, 480 in the United Kingdom, and 450 in Taiwan). Responses were received from 696 participants (56.58 %), with a greater number of respondents from Taiwan (n = 413, response rate: 91 %), compared to Australia (n = 137, response rate: 45 %) or the United Kingdom (n = 144, response rate: 30 %). It should be noted that the response rate from Taiwan was notably higher compared to that of Australian and United Kingdom sample groups. This was likely to be as a result of attitudinal differences between the participant groups. Previous research has indicated that participants from different cultural backgrounds often have different attitudes towards voluntary participation in research studies [36, 37]. Taiwan is considered to be a collectivist culture [38], which is associated with higher levels of compliance than in individualistic cultures [39] such as Australia and the United Kingdom.

The sample consisted of 519 female and 177 male participants, with the majority aged between 20 and 39 years (77 %). Most of the participants reported their highest level of occupational therapy qualification as a bachelor’s degree (78 %), were working full-time (86 %), and held the position of an occupational therapy clinician for an employer (81 %). The majority of respondents reported their primary area of responsibility as clinical (87 %), had a caseload of outpatient/community based clients (41 %), and reported being employed in a general hospital setting (42 %) (see Table 1).

Table 1 Participant demographic data (n = 696)

Initial model analysis

Since confirmatory factor analysis is often considered overly stringent for exploratory investigations of data, [40] principal components analysis with Procrustes transformation was used [33]. This method compares the rotated solution to an ideal matrix where items either load completely or not at all; providing an estimate of how well items fit. For consistency and ease of presentation, the labels of the BARRIERS Scale items have been drawn from the original scale names. Items were numbered in the order presented in the original publication of the BARRIERS Scale [10].

The 28 items of the BARRIERS Scale [10] were subjected to Principal Components Analysis with Varimax rotation for each sample. Kaiser–Meyer–Olkin Measure of Sampling Adequacy values ranged from 0.789 to 0.966, and Bartlett’s Test of Sphericity was significant, supporting the factorability of the correlation matrix for each population [4143].

Up to eight dimensions with eigenvalues exceeding 1 were found across the participant groups. However Cattell’s scree test [44] and Parallel Analysis [45] generally provided support for a four factor model, although there was slight support for a three component model in the United Kingdom sample and some support for a two component model for the Taiwanese sample. A forced four factor PCA with Varimax rotation was conducted for each sample, and a Procrustes transformation was performed using Orthosim version 2.01 [34]. The results of the initial model are presented in Table 2.

Table 2 Initial PCA loadings with Procrustes transformation by country

From Table 2, it can be seen that a considerable number of items demonstrated a primary loading on components other than those originally defined by Funk et al. [10] Item cross-loading is evident for many items and across all three participant groups. The Procrustes transformation provides an indication of item fit to an ideal matrix, with values below 0.80 suggesting less than ideal item fit [46]. Overall model fit was also calculated (Australia = 0.81; United Kingdom = 0.78; Taiwan = 0.77), with values below 0.85 indicative of less than ideal model fit [47, 48].

Final model analysis

The final model which provided the best overall fit is described in Table 3.

Table 3 Final PCA loadings with procrustes transformation by country

For each the three participant group, the Kaiser–Meyer–Olkin Measure of Sampling Adequacy values ranged from 0.797 to 0.958. Additionally, the Bartlett’s Test of Sphericity was significant, supporting the factorability of the correlation matrix [4143]. Similarly to the initial model, a four component solution was more strongly supported in the Australian sample, while there was slight support for a three component solution for the United Kingdom sample, and some support for a two component solution for the Taiwan sample. While this could be indicative of lower discriminant validity in the later samples, it is also likely to occur if the response styles are similar across items within countries. Figure 1 details the scree plots by country.

Fig. 1

Scree plot for components by country

For the Australian participants, the four factor model explained 26.36, 12.95, 9.27 and 8.43 % of the variance. For the British participants, the four factor model explained 37.79, 11.71, 7.56 and 5.57 % of the variance. For Taiwanese participants, the four factor model explained 57.67, 8.04, 5.29 and 4.44 % of the variance. From Table 3, it is evident that across all populations, the issues identified in the initial model are greatly reduced. Overall model fit was also significantly improved, with the Procrustes model fit congruence coefficients for Australia, United Kingdom and Taiwan calculated to be 0.95, 0.88, and 0.85 respectively.

An alternative application for the Procrustes transformation procedure is the comparison of component loadings between samples, rather than to an ideal matrix. This provides an indication of how well the component structure of one sample compares to another. The cross-cultural comparison demonstrated excellent results, with Procrustes model fit congruence coefficients for Australia–United Kingdom, Australia Taiwan, and United Kingdom–Taiwan calculated to be 0.93, 0.92, and 0.94 respectively. Despite the support for two or three component solutions in the Taiwan and United Kingdom samples respectively, the cross-cultural congruence coefficients provide support for the interpretability of the four component solution across cultures. Item level congruence coefficients are detailed in Table 4, while the content of the final 19 items can be found in Table 5.

Table 4 Cross-cultural procrustes congruence coefficients
Table 5 Statements for 19 item BARRIERS Scale

Validity and reliability of the final model

To investigate the validity of the scales in the final model, Pearson’s correlations were calculated for scale scores calculated using the original structure, and the adapted final structure. The resulting correlations were all highly significant, indicating that the scales in the final model are comparable to those used in the original model; and as such, interpretation of the final model scales and comparison to previously published studies will be meaningful. The correlations are reported in Table 6.

Table 6 Cronbach’s α reliability coefficients for adapted scales by country

Reliabilities for the adapted final model scales were calculated using Cronbach’s alpha. All scales indicated good or better reliability across samples with the exception of the Communication subscale, which demonstrated lower (although acceptable) reliability in the Australian and United Kingdom populations. Scale reliabilities are detailed in Table 7.

Table 7 Pearson’s correlation for original and adapted scales by country


The purpose of the study was to investigate the dimensional structure and stability of the BARRIERS Scale across three cross-cultural cohorts of occupational therapy clinicians. The findings indicate that the 19 item, four component solution provided a superior fit for structure and dimensionality compared to the original 28 item solution when considered across cultures. It also provided evidence of the component stability when compared between the three participant groups. The final model obtained in this study largely mirrored the original factor structure obtained by Funk et al. [10] which was supported by the high correlations between the original scale scores and final adapted scale scores. The primary difference between the original BARRIERS Scale version and the final model version obtained derived in this study was the number of items with the new version only being composed of 19 of the original items. The final model also demonstrated internal reliability equal to or better than that published by Funk et al. [10, 13].

The current study included a slight change to the administration of the BARRIERS Scale, with the removal of the “no opinion” response. Endorsement of not having an opinion acts as a confounder to meaningful statistical analysis. Missing value replacement protocols rely on an assumption of data missing at random or completely at random [35] which is not able to be adequately established. Previous studies which have investigated the dimensional validity of the BARRIERS Scale have chosen to either to exclude those cases (leading to a severe loss of sample size) [10] or substitute for the first response “to no extent” [12]. However to do the later suggests that having “no opinion” is equivalent to a barrier impacting on the participant “to no extent”, and that this assumption holds across all participants—an assumption that is optimistic at best, and seriously flawed at worst.

In an Australian study, Hutchinson and Johnston reported achieving an acceptable four factor solution using 27 items of the BARRIERS Scale [14]; however an investigation of the solution revealed a number of items which had primary loadings on incorrect factors, and other items with significant cross-loading. This is consistent with the initial model in the current study, and was the rationale for conducting further item removal. Similarly, in a study conducted in the United Kingdom, the factor structure of the original scale did not emerge, and the results suffered from poor validity and reliability. [27] A later study which focused on revised items fared little better. [9] These results were also consistent with the initial results found in the United Kingdom sample in the current study, where items were loading on incorrect factors and there were a considerable number of items with poor congruence coefficients.

No previous research was identified that described the use of the BARRIERS Scale in Taiwan, or where the BARRIERS Scale had been translated into Mandarin Chinese. Previous translations have been applied in countries such as Korea [19] and Iran [49]. The rigorous translation and back-translation procedure as described by Cha et al. [31] and Wang et al. [32] preserved the meaning and content of the BARRIERS Scale items, with the final model demonstrating acceptable fit. While several items demonstrated less than ideal fit in the Taiwanese participant group, these items proved to be not problematic in the overall fit. One item demonstrated a primary loading on the incorrect factor; however the difference between the primary and secondary loading was only 0.01.

To achieve acceptable model fit across the three samples, a total of nine items were removed. Three criteria were used to determine which items to remove, including low communalities, significant cross-loading, and items which demonstrated a primary loading on a component other than as originally hypothesised. Items with low communality are problematic, suggesting that the items offer little information of use while confounding the desired constructs. Significant cross-loading of items suggests a lack of discriminant validity, which can lead to inflated correlations between constructs [35]. Items with a primary loading on the incorrect component are not consistent with the theoretical underpinnings of the scale and removal of such items is worthy of consideration. While it could be argued that item removal is a threat to the depth of each construct, if the removal results in an increase in reliability, discriminant validity, and dimensional stability across samples and cultures, then the item removals are warranted.

Within the Adopter subscale, four items were removed. ADOPT7 (The clinician is isolated from knowledgeable colleagues with whom to discuss the research) and ADOPT8 (The clinician is unaware of the research) demonstrated low communality in the Australian sample, cross-loading across all samples, and a primary loading on the Organisation subscale across all samples, suggesting that these items offer little discriminant validity and may be more indicative of organisational barriers to research utilisation—albeit poorly performing ones. ADOPT6 (The clinician does not feel capable of evaluating the quality of the research) demonstrated cross-loading in all samples and primary loadings on the Organisation and Communication subscales for the United Kingdom and Taiwan samples respectively, suggesting that this item offered little discrimination. ADOPT4 (There is not a documented need to change practice) demonstrated low communality in the Australian sample, which may have been indicative of method error, given that all other items in Adoption were worded from the perspective of “the clinician” which was noticeably absent in this item.

Within the Organisation subscale, two items were removed. ORGAN7 (The clinician does not have time to read research) and ORGAN8 (The clinician feels results are not generalisable to own setting) demonstrated low communalities in the Australian sample and primary loadings on the Communication and Adopter subscales respectively. While this was not evident in the other samples, significant cross-loading was noted, suggesting that while not as problematic in the United Kingdom and Taiwan samples as the Australian sample, nevertheless the items were a threat to cross-cultural comparisons.

Within the Innovation subscale, only one item was removed. INNOV6 (Research reports/articles are not published fast enough) demonstrated low communality in the Australian sample, and primary loading on the Communication subscale across all samples. Semantically, the relationship between the timely availability of research and accessibility of research appears to be meaningful, however given that the item was not hypothesised to load on this component, removal was justified.

Within the Communication subscale, two items were removed. COMM1 (Implications for practice are not made clear) demonstrated significant cross-loading in the Australian sample. COMM6 (The research is not relevant to the clinician’s practice) performed rather inconsistently across samples, loading on Innovation in the Australian sample, demonstrating low communality in the United Kingdom sample, and cross-loading in the Taiwan sample.

Cross-cultural comparisons using Procrustes transformation allowed for further evaluation of the structural integrity and dimensionality of the 19 item BARRIERS Scale. While overall fit was excellent, two items indicated less than ideal fit across populations. Between Australia and Taiwan, the fit for COMM2 (Research reports/articles are not readily available) was less than ideal, and a review of the component loadings suggested that this was mostly as a result of larger secondary loadings in the Taiwanese sample, with the primary loadings differing between participant groups by only 0.02. Between Australia and the United Kingdom, COMM5 (The relevant literature is not compiled in one place) also indicated misfit, with a lower primary loading and cross-loading in the United Kingdom sample. This pattern may indicate that the availability of relevant literature is less of a concern in the United Kingdom than in Australia.

The Communication subscale demonstrated lower reliability in both the original study [10] and in the final model of the current study. This indicates that the Communication subscale has been consistently underperforming when compared to the other BARRIERS subscales, and research may be warranted in further refining this subscale through the development of additional items. Interestingly, the Communication subscale demonstrated much higher reliability in the Taiwanese sample, which suggests that the content of the items in the final model of the subscale may be more indicative of barriers to research relating to presentation and accessibility in Taiwan than in Australia or the United Kingdom.

A brief comment is warranted on the variance explained by the 19 item four component model across samples. The total variance explained for Australia, United Kingdom, and Taiwan was 57.01, 62.63, and 75.44 % respectively. Some explanation can be found in the scree plots (see Fig. 1). While a four component solution was clearly demonstrated for the Australian sample, there was a slight inclination towards a three component solution for the United Kingdom sample, and some evidence for a two component solution for the Taiwan sample. This suggests that there may be somewhat less discriminant validity between the components in the United Kingdom and Taiwan samples, or an element of “blurred boundaries” between the components [50]. As more components are extracted, progressively more variance is explained which contributes to the higher total variance reported in these later samples. When the forced four component model was compared cross-culturally, the slightly reduced discriminant validity was not problematic given the excellent model fit, supporting the use of the four component model in the cross-cultural comparison. Of the 13 studies identified by Kajermo et al. [17] which investigated the structure of the BARRIERS scale, eight reported values for total variance explained. Interestingly, the mean variance explained across studies was only 44.03 % (range 38.90–47.50 %), which provides further support for the item removal process in the current study given the increase in variance explained by the 19 item model.

Study limitations

The study has several potential limitations. Firstly, the use of self-report data has a number of bias issues such as answering scale items in a socially desirability manner. Secondly, the use of convenience sampling, while easier to recruit participants, can limit the generalisability of results, although three different sample groups strengthen the external validity. The current study also used relatively small sample sizes for Australia and the United Kingdom, although these proved to be not problematic. Response rates varied across samples, however this may be somewhat explained by cultural differences in compliance [37].

Future research

Several suggestions for future research are apparent. The consistent lower reliability of the Communication subscale suggests that further research is warranted to more accurately identify and measure these barriers to research. The final model presented in the current study requires fewer items while preserving the meaning of the original BARRIERS Scale factor structure, and demonstrated cross-cultural structural and dimensional stability. The rigorous psychometric evaluation presented in the current study allows for meaningful comparisons to be drawn between paediatric occupational therapists in Australia, the United Kingdom, and Taiwan. While not explicitly tested in the current study, it would be worth exploring the extent to which the same dimensional stability is evident across other cultures and professions, to assist in informing evidence based practice. It is also suggested that the 19 item, four component version of the BARRIERS Scale be used by clinicians who might be interested in identifying key self-reported barriers to research utilization and evidence based practice in health care settings.


The 19 item version of the BARRIERS Scale is a robust, valid and reliable measure. It appears to have applicability in cross-cultural settings and also has decreased respondent burden since it has nine fewer items compared to the original 28 item version. The 19 item version of the BARRIERS Scale demonstrated acceptable dimensional stability when completed by participants from three different cross-cultural contexts. Further evaluation of the measurement properties of the 19 item BARRIERS Scale across health professions is recommended.


  1. 1.

    Turpin M, Higgs J. Clinical reasoning and EBP. In: Hoffmann T, Bennett S, Del Mar C. Chatswood, editors. EBP across the health professions. Australia: Elsevier; 2010. p. 300–317.

  2. 2.

    Fink R, Thompson CJ, Bonnes D. Overcoming barriers and promoting the use of research in practice. J Nurs Adm. 2005;35(3):121–9.

    Article  PubMed  Google Scholar 

  3. 3.

    Metcalf C, Lewin R, Wisher S, Perry S, Banningan K, Moffett JK. Barriers to implementing the evidence base in four NHS therapies: dietitians, occupational therapists, physiotherapists, speech and language therapists. Physiotherapy. 2001;87(8):433–41.

    Article  Google Scholar 

  4. 4.

    Thompson DS, Chau JPC, Lopez V. Barriers to, and facilitators of, research utilization: a survey of Hong Kong regsitered nurses. Int J Evid Based Healthc. 2006;4(2):77–82.

    PubMed  Google Scholar 

  5. 5.

    Cameron K, Ballantyne S, Kulbitsky A, Morgolis-Gal M, Daugherty T, Ludwig F. Utilization of EBP by registered occupational therapists. Occup Therapy Int. 2005;12:123–36.

    Article  Google Scholar 

  6. 6.

    Gustaffson L, Yates K. Are we applying interventions with research evidence when targeting secondary complications of the stroke-affected upper limb. Aust Occup Ther J. 2009;56:428–36.

    Article  Google Scholar 

  7. 7.

    Parahoo K. Barriers to, and facilitators of, research utilization among nurses in Northern Ireland. J Adv Nurs. 2000;31(1):89–98.

    CAS  Article  PubMed  Google Scholar 

  8. 8.

    Boström AM, Kajermo KN, Nordström G, Wallin L. Barriers to research utilization and research use amongst registered nurses working in the care of older people: does the BARRIERS Scale discriminate between research users and non-research users on perceptions of barriers? Implement Sci. 2008;3:24.

    PubMed Central  Article  PubMed  Google Scholar 

  9. 9.

    Marsh GW, Nolan M, Hopkins S. Testing the revised barriers to research utilization scale for use in the UK. Clin Eff Nurs. 2001;5(2):66–72.

    Article  Google Scholar 

  10. 10.

    Funk SG, Champagne MT, Wiese RA, Tornquist EM. BARRIERS: the barriers to research utilization scale. Appl Nurs Res. 1991;4(1):39–45.

    CAS  Article  PubMed  Google Scholar 

  11. 11.

    Retsas A. Barriers to using research evidence in nursing practice. J Adv Nurs. 2000;31:599–606.

    CAS  Article  PubMed  Google Scholar 

  12. 12.

    Closs SJ, Bryar RM. The BARRIERS Scale: does it ‘fit’ the current NHS research culture? Nurs Times Res. 2001;6(5):853–65.

    Article  Google Scholar 

  13. 13.

    Funk SG, Champagne MT, Wiese RA, Tornquist EM. Barriers to using research findings in practice: the clinician’s perspective. Appl Nurs Res. 1991;4(2):90–5.

    CAS  Article  PubMed  Google Scholar 

  14. 14.

    Hutchinson AM, Johnston L. Beyond the BARRIERS Scale: commonly reported barriers to research use. J Nurs Adm. 2006;36(4):189–99.

    Article  PubMed  Google Scholar 

  15. 15.

    Rogers E. Diffusion of innovations. 4th ed. New York: Free Press; 1995.

    Google Scholar 

  16. 16.

    McCleary L, Brown GT. Barriers to paediatric nurses’ research utilization. J Adv Nurs. 2003;42(4):364–72.

    Article  PubMed  Google Scholar 

  17. 17.

    Kajermo KN, Boström A-M, Thompson DS, Hutchinson AM, Estabrooks CA, Wallin L. The BARRIERS Scale—the barriers to research utilization scale: a systematic review. Implement Sci. 2010;32(5):1–22.

    Google Scholar 

  18. 18.

    LaPierre E, Ritchey K, Newhouse R. Barriers to research use in the PACU. J PeriAnesthesia Nurs. 2004;19(2):78–83.

    Article  Google Scholar 

  19. 19.

    Oh EG. Research activities and perceptions of barriers to research utilization among critical care nurses in Korea. Intensive Crit Care Nurs. 2008;24(5):314–22.

    Article  PubMed  Google Scholar 

  20. 20.

    Strickland R, O’Leary-Kelley C. Clinical nurse educators’ perceptions of research utilization: barriers and facilitators to change. J Nurses Staff Dev. 2009;25(4):164–71.

    Article  PubMed  Google Scholar 

  21. 21.

    Samuelsson K, Wressle E. Turning evidence into practice: Barriers to research use among occupational therapists. Br J Occup Therapy. 2015;78(3):175–81.

    Article  Google Scholar 

  22. 22.

    Closs S, Lewin B. Perceived barriers to research utilisation: a survey of four therapies. Br J Therapy Rehabil. 1998;5(3):151–5.

    Article  Google Scholar 

  23. 23.

    Diamantopoulos A, Sarstedt M, Fuchs C, Wilczynski P, Kaiser S. Guidelines for choosing between multi-item and single-item scales for construct measurement: a predictive validity perspective. Acad Mark Sci J. 2012;40(3):434–49.

    Article  Google Scholar 

  24. 24.

    Hutchinson AM, Johnston L. Clinical nursing issues bridging the divide: a survey of nurses’ opinions regarding barriers to, and facilitators of, research utilization in practice setting. J Clin Nurs. 2004;13:304–15.

    Article  PubMed  Google Scholar 

  25. 25.

    Bayik TA, Uysala A, Ardahan M, Ozkahraman S. Barriers to research utilization scale: psychometric properties of the Turkish version. J Adv Nurs. 2010;66(2):456–64.

    Article  Google Scholar 

  26. 26.

    Gunderman RB, Meesa IR. The adoption of innovation. Radiology. 2008;264(3):659–61.

    Article  Google Scholar 

  27. 27.

    Dunn V, Crichton N, Roe B, Seers K, Williams K. Using research for practice: A UK experience of the BARRIERS Scale. J Adv Nurs. 1998;26(6):1203–10.

    Article  Google Scholar 

  28. 28.

    Retsas A, Nolan M. Barriers to nurses’ research: an Australian hospital study. Int J Nurs Stud. 1999;36:335–43.

    CAS  Article  PubMed  Google Scholar 

  29. 29.

    Ashley JS. Barriers and facilitators to research utilization as perceived by critical care nurses. Doctoral Dissertation. 2005.

  30. 30.

    Carlson CL, Plonczynski DJ. Has the BARRIERS Scale changed nursing practice? J Adv Nurs. 2008;63(4):322–33.

    Article  PubMed  Google Scholar 

  31. 31.

    Cha E-S, Kim KH, Erlen JA. Translation of scales in cross-cultural research: Issues and techniques. J Adv Nurs. 2007;58(4):386–95.

    Article  PubMed  Google Scholar 

  32. 32.

    Wang W, Lee H, Fetzer SJ. Challenges and strategies of instrument translation. West J Nurs Res. 2006;28:310–21.

    CAS  Article  PubMed  Google Scholar 

  33. 33.

    McCrae R, Zonderman A, Costa P, Bond M, Paunonen S. Evaluating replicability of factors in the revised NEO personality inventory: confirmatory factor analysis versus procrustes rotation. J Pers Soc Psychol. 1996;70(3):552–66.

    Article  Google Scholar 

  34. 34.

    Barrett P. Orthosim 2.01. [computer software]. Auckland: New Zealand; 2006.

  35. 35.

    Tabachnick BG, Fidell LS. Using multivariate statistics. Boston: Pearson Education Limited; 2013.

    Google Scholar 

  36. 36.

    Glickman SW, Anstrom KJ, Lin L, Chandra A, Laskowitz DT, et al. Challenges in enrollment of minority, pediatric, and geriatric patients in emergency and acute care clinical research. Ann Emerg Med. 2008;51:775–80.

    Article  PubMed  Google Scholar 

  37. 37.

    Shavers VL, Lynch CF, Burmeister LF. Racial differences in factors that influence the willingness to participate in medical research studies. Ann Epidemiol. 2002;12:248–56.

    Article  PubMed  Google Scholar 

  38. 38.

    Ali AJ, Lee M, Hsieh YC, Krishnan K. Individualism and collectivism in Taiwan. Cross Cult Manag Int J. 2005;12(4):3–16.

    Article  Google Scholar 

  39. 39.

    Se Hyung Oh. Do collectivists conform more than individualists? Cross-cultural differences in compliance and internalization. Soc Behav Personal Int J. 2013;41(6):981–94.

    Article  Google Scholar 

  40. 40.

    Browne MW. An overview of analytic rotation in exploratory factor analysis. Multivar Behav Res. 2001;36(1):111–50.

    Article  Google Scholar 

  41. 41.

    Bartlett M. A note on the multiplying factors for various Chi square approximations. J R Stat Soc 1954;16(Series B):296–298.

  42. 42.

    Kaiser H. A second generation little iffy. Psychometrika. 1970;35:401–15.

    Article  Google Scholar 

  43. 43.

    Kaiser H. An index of factorial simplicity. Psychometrika. 1974;39:31–6.

    Article  Google Scholar 

  44. 44.

    Cattell RB. The scree test for the number of factors. Multivar Behav Res. 1966;1(2):245–76.

    Article  Google Scholar 

  45. 45.

    Watkins MW. Monte Carlo PCA for parallel analysis [computer software]. State College, PA: Ed & Psych Associates; 2000.

  46. 46.

    Barrett P. Factor comparison: an examination of three methods. Personal Individ Differ. 1986;7(3):327–40.

    Article  Google Scholar 

  47. 47.

    Mulaik S. The foundations of factor analysis. New York: McGraw-Hill; 1972.

    Google Scholar 

  48. 48.

    ten Berge JMF. Rotation to perfect congruence and the cross validation of component weights across populations. Multivar Behav Res. 1986;21(1):41–64.

    Article  Google Scholar 

  49. 49.

    Mehrdad N, Salsali M, Kazemnejad A. The spectrum of barriers to and facilitators of research utilization in Iranian nursing. J Clin Nurs. 2008;17(16):2194–202.

    Article  PubMed  Google Scholar 

  50. 50.

    DeVellis RF. Scale development: theory and applications. 3rd ed. Thousand Oaks: Sage Publications, Inc; 2012.

    Google Scholar 

Download references

Authors’ contributions

TB was involved in the study design, implementation, data collection, data analysis and drafting of the manuscript. BW and SC were involved in the data analysis and drafting of the manuscript. All authors read and approved the final manuscript.


We would like to thank the Australian, United Kingdom and Taiwanese occupational therapists for taking the time to complete the questionnaire.

Competing interests

The authors declare that they have no competing interests. The authors alone are responsible for the content and writing of this paper.

Author information



Corresponding author

Correspondence to Brett Williams.

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (, which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver ( applies to the data made available in this article, unless otherwise stated.

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Cite this article

Williams, B., Brown, T. & Costello, S. A cross-cultural investigation into the dimensional structure and stability of the Barriers to Research and Utilization Scale (BARRIERS Scale). BMC Res Notes 8, 601 (2015).

Download citation


  • BARRIERS Scale
  • Validity
  • Psychometric properties
  • Research utilization
  • Factor structure
  • Dimensional stability
  • EBP