Introduction
Improving healthcare systems is essential globally as they continually evolve to meet patient expectations, enhance population health, and improve the quality of care [1]. Effective communication and appropriate feedback mechanisms are vital in achieving these goals, especially among healthcare professionals who must work cohesively. Nurses, in particular, are pivotal in providing quality care and ensuring patient safety; their insights and feedback are crucial for the success of healthcare systems [2, 3].
Numerous literature reviews have shown that healthcare team members often refrain from addressing issues concerning patient safety and unprofessional behavior until they escalate into serious problems [2, 4–6]. This reluctance to speak up can harm the organization [7] and diminish commitment to core values [8]. Ethical concerns are a primary reason for nurses’ reluctance to voice their opinions [9], and feelings of being unwelcome or excluded can further contribute to this silence [7]. Other factors include perceiving one’s role as passive [10, 11], a lack of organizational identification [12], the integrity of leaders, and concerns about negatively impacting one’s academic career [13].
The importance of psychometric research in healthcare settings cannot be overstated. Several theories have been proposed to elucidate organizational silence, with Organizational Culture and Communication Theory being the most prominent. According to Organizational Culture Theory, an organization’s culture significantly impacts employees’ inclination to voice their opinions. A culture that promotes open dialogue creates an environment where employees feel comfortable expressing their concerns without fear of reprisal. Conversely, a culture discouraging dissenting views leads to heightened silence among team members [14]. Communication Theory underscores the significance of effective communication channels in addressing organizational silence. Studies have demonstrated that effective communication among healthcare professionals reduces errors and improves patient outcomes [15]. Organizations that cultivate open communication channels experience fewer adverse events, underscoring the necessity of a supportive communication framework [16]. Additionally, Social Exchange Theory suggests that employees weigh the costs and benefits of voicing their concerns. If they perceive the risks (e.g., retaliation, lack of support) to outweigh the benefits, they are more likely to remain silent [17].
Previous studies have explored the impact of communication and feedback systems on patient safety and organizational effectiveness. Research indicates that healthcare professionals often feel inhibited from voicing concerns due to fear of retaliation or lack of support from management [10, 18]. This aligns with findings suggesting that organizational culture is pivotal in encouraging or discouraging open dialogue among team members [7].
Various questionnaires have been developed to assess this concept in organizational silence. While tools like the Organizational Silence Scale by Van Dyne et al. [19] and the Organizational Silence Inventory by Knoll and Van Dick [20] offer valuable insights, they lack specificity regarding the unique challenges faced by healthcare professionals in high-stakes environments. One significant limitation of these instruments is their general applicability, which often fails to account for the specific dynamics and pressures within healthcare settings. For example, healthcare professionals frequently encounter situations where speaking up about patient safety issues is critical [21]. However, research indicated that many team members refrain from voicing their concerns until these issues escalate to critical levels. This delay in communication can severely compromise patient safety and care quality, leading to adverse outcomes [22].
Moreover, the existing tools do not adequately address the psychological barriers that healthcare professionals face. Factors such as fear of retaliation—where individuals worry about negative repercussions from colleagues or management—play a crucial role in fostering silence [23, 24]. Additionally, inadequate support from management can exacerbate feelings of isolation and disempowerment among staff, making them less likely to speak up. The existing questionnaires often overlook these emotional and contextual factors, resulting in a lack of actionable insights for improving team communication [21, 24].
In Turkey, a scale was developed by Yalçın and Baykal to measure organizational silence behavior among nurses to minimize the gap in the literature review. The reassessment of the organizational silence behavior scale (OSBS) for healthcare professionals is crucial due to the evolving nature of healthcare systems and the increasing recognition of the impact of organizational silence on patient outcomes. Unlike existing scales, the OSBS is specifically tailored to reflect the dynamics of healthcare settings, incorporating factors such as hierarchical structures and interdisciplinary communication challenges prevalent in clinical environments [21].
The importance of developing a culturally appropriate tool for measuring organizational silence in the Iranian context cannot be overstated. A tailored instrument, such as the OSBS-P, is necessary to accurately reflect the organizational dynamics and communication challenges faced by healthcare professionals in Iran. This study aims to fill the research gap by providing a psychometrically validated scale that captures the nuances of organizational silence behaviors specific to Iranian clinical settings.
This research contributes to the existing body of knowledge by offering new insights into the factors influencing organizational silence among clinical nurses in Iran. It emphasizes the need for a new tool with valid and reliable psychometric properties, addressing the limitations of existing scales.
Statement of the problem
It is crucial to investigate the phenomenon of organizational silence in Iranian healthcare environments [25]. Organizational silence, where employees are hesitant to address critical issues, is a significant concern in healthcare settings due to its potential adverse impact on patient safety and care quality [10]. By delving into the causes and manifestations of organizational silence, the research aimed to gain insights that can guide targeted interventions to mitigate its effects and promote a culture of open communication and transparency within healthcare organizations. Ultimately, the findings of this research have the potential to improve healthcare outcomes by illuminating the dynamics of organizational silence and facilitating the implementation of interventions to foster open communication and employee voice, thereby enhancing patient safety and care quality in Iranian healthcare settings. The primary goal of this study was to translate and evaluate the psychometric properties of the Persian version of the Organizational Silence Behavior Scale (OSBS-P) among clinical nurses.
Materials and methods
Design and setting
From October 2023 to February 2024, we conducted a methodological study to assess the reliability and validity of the Persian version of the Organizational Silence Behavior Scale (OSBS-P) among clinical nurses. This validation process was a significant aspect of our research. We used convenience sampling to gather participants efficiently within our limited research timeframe. This approach allowed us to recruit participants from five educational-therapeutic hospitals in Ardabil province, Northwestern Iran.
Given our time and resource constraints, convenience sampling proved especially advantageous and enabled us to collect data from readily available and willing clinical nurses. To further clarify our methodological design, we referenced previous studies that employed convenience sampling in psychological research, underlining its effectiveness in similar contexts. For example, studies by Yalçın and Baykal [21] and Chung and Hwang [23] successfully utilized convenience sampling to investigate related phenomena, affirming the reliability of findings obtained through this method. Convenience sampling can yield valuable insights, especially in exploratory research or when access to a larger population is limited. However, it may also introduce selection bias and restrict the generalizability of the findings. Therefore, while we reference these studies to advocate for convenience sampling in our research context, we must consider its limitations and interpret the results cautiously. This balanced perspective fosters a deeper understanding of the implications of employing convenience sampling in our study [26].
Participants
A survey was conducted on 338 clinical nurses who work in educational-therapeutic hospitals of Ardabil University of Medical Sciences in Iran. To be eligible for the study, participants had to voluntarily agree and possess at least a bachelor’s degree in nursing or a higher qualification. Additionally, they needed to have a minimum of six months of work experience in the field. The exclusion criterion for disqualification from the research was unanswered questions in the questionnaire.
Instrument
The tool comprised two sections. The initial section consisted of a questionnaire to collect demographic information, including age, gender, work experience, and other relevant details. The second part of the questionnaire was the Organizational Silence Behavior Scale (OSBS-P), designed by Yalçın and Baykal [21] and administered to clinical nurses. The OSBS-P is a scale that comprises 32 items, which are further divided into four subscales. These subscales include silence climate (5 items), silence based on fear (12 items), acquiescence silence (10 items), and silence based on protecting the organization (5 items). The items in the scale are rated on a 5-point Likert scale ranging from 1 (never stay silent) to 5 (always stay silent). As the total score on the scale increases, the level of silence also increases (S1 File).
The Content Validity Index (CVI) for the original scale was measured at 0.86, indicating an adequate level of content validity. Before performing Confirmatory Factor Analysis (CFA), Exploratory Factor Analysis (EFA) was conducted to determine the suitability of the dataset for CFA. The KMO index was 0.945, and the results of Bartlett’s test showed χ2 = 14903.485; p < 0.001 (n = 485), indicating that the data were appropriate for factor analysis. The CFA results revealed a CMIN/DF of 3.54, RMSEA of 0.07, GFI of 0.92, CFI of 0.92, and NFI of 0.90, demonstrating a good fit for the model. Additionally, the Cronbach’s alpha for the overall scale was 0.93. Specifically, the reliability coefficients for the subscales were as follows: ’Silence based on fear’ (0.93), ’Silence based on protecting the organization’ (0.85), ’Silence climate’ (0.93), and ’Acquiescent silence’ (0.91).
Psychometric evaluation
Translation procedure.
The researchers sought permission from Dr. Begüm Yalçın, the developer of OSBS, to use the tool with Iranian clinical nurses. They followed the forward-backward method recommended by the WHO guidelines to translate the tool from English to Persian and back to ensure accuracy [27]. Two professional translators worked independently to translate the tool from English to Persian, which was evaluated and discussed by an expert panel comprising two nursing professors and two translators. The Persian version was finalized and subsequently back-translated into English by a translator without previous knowledge of the original version or the study’s objectives. The expert panel reviewed and refined this back-translated version against the original version (S2 File) to create the final Persian tool.
Face validity.
A scale’s face validity was assessed using quantitative and qualitative approaches. Ten clinical nurses from different units were interviewed to evaluate the scale’s face validity. The nurses were asked to provide feedback on the items’ difficulty level, relevance to the measured concept, and presence of ambiguous language. The researchers recorded and reviewed the participants’ comments. Based on the feedback, any required modifications to the scale were made.
A structured approach was employed to assess the quantitative face validity of the scale. Clinical nurses were asked to evaluate the significance of each item on a five-point scale ranging from "unimportant" (1) to "important" (5). The impact score for each item was calculated by multiplying the percentage of respondents who rated the item as important (scores of 4 or 5) by the average importance rating assigned to that item. An item was considered significant if it had an impact score greater than 1.5. This method ensured that only items deemed relevant and meaningful by the target population of clinical nurses were included in the final scale [28].
Content validity.
The content validity of the OSBS-P was evaluated using qualitative and quantitative methods. For the qualitative method, we conducted purposive sampling and administered the OSBS-P to ten experts, specifically nursing faculty members with diverse backgrounds. These faculty members were selected based on their expertise in various areas of nursing, including clinical practice, nursing education, and research. Their collective experience ensured a comprehensive evaluation of the scale. Following the qualitative examination, the experts provided feedback on several aspects of the OSBS-P, including grammar, terminology, and overall relevance to clinical nursing practice. This feedback was instrumental in refining the scale to meet the target population’s needs better.
To evaluate the precision of our scale, we conducted an assessment of its quantitative content validity. This was done using the content validity ratio (CVR) and the content validity index (CVI). A 3-point Likert scale was used to rate each item to determine the CVR. Experts rated each item as "essential," "useful but not essential," or "not essential" for the scale. The CVR formula was utilized to calculate the CVR. The formula considers the total number of invited experts (N) and the number of experts who ranked the item (Ne). The resulting ratio is CVR = (Ne—N/2) / (N/2). According to Lawshe’s table, a CVR value of 0.62 or higher is acceptable. Also, the items were rated by experts on a 4-point Likert scale to evaluate their relevance, clarity, and simplicity. The scale ranged from 1 for "not relevant" to 4 for "completely relevant." Items with CVIs greater than 0.79 were deemed acceptable [29].
To ensure an accurate assessment, an examination was conducted to detect any floor and ceiling effects that could suggest content validity issues. A floor effect occurs when over 15% of participants attain the lowest possible score on a particular tool, indicating that the tool lacks items representing the minimum intensity of the phenomenon. Conversely, a ceiling effect occurs when more than 15% of participants achieve the highest possible score on the tool, suggesting a lack of items representing the maximum intensity of the phenomenon. If these effects exceed 15%, the tool may be considered to have inadequate content validity [30].
Construct validity.
The scale OSBS-P was created by utilizing exploratory factor analysis (EFA). In contrast, this research employed confirmatory factor analysis (CFA) to assess the construct validity [31]. Several methods are available to determine the construct validity of a measurement instrument, such as Confirmatory Factor Analysis (CFA) and Exploratory Factor Analysis (EFA). However, research suggests that CFA is generally considered more effective than EFA. CFA allows researchers to test a pre-specified theoretical model of the underlying constructs. At the same time, EFA involves an exploratory approach that may not necessarily align with the theoretical model. Therefore, CFA is often preferred when the goal is to confirm a specific hypothesis or construct [32]. The current study estimated the parameters using Maximum Likelihood Estimation (MLE). The model fit indices were evaluated based on the following criteria: χ2/df ratio less than 3, RMSEA less than 0.08, GFI greater than 0.90, CFI greater than 0.90, TLI greater than 0.90, NFI greater than 0.90, and AGFI greater than 0.80 [33]. Statistically significant factor loadings were those above 0.3 and T-values above 1.96, as per Hair et al. [34].
There are varying opinions on what constitutes an adequate sample size for a study. Some experts recommend a minimum of 100 participants, while others suggest a larger sample size of 400 or more. The number of participants needed may also depend on other factors, such as the complexity and number of items in the measuring tool [34–40]. In our study, we employed the OSBS-P, which comprised 32 items. Adhering to the general guideline of having 5–10 participants for each item [41], we recruited 12 participants per item, culminating in 384 clinical nurses through convenience sampling. Notably, 88% of these participants completed the survey. After carefully analyzing the valid responses, we selected 338 participants for the CFA. This sample size was adequate to ensure a robust structural validity analysis while accurately representing the target population and minimizing the risk of bias or error.
Convergent and discriminant validity.
Hair et al. defined convergent validity as the degree to which a construct is characterized by its items’ similarity and high common variance. In contrast, divergent validity is demonstrated when a construct’s items are dissimilar and unrelated [34]. To evaluate the accuracy and consistency of the OSBS-P scale, we used the Fornell and Larcker criterion, along with key parameters such as Composite Reliability (CR), Average Variance Extracted (AVE), Maximum Shared Squared Variance (MSV), and Average Shared Squared Variance (ASV) [42, 43]. To establish convergent validity, we looked for AVE values greater than 0.5 and CR values greater than 0.7, with CR being greater than AVE [44]. It should be noted that if the AVE exceeds both the MSV and the ASV, it is a sign that the measurements employed in the analysis possess reliable discriminant validity [34]. To confirm discriminant validity, the researchers utilized the Heterotrait-Monotrait (HTMT) correlation ratio criterion developed by Henseler et al. [45]. As per Henseler et al.’s recommendation, all values in the HTMT matrix must be less than 0.9 [45].
Reliability.
We assessed the internal consistency of the instrument and its dimensions using Cronbach’s alpha coefficient (α), McDonald’s omega (ω), Coefficient H, and mean inter-item correlation (ρ) to confirm that the items within each scale measure the same construct reliably. For the results to be acceptable, the values of α, ω, and H should be greater than 0.7 [46, 47]. Furthermore, Coefficient H demonstrated the highest level of reliability among the subscales [46–48]. It’s worth noting that internal consistency can be measured regardless of the number of items or sample size since mean inter-item correlation remains independent [43, 49, 50]. An inter-item correlation ranging from 0.15 to 0.5 is commonly considered optimal. If the correlation falls below 0.15, it may suggest a weak relationship between the items. Conversely, the correlation above 0.5 may indicate redundancy among the items [51].
The Intraclass Correlation Coefficient (ICC) was used to determine the stability of the OSBS-P. The data were collected from 40 clinical nurses by simple random sampling over two weeks to assess test-retest reliability. The acceptable threshold was an ICC value of 0.75 or above [52].
Data analysis
Two methods were used to detect outliers in the data. The first method involved examining the skewness and kurtosis values for univariate outliers. Values larger than ±3 and ±8, respectively, were considered outliers. The second approach employed the Mahalanobis squared distance technique to identify multivariate outliers. The approach sets a statistical significance threshold at p < 0.001, indicating that only outliers with a very low probability of occurring by chance will be identified. The suitability of a normal distribution for multivariate variables was assessed using Mardia’s coefficient. If the computed Mardia coefficient was above 8, it was determined that the data did not adhere to a normal distribution [53].
The study employed IBM SPSS Statistics for Windows, version 24.0 (IBM Corp., Armonk, NY, USA) for performing descriptive statistical analysis (such as mean, standard deviation (SD), number, and percentage). The CFA and structured equation modeling (SEM) were also conducted using IBM SPSS AMOS, version 24.0. The predetermined statistical significance level was established at p < 0.05.
Ethical considerations
The Research Ethics Committee of Ardabil University of Medical Sciences, Iran, has approved the research proposal with the ethics code (IR.ARUMS.REC.1402.062). The research project followed the principles outlined in the Declaration of Helsinki. Before the commencement of the study, written informed consent was acquired from all participating individuals. Detailed information regarding the study’s objectives, methods, nature, and duration was provided to each participant thoroughly. This included a comprehensive explanation of the potential benefits, risks, and any potential discomforts that may be experienced. Participants were also informed of their right to withdraw from the study at any time.
Results
Participant characteristics
The demographic characteristics of the participants in our study, which included 338 individuals, revealed a mean age of 35.75 years (SD = 4.33) and an average working experience of 9.52 years (SD = 4.25). The sample comprised 39.1% male (n = 132) and 60.9% female (n = 206) participants. Regarding marital status, 41.7% of the participants were single (n = 141), while 58.3% were married (n = 197). Education-wise, the majority held a bachelor’s degree (71.6%, n = 242), with 28.4% possessing a master’s degree (n = 96). Participants were also divided based on their working departments, with 36.4% in medical (n = 123), 20.1% in surgical (n = 68), 9.5% in emergency (n = 32), 9.5% in ICU (n = 32), 6.8% in pediatric (n = 23), and 17.7% in other departments (n = 60).
Face validity
During the face validity process, we received feedback from clinical nurses regarding the translated version of the OSBS-P scale. The nurses highlighted several key areas for improvement: some items were a little complex, prompting suggestions to simplify the language for better clarity. Instances of ambiguous language were identified, and the nurses recommended rephrasing these items to ensure consistent interpretation among respondents. The researchers recorded and reviewed the participants’ comments. Based on this feedback, any required modifications to the scale were made, enhancing its clarity and relevance to clinical practice. As a result, all items in the translated version received a score of 1.5 or higher, ranging from 2.5 to 3.8 (Table 1). Accordingly, we discovered that all the items present in the original OSBS-P scale were deemed necessary by the participants, and none of them were removed during the face validity process. Keeping these findings in mind, we retained all the items to proceed with further stages of our study.
[Figure omitted. See PDF.]
Content validity
According to the results, each item had a CVR value above 0.62, varying from 0.7 to 1. The Waltz and Bausell method yielded a CVI of over 0.79 for each item (ranging from 0.8 to 1) (Table 1). During the qualitative phase of the study, the experts found that the 5-point Likert scale ranging from 1 (never stay silent) to 5 (always stay silent) had an appropriate ranking scale that effectively measured the intended constructs. Additionally, they concluded that the grammar, wording, item allocation, and scaling of the response category were all suitable for the targeted population and research objectives. This suggests that the response category used in the survey was well-designed and reliable for collecting valuable data.
Moreover, no floor or ceiling effects were observed when most participants scored at the lowest or highest possible level, respectively. Accordingly, the percentage of participants who scored at the lowest or highest possible level was less than 15% for all measures. This indicates that the OSBS-P is a reliable and valid tool for measuring the constructs it was designed to assess (Table 2).
[Figure omitted. See PDF.]
Descriptive statistics of the 32-item OSBS-P
The scale’s overall mean score was 3.11 (0.47). The mean scores for the subscales of silence climate, silence based on fear, acquiescence silence, and silence based on protecting the organization were 3.61 (0.89), 3.02 (0.68), 3.12 (0.67), and 3.48 (0.85), respectively. According to the findings, the study’s results showed that the overall scale and subscales had negative skewness. The negative skewness means that most participants had favorable attitudes towards the items in the scale, which could suggest a high degree of satisfaction or agreement with the statements presented. This information provides valuable insight into the perceptions and attitudes of the participants towards the topic under investigation (Table 2).
Construct validity
Fig 1 shows the OSBS-P structure, where the latent factors 1 to 4 are silence climate, silence based on fear, acquiescence silence, and silence based on protecting the organization, respectively. The CFA findings conducted using a four-factor model revealed that all the items in the model had a factor loading of 0.3 or higher, ranging from 0.54 to 0.81. This suggests that none of the items needed to be eliminated from the model, and the results were statistically significant with a p < 0.001. The statistical analysis using T-value tests revealed a significant relationship between all latent factors and their corresponding items. It is worth mentioning that the T-value for all items was greater than 1.96, indicating a strong association between the factors and their observed variables. The proposed model of OSBS-P and its concepts are deemed acceptable overall based on the goodness-of-fit indices: χ2 = 1254.65, df = 456, p < 0.001, χ2/df = 2.751, RMSEA = 0.052, GFI = 0.913, CFI = 0.938, TLI = 0.933, NFI = 0.939, and AGFI = 0.877 (Table 3). It was determined that four latent factors approved this structure.
[Figure omitted. See PDF.]
[Figure omitted. See PDF.]
Convergent and discriminant validity
The results presented in Table 4 indicate that the latent factors under study exhibit strong convergent validity. This is evidenced by their high CR and AVE values, surpassing the recommended thresholds of 0.7 and 0.5, respectively, with CR being greater than AVE. These findings suggest that the measurement model used to assess the latent constructs is reliable and valid, providing robust support for the study’s theoretical framework. The discriminant validity of the latent factors is confirmed by the MSV and ASV values being lower than AVE. The HTMT matrix analysis also reveals that all values are less than 0.9, as presented in Table 4.
[Figure omitted. See PDF.]
Reliability
The internal consistency of the scale was assessed using various statistical measures such as alpha coefficient (α = 0.969), MacDonald’s omega coefficient (ω = 0.971), coefficient H (H = 0.977), and mean inter-item correlation (ρ = 0.494). These measures indicate that the OSBS-P scale is a reliable and consistent tool for assessing the construct it aims to measure. Also, MacDonald’s omega coefficient and H coefficient values for four factors (subscales) were above 0.7. On the other hand, mean inter-item correlation values for all factors were between 0.15 and 0.5, which indicates excellent and acceptable internal consistency (Table 5).
[Figure omitted. See PDF.]
The ICC was utilized to assess the OSBS-P’s stability across all dimensions. The overall ICC was 0.951. Notably, the ICC values for all dimensions were above 0.75, demonstrating a high level of agreement among the raters. These findings reinforce the OSBS-P’s reliability, instilling confidence in its use in future research and clinical settings (Table 5).
Discussion
This study aimed to translate the Organizational Silence Behaviors Scale (OSBS) into Persian and evaluate its psychometric properties among clinical nurses working in educational-therapeutic hospitals of Ardabil, Iran. The study involved the participation of 338 clinical nurses selected through convenience sampling. The translated version of the OSBS was administered to the participants, and the data was analyzed using various statistical methods to assess the reliability and validity of the scale. This scale can be utilized as an effective instrument to measure and objectively represent the extent of organizational silence within the healthcare environment. Furthermore, this tool fulfills the requirement stated in prior studies that emphasized the importance of avoiding organizational silence, improving the working conditions for healthcare professionals, detecting trends in organizational silence, and encouraging healthcare professionals to communicate comfortably with their superiors and subordinates [54].
The face validity assessment indicated that clinical nurses found the translated OSBS-P items to be relevant and necessary, with scores ranging from 2.5 to 3.8. This valuable feedback prompted significant improvements in item clarity and language simplicity. In contrast, the original Turkish version [21] encountered challenges with item complexity, leading to the removal of nine items due to low correlation coefficients (r = 0.19 to 0.74). Our proactive approach in incorporating feedback from end-users not only enhanced the scale’s clarity but also ensured its resonance with the target population. This participatory method is crucial as it aligns with best practices in survey design, emphasizing the importance of user-centered approaches in developing valid measurement tools.
Validity is a concept that refers to the extent to which a measurement tool accurately measures what it is intended to measure in the real world. It measures how well the tool aligns with the natural world and accurately captures the phenomenon being studied [55]. The scale’s content validity results further confirmed its robustness, with all items having CVR values exceeding 0.62 and CVI values above 0.79. In contrast to the original scale [21], which underwent item reduction based on similar validity assessments and resulted in a final count of 32 items, our study retained all items. We enhanced their relevance and clarity, improving the scale’s applicability in clinical settings. Our data showed no floor or ceiling effects, indicating that the OSBS-P scale could effectively capture a wide range of responses and reinforce its reliability as a measurement tool.
The construct validity of a scale refers to how accurately and consistently it measures the concept or construct it is designed to measure. In simpler terms, it determines whether the assessment instrument is measuring what it is intended to measure and whether the results obtained from the instrument reflect the underlying construct being measured [56]. The results of the CFA confirmed the statistical robustness of our four-factor model, with all items showing significant factor loadings (≥ 0.3). This represented a considerable improvement over the original scale, where several items had to be removed due to insufficient factor loadings. Our CFA revealed strong connections between the latent factors and their respective items, providing further support for the theoretical framework of the OSBS-P. The goodness-of-fit indices indicated that our model was acceptable, underscoring the effectiveness of our adaptation process. The original version’s use of EFA before CFA had resulted in a somewhat unstable model, requiring the removal of multiple items to attain satisfactory fit indices [21]. Our study’s methodology, incorporating qualitative and quantitative validation techniques, produced a more dependable and valid representation of the construct.
Based on the results of a reliability analysis, it has been determined that for a scale to be considered reliable, it must consistently reflect the underlying structure it intends to measure. This means that the scale should produce consistent and reliable results over time and accurately assess the construct it is designed to measure [57]. The internal consistency of the OSBS-P was evaluated using various statistical measures, resulting in a Cronbach’s alpha of 0.969, signifying excellent reliability. This surpasses the original Turkish version [21], which reported a Cronbach’s alpha of 0.93. The strong intraclass correlation coefficient (ICC = 0.951) further reinforces the scale’s stability across dimensions. These results indicate that our adapted version of the OSBS-P maintains high reliability and enhances the original tool’s consistency. The rigorous methodology we employed in assessing internal consistency, including the calculation of MacDonald’s omega coefficient, coefficient H, and mean inter-item correlations, establishes a strong basis for the reliability of our findings.
Our study showed that the translated OSBS-P is a valid and reliable tool for evaluating organizational silence in clinical settings. By overcoming previous research limitations and integrating expert input into our validation process, we have developed a strong instrument to improve comprehension of communication barriers within healthcare organizations and to facilitate future studies on their influence on clinical practice.
The authors have identified several limitations in evaluating the validity and reliability of the OSBS-P. The process of conducting CFA involved selecting participants using a convenience sampling method and recruiting individuals from educational and therapeutic centers affiliated with Ardabil University of Medical Sciences. While this approach facilitated efficient participant recruitment, it’s crucial to acknowledge that the external validity of the findings may be limited. Convenience sampling may not accurately represent the broader population, potentially weakening the generalizability of the results. It’s important to always consider this limitation when interpreting the findings of the CFA. Additionally, while the study included 338 clinical nurses, surpassing the minimum sample size requirement, it is crucial to recognize that the sample’s demographic characteristics may not reflect the diversity of clinical nurses in other regions or countries. Factors such as age, gender, years of experience, and cultural background could influence the psychometric properties of the OSBS-P, and the lack of representation from various demographics may limit the applicability of the findings. Another limitation is the reliance on self-reported measures, which can introduce biases such as social desirability bias, where participants may respond in a manner they believe is favorable rather than providing honest answers. This could affect the accuracy of the data collected and, consequently, the results of the CFA. Lastly, the study’s limited data collection period does not allow for examining changes over time or the causal relationships between the variables. Longitudinal studies would be beneficial in understanding how the psychometric properties of the OSBS-P may evolve in different settings or populations.
Despite the limitations, the study presented several notable strengths that contributed to understanding the OSBS-P’s psychometric properties. First and foremost, the study successfully conducted CFA on a substantial sample of 338 clinical nurses, which exceeded the minimum sample size requirements for robust statistical analysis. This large sample enhanced the reliability of the findings and provided a solid foundation for evaluating the scale’s psychometric characteristics. Furthermore, the study’s methodological rigor in employing CFA reflected a commitment to ensuring the OSBS-P’s psychometric soundness. By utilizing established statistical techniques, the authors provided a comprehensive scale evaluation, adding credibility to the results and reinforcing psychometric assessment’s importance in instrument development.
Conclusions
The Persian version of the Organizational Silence Behavior Scale (OSBS-P) demonstrates sound psychometric properties, including validity, reliability, and internal consistency. Due to its ease of administration, this tool can assess organizational silence behavior among clinical nurses. Therefore, future research should utilize larger sample sizes and include various hospital departments and groups to improve OSBS-P among clinical nurses in the healthcare environment. Thus, the study that has been conducted can prove to be extremely beneficial for nursing managers. It aims to identify and analyze the factors contributing to organizational silence behavior among clinical nurses. This analysis is based on using OSBS’ dimensions, which can help identify the strengths and weaknesses of the clinical nursing staff. By examining these factors, healthcare system managers and nursing policymakers can take necessary steps to improve the workplace environment and ensure clinical nurses can perform their duties effectively and efficiently. With this study, it is hoped that the issues related to organizational silence behavior among clinical nurses can be addressed and resolved, ultimately leading to better patient care and outcomes.
Supporting information
S1 File. Persian version of the OSBS.
https://doi.org/10.1371/journal.pone.0314155.s001
(DOCX)
S2 File. Original scale (Turkish).
https://doi.org/10.1371/journal.pone.0314155.s002
(DOCX)
Acknowledgments
The authors would like to thank all the clinical nurses and the officials of the hospitals affiliated with the Ardabil University of Medical Sciences, the Student Research Committee of Ardabil University of Medical Sciences, the Vice Chancellor for Research of Ardabil University of Medical Sciences, and all those who helped us in this study.
References
1. 1. Cho SH, Lee JY, You SJ, Song KJ, Hong KJ. Nurse staffing, nurses prioritization, missed care, quality of nursing care, and nurse outcomes. International journal of nursing practice. 2020;26(1): e12803. pmid:31850645
* View Article
* PubMed/NCBI
* Google Scholar
2. 2. Hickman SE, Lum HD, Walling AM, Savoy A, Sudore RL. The care planning umbrella: The evolution of advance care planning. Journal of the American Geriatrics Society. 2023;71(7):2350. pmid:36840690
* View Article
* PubMed/NCBI
* Google Scholar
3. 3. Vakilabad RN, Kheiri R, Islamzadeh N, Afshar PF, Ajri-Khameslou M. A survey of social well-being among employees, retirees, and nursing students: a descriptive-analytical study. BMC nursing. 2023;22(1):1–9.
* View Article
* Google Scholar
4. 4. AlThubaity DD, Mahdy Shalby AY. Perception of health teams on the implementation of strategies to decrease nursing errors and enhance patient safety. Journal of Multidisciplinary Healthcare. 2023:693–706. pmid:36938486
* View Article
* PubMed/NCBI
* Google Scholar
5. 5. Imam A, Obiesie S, Gathara D, Aluvaala J, Maina M, English M. Missed nursing care in acute care hospital settings in low-income and middle-income countries: a systematic review. Human resources for health. 2023;21(1):19. pmid:36918941
* View Article
* PubMed/NCBI
* Google Scholar
6. 6. Jafari MJ, Mostafazadeh P, Mojebi MR, Nemati-Vakilabad R, Mirzaei A. Identifying predictors of patient safety competency based on sleep quality in student faculty of nursing and midwifery during the internship period: a multidisciplinary study. BMC nursing. 2024;23(1):67. pmid:38267940
* View Article
* PubMed/NCBI
* Google Scholar
7. 7. De los Santos JAA, Rosales RA, Falguera CC, Firmo CN, Tsaras K, Labrague LJ, editors. Impact of organizational silence and favoritism on nurse’s work outcomes and psychological well‐being. Nursing Forum; 2020: Wiley Online Library.
8. 8. Yang J, Yang H, Wang B. Organizational silence among hospital nurses in China: a cross-sectional study. BioMed Research International. 2022;2022.
* View Article
* Google Scholar
9. 9. She J, Zhang R, Li Y, Mei Y, Li H. Effect of Ethical Leadership on Nurses’ Organizational Silence: The Mediating Role of Organizational Justice. Journal of Nursing Management. 2023;2023(1):9929435.
* View Article
* Google Scholar
10. 10. Doo EY, Kim M. Effects of hospital nurses’ internalized dominant values, organizational silence, horizontal violence, and organizational communication on patient safety. Research in nursing & health. 2020;43(5):499–510. pmid:32810300
* View Article
* PubMed/NCBI
* Google Scholar
11. 11. Farghaly Abdelaliem SM, Abou Zeid MAG. The relationship between toxic leadership and organizational performance: the mediating effect of nurses’ silence. BMC nursing. 2023;22(1):4. pmid:36600211
* View Article
* PubMed/NCBI
* Google Scholar
12. 12. Oyerinde OF. Leadership style, work environment, organizational silence and institutional effectiveness of polytechnic libraries, South-West Nigeria. International Information & Library Review. 2020;52(2):79–94.
* View Article
* Google Scholar
13. 13. Dehkharghani LL, Paul J, Maharati Y, Menzies J. Employee silence in an organizational context: A review and research agenda. European Management Journal. 2022.
* View Article
* Google Scholar
14. 14. Everett JL. Organizational culture and ethnoecology in public relations theory and practice. Public relations research annual: Routledge; 2020. p. 235–52.
15. 15. Van Ruler B. Communication theory: An underrated pillar on which strategic communication rests. Future directions of strategic communication: Routledge; 2020. p. 39–53.
16. 16. Brinsfield CT, Edwards MS. Employee voice and silence in organizational behavior. Handbook of research on employee voice: Edward Elgar Publishing; 2020. p. 103–20.
17. 17. Ahmad R, Nawaz MR, Ishaq MI, Khan MM, Ashraf HA. Social exchange theory: Systematic review and future directions. Frontiers in Psychology. 2023;13:1015921. pmid:36710813
* View Article
* PubMed/NCBI
* Google Scholar
18. 18. Labrague LJ, De los Santos JA. Association between nurse and hospital characteristics and organisational silence behaviours in nurses: A cross‐sectional study. Journal of Nursing Management. 2020;28(8):2196–204. pmid:32668491
* View Article
* PubMed/NCBI
* Google Scholar
19. 19. Dyne LV, Ang S, Botero IC. Conceptualizing employee silence and employee voice as multidimensional constructs. Journal of management studies. 2003;40(6):1359–92.
* View Article
* Google Scholar
20. 20. Knoll M, Van Dick R. Do I hear the whistle…? A first attempt to measure four forms of employee silence and their correlates. Journal of business ethics. 2013;113:349–62.
* View Article
* Google Scholar
21. 21. Yalçın B, Baykal Ü. Development and psychometric testing of the organizational silence behavior scale for healthcare professionals. Nursing & health sciences. 2019;21(4):454–60. pmid:31250957
* View Article
* PubMed/NCBI
* Google Scholar
22. 22. Lainidi O, Jendeby MK, Montgomery A, Mouratidis C, Paitaridou K, Cook C, et al. An integrative systematic review of employee silence and voice in healthcare: what are we really measuring? Frontiers in Psychiatry. 2023;14:1111579. pmid:37304444
* View Article
* PubMed/NCBI
* Google Scholar
23. 23. Chung S, Hwang J-I. Development of the Hospital Nurses’ Silence Behavior Scale. Journal of Korean Academy of Nursing. 2024;54(2):279–95.
* View Article
* Google Scholar
24. 24. Nemati-Vakilabad R, Mostafazadeh P, Mirzaei A. Investigating the Impact of Organizational Justice on the Relationship Between Organizational Learning and Organizational Silence in Clinical Nurses: A Structural Equation Modeling Approach. Journal of Nursing Management. 2024;2024(1):7267388.
* View Article
* Google Scholar
25. 25. Saeidipour B, Akbari P, Alizadeh Z. The role of organizational silence & organizational mobbing on the turnover intention. International Journal of Ethics and Society. 2021;3(1):59–69.
* View Article
* Google Scholar
26. 26. Etikan I, Musa SA, Alkassim RS. Comparison of convenience sampling and purposive sampling. American journal of theoretical and applied statistics. 2016;5(1):1–4.
* View Article
* Google Scholar
27. 27. World Health Organization (WHO). Process of translation and adaptation of instruments. 2007 [Online] 1 April 2017 Available from: http://www.who.int/substance_abuse/research_tools/translation/en/.
* View Article
* Google Scholar
28. 28. Polit DF, Yang FM. Measurement and the measurement of change: a primer for the health professions: Wolters Kluwer Philadelphia; 2016.
29. 29. Almanasreh E, Moles R, Chen TF. Evaluation of methods used for estimating content validity. Research in social and administrative pharmacy. 2019;15(2):214–21. pmid:29606610
* View Article
* PubMed/NCBI
* Google Scholar
30. 30. Terwee CB, Bot SD, de Boer MR, van der Windt DA, Knol DL, Dekker J, et al. Quality criteria were proposed for measurement properties of health status questionnaires. Journal of clinical epidemiology. 2007;60(1):34–42. pmid:17161752
* View Article
* PubMed/NCBI
* Google Scholar
31. 31. De Vet HC, Terwee CB, Mokkink LB, Knol DL. Measurement in medicine: a practical guide: Cambridge university press; 2011.
* View Article
* Google Scholar
32. 32. Hox JJ. Confirmatory factor analysis. The encyclopedia of research methods in criminology and criminal justice. 2021;2:830–2.
* View Article
* Google Scholar
33. 33. Shi D, Maydeu-Olivares A. The effect of estimation methods on SEM fit indices. Educational and psychological measurement. 2020;80(3):421–45. pmid:32425213
* View Article
* PubMed/NCBI
* Google Scholar
34. 34. Hair J, Black W, Babin B, Anderson R, Tatham R. Multivariate Data Analysis, Pearson Prentice-Hall, Upper Saddle River, NJ. 2006.
35. 35. Kline P. An easy guide to factor analysis: Routledge; 2014.
* View Article
* Google Scholar
36. 36. Anderson JC, Gerbing DW. The effect of sampling error on convergence, improper solutions, and goodness-of-fit indices for maximum likelihood confirmatory factor analysis. Psychometrika. 1984;49:155–73.
* View Article
* Google Scholar
37. 37. Boomsma A. Nonconvergence, improper solutions, and starting values in LISREL maximum likelihood estimation. Psychometrika. 1985;50:229–42.
* View Article
* Google Scholar
38. 38. Lt Hu, Bentler PM. Cutoff criteria for fit indexes in covariance structure analysis: Conventional criteria versus new alternatives. Structural equation modeling: a multidisciplinary journal. 1999;6(1):1–55.
* View Article
* Google Scholar
39. 39. Cattell R. The scientific use of factor analysis in behavioral and life sciences: Springer Science & Business Media; 2012.
40. 40. Muthén LK, Muthén BO. How to use a Monte Carlo study to decide on sample size and determine power. Structural equation modeling. 2002;9(4):599–620.
* View Article
* Google Scholar
41. 41. Wolf EJ, Harrington KM, Clark SL, Miller MW. Sample Size Requirements for Structural Equation Models: An Evaluation of Power, Bias, and Solution Propriety. Educ Psychol Meas. 2013;76(6):913–34. pmid:25705052
* View Article
* PubMed/NCBI
* Google Scholar
42. 42. Fornell C, Larcker DF. Evaluating structural equation models with unobservable variables and measurement error. Journal of marketing research. 1981;18(1):39–50.
* View Article
* Google Scholar
43. 43. Nemati-Vakilabad R, Khoshbakht-Pishkhani M, Maroufizadeh S, Javadi-Pashaki N. Translation and validation of the Persian version of the perception to care in acute situations (PCAS-P) scale in novice nurses. BMC Nurs. 2024;23(1):108. pmid:38326865
* View Article
* PubMed/NCBI
* Google Scholar
44. 44. Shrestha N. Factor analysis as a tool for survey analysis. American journal of Applied Mathematics and statistics. 2021;9(1):4–11.
* View Article
* Google Scholar
45. 45. Henseler J, Ringle CM, Sarstedt M. A new criterion for assessing discriminant validity in variance-based structural equation modeling. Journal of the academy of marketing science. 2015;43:115–35.
* View Article
* Google Scholar
46. 46. McNeish D. Thanks coefficient alpha, we’ll take it from here. Psychological methods. 2018;23(3):412. pmid:28557467
* View Article
* PubMed/NCBI
* Google Scholar
47. 47. Kalkbrenner MT. Alpha, omega, and H internal consistency reliability estimates: Reviewing these options and when to use them. Counseling Outcome Research and Evaluation. 2023;14(1):77–88.
* View Article
* Google Scholar
48. 48. Hashemian Moghadam A, Nemati-Vakilabad R, Imashi R, Yaghoobi Saghezchi R, Dolat Abadi P, Jamshidinia M, et al. The psychometric properties of the Persian version of the innovation support inventory (ISI-12) in clinical nurses: a methodological cross-sectional study. BMC Nursing. 2024;23(1):699. pmid:39342141
* View Article
* PubMed/NCBI
* Google Scholar
49. 49. Streiner DL, Norman GR, Cairney J. Health measurement scales: a practical guide to their development and use: Oxford University Press, USA; 2015.
50. 50. Spiliotopoulou G. Reliability reconsidered: Cronbach’s alpha and paediatric assessment in occupational therapy. Australian Occupational Therapy Journal. 2009;56(3):150–5. pmid:20854508
* View Article
* PubMed/NCBI
* Google Scholar
51. 51. Clark LA, Watson D. Constructing validity: Basic issues in objective scale development. 2016.
* View Article
* Google Scholar
52. 52. Xue C, Yuan J, Lo GG, Chang AT, Poon DM, Wong OL, et al. Radiomics feature reliability assessed by intraclass correlation coefficient: a systematic review. Quantitative imaging in medicine and surgery. 2021;11(10):4431. pmid:34603997
* View Article
* PubMed/NCBI
* Google Scholar
53. 53. Vinzi VE, Chin WW, Henseler J, Wang H. Perspectives on partial least squares. Handbook of partial least squares: Concepts, methods and applications: Springer; 2009. p. 1–20.
54. 54. Cetin A. Organizational silence and organizational commitment: a study of Turkish sport managers. Annals of Applied Sport Science. 2020;8(2):0-.
* View Article
* Google Scholar
55. 55. Sürücü L, Maslakci A. Validity and reliability in quantitative research. Business & Management Studies: An International Journal. 2020;8(3):2694–726.
* View Article
* Google Scholar
56. 56. Tavakol M, Wetzel A. Factor Analysis: a means for theory and instrument development in support of construct validity. International journal of medical education. 2020;11:245. pmid:33170146
* View Article
* PubMed/NCBI
* Google Scholar
57. 57. Heale R, Twycross A. Validity and reliability in quantitative studies. Evidence-based nursing. 2015;18(3):66–7. pmid:25979629
* View Article
* PubMed/NCBI
* Google Scholar
Citation: Mirzaei A, Jamshidinia M, Aghabarari M, Abadi PD, Nemati-Vakilabad R (2024) Psychometric evaluation and translation of the Persian version of the Organizational Silence Behavior Scale (OSBS-P) for clinical nurses. PLoS ONE 19(12): e0314155. https://doi.org/10.1371/journal.pone.0314155
About the Authors:
Alireza Mirzaei
Roles: Conceptualization, Data curation, Formal analysis, Writing – original draft, Writing – review & editing
Affiliation: Department of Emergency Nursing, School of Nursing and Midwifery, Ardabil University of Medical Sciences, Ardabil, Iran
ORICD: https://orcid.org/0000-0003-4242-2108
Mobina Jamshidinia
Roles: Data curation, Investigation, Methodology, Project administration, Validation, Visualization, Writing – original draft, Writing – review & editing
Affiliation: Student Research Committee, School of Nursing and Midwifery, Guilan University of Medical Sciences, Rasht, Iran
ORICD: https://orcid.org/0000-0002-4083-2907
Mehrzad Aghabarari
Roles: Data curation, Investigation, Methodology, Writing – original draft, Writing – review & editing
Affiliation: Student Research Committee, School of Nursing and Midwifery, Guilan University of Medical Sciences, Rasht, Iran
ORICD: https://orcid.org/0009-0001-9457-9011
Pouya Dolat Abadi
Roles: Conceptualization, Data curation, Writing – original draft, Writing – review & editing
Affiliation: Student Research Committee, School of Nursing and Midwifery, Tehran University of Medical Sciences, Tehran, Iran
ORICD: https://orcid.org/0000-0002-7281-2863
Reza Nemati-Vakilabad
Roles: Conceptualization, Methodology, Supervision, Writing – original draft, Writing – review & editing
E-mail: [email protected]
Affiliation: Department of Medical-Surgical Nursing, School of Nursing and Midwifery, Ardabil University of Medical Sciences, Ardabil, Iran
ORICD: https://orcid.org/0000-0002-8016-8625
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
[/RAW_REF_TEXT]
1. Cho SH, Lee JY, You SJ, Song KJ, Hong KJ. Nurse staffing, nurses prioritization, missed care, quality of nursing care, and nurse outcomes. International journal of nursing practice. 2020;26(1): e12803. pmid:31850645
2. Hickman SE, Lum HD, Walling AM, Savoy A, Sudore RL. The care planning umbrella: The evolution of advance care planning. Journal of the American Geriatrics Society. 2023;71(7):2350. pmid:36840690
3. Vakilabad RN, Kheiri R, Islamzadeh N, Afshar PF, Ajri-Khameslou M. A survey of social well-being among employees, retirees, and nursing students: a descriptive-analytical study. BMC nursing. 2023;22(1):1–9.
4. AlThubaity DD, Mahdy Shalby AY. Perception of health teams on the implementation of strategies to decrease nursing errors and enhance patient safety. Journal of Multidisciplinary Healthcare. 2023:693–706. pmid:36938486
5. Imam A, Obiesie S, Gathara D, Aluvaala J, Maina M, English M. Missed nursing care in acute care hospital settings in low-income and middle-income countries: a systematic review. Human resources for health. 2023;21(1):19. pmid:36918941
6. Jafari MJ, Mostafazadeh P, Mojebi MR, Nemati-Vakilabad R, Mirzaei A. Identifying predictors of patient safety competency based on sleep quality in student faculty of nursing and midwifery during the internship period: a multidisciplinary study. BMC nursing. 2024;23(1):67. pmid:38267940
7. De los Santos JAA, Rosales RA, Falguera CC, Firmo CN, Tsaras K, Labrague LJ, editors. Impact of organizational silence and favoritism on nurse’s work outcomes and psychological well‐being. Nursing Forum; 2020: Wiley Online Library.
8. Yang J, Yang H, Wang B. Organizational silence among hospital nurses in China: a cross-sectional study. BioMed Research International. 2022;2022.
9. She J, Zhang R, Li Y, Mei Y, Li H. Effect of Ethical Leadership on Nurses’ Organizational Silence: The Mediating Role of Organizational Justice. Journal of Nursing Management. 2023;2023(1):9929435.
10. Doo EY, Kim M. Effects of hospital nurses’ internalized dominant values, organizational silence, horizontal violence, and organizational communication on patient safety. Research in nursing & health. 2020;43(5):499–510. pmid:32810300
11. Farghaly Abdelaliem SM, Abou Zeid MAG. The relationship between toxic leadership and organizational performance: the mediating effect of nurses’ silence. BMC nursing. 2023;22(1):4. pmid:36600211
12. Oyerinde OF. Leadership style, work environment, organizational silence and institutional effectiveness of polytechnic libraries, South-West Nigeria. International Information & Library Review. 2020;52(2):79–94.
13. Dehkharghani LL, Paul J, Maharati Y, Menzies J. Employee silence in an organizational context: A review and research agenda. European Management Journal. 2022.
14. Everett JL. Organizational culture and ethnoecology in public relations theory and practice. Public relations research annual: Routledge; 2020. p. 235–52.
15. Van Ruler B. Communication theory: An underrated pillar on which strategic communication rests. Future directions of strategic communication: Routledge; 2020. p. 39–53.
16. Brinsfield CT, Edwards MS. Employee voice and silence in organizational behavior. Handbook of research on employee voice: Edward Elgar Publishing; 2020. p. 103–20.
17. Ahmad R, Nawaz MR, Ishaq MI, Khan MM, Ashraf HA. Social exchange theory: Systematic review and future directions. Frontiers in Psychology. 2023;13:1015921. pmid:36710813
18. Labrague LJ, De los Santos JA. Association between nurse and hospital characteristics and organisational silence behaviours in nurses: A cross‐sectional study. Journal of Nursing Management. 2020;28(8):2196–204. pmid:32668491
19. Dyne LV, Ang S, Botero IC. Conceptualizing employee silence and employee voice as multidimensional constructs. Journal of management studies. 2003;40(6):1359–92.
20. Knoll M, Van Dick R. Do I hear the whistle…? A first attempt to measure four forms of employee silence and their correlates. Journal of business ethics. 2013;113:349–62.
21. Yalçın B, Baykal Ü. Development and psychometric testing of the organizational silence behavior scale for healthcare professionals. Nursing & health sciences. 2019;21(4):454–60. pmid:31250957
22. Lainidi O, Jendeby MK, Montgomery A, Mouratidis C, Paitaridou K, Cook C, et al. An integrative systematic review of employee silence and voice in healthcare: what are we really measuring? Frontiers in Psychiatry. 2023;14:1111579. pmid:37304444
23. Chung S, Hwang J-I. Development of the Hospital Nurses’ Silence Behavior Scale. Journal of Korean Academy of Nursing. 2024;54(2):279–95.
24. Nemati-Vakilabad R, Mostafazadeh P, Mirzaei A. Investigating the Impact of Organizational Justice on the Relationship Between Organizational Learning and Organizational Silence in Clinical Nurses: A Structural Equation Modeling Approach. Journal of Nursing Management. 2024;2024(1):7267388.
25. Saeidipour B, Akbari P, Alizadeh Z. The role of organizational silence & organizational mobbing on the turnover intention. International Journal of Ethics and Society. 2021;3(1):59–69.
26. Etikan I, Musa SA, Alkassim RS. Comparison of convenience sampling and purposive sampling. American journal of theoretical and applied statistics. 2016;5(1):1–4.
27. World Health Organization (WHO). Process of translation and adaptation of instruments. 2007 [Online] 1 April 2017 Available from: http://www.who.int/substance_abuse/research_tools/translation/en/.
28. Polit DF, Yang FM. Measurement and the measurement of change: a primer for the health professions: Wolters Kluwer Philadelphia; 2016.
29. Almanasreh E, Moles R, Chen TF. Evaluation of methods used for estimating content validity. Research in social and administrative pharmacy. 2019;15(2):214–21. pmid:29606610
30. Terwee CB, Bot SD, de Boer MR, van der Windt DA, Knol DL, Dekker J, et al. Quality criteria were proposed for measurement properties of health status questionnaires. Journal of clinical epidemiology. 2007;60(1):34–42. pmid:17161752
31. De Vet HC, Terwee CB, Mokkink LB, Knol DL. Measurement in medicine: a practical guide: Cambridge university press; 2011.
32. Hox JJ. Confirmatory factor analysis. The encyclopedia of research methods in criminology and criminal justice. 2021;2:830–2.
33. Shi D, Maydeu-Olivares A. The effect of estimation methods on SEM fit indices. Educational and psychological measurement. 2020;80(3):421–45. pmid:32425213
34. Hair J, Black W, Babin B, Anderson R, Tatham R. Multivariate Data Analysis, Pearson Prentice-Hall, Upper Saddle River, NJ. 2006.
35. Kline P. An easy guide to factor analysis: Routledge; 2014.
36. Anderson JC, Gerbing DW. The effect of sampling error on convergence, improper solutions, and goodness-of-fit indices for maximum likelihood confirmatory factor analysis. Psychometrika. 1984;49:155–73.
37. Boomsma A. Nonconvergence, improper solutions, and starting values in LISREL maximum likelihood estimation. Psychometrika. 1985;50:229–42.
38. Lt Hu, Bentler PM. Cutoff criteria for fit indexes in covariance structure analysis: Conventional criteria versus new alternatives. Structural equation modeling: a multidisciplinary journal. 1999;6(1):1–55.
39. Cattell R. The scientific use of factor analysis in behavioral and life sciences: Springer Science & Business Media; 2012.
40. Muthén LK, Muthén BO. How to use a Monte Carlo study to decide on sample size and determine power. Structural equation modeling. 2002;9(4):599–620.
41. Wolf EJ, Harrington KM, Clark SL, Miller MW. Sample Size Requirements for Structural Equation Models: An Evaluation of Power, Bias, and Solution Propriety. Educ Psychol Meas. 2013;76(6):913–34. pmid:25705052
42. Fornell C, Larcker DF. Evaluating structural equation models with unobservable variables and measurement error. Journal of marketing research. 1981;18(1):39–50.
43. Nemati-Vakilabad R, Khoshbakht-Pishkhani M, Maroufizadeh S, Javadi-Pashaki N. Translation and validation of the Persian version of the perception to care in acute situations (PCAS-P) scale in novice nurses. BMC Nurs. 2024;23(1):108. pmid:38326865
44. Shrestha N. Factor analysis as a tool for survey analysis. American journal of Applied Mathematics and statistics. 2021;9(1):4–11.
45. Henseler J, Ringle CM, Sarstedt M. A new criterion for assessing discriminant validity in variance-based structural equation modeling. Journal of the academy of marketing science. 2015;43:115–35.
46. McNeish D. Thanks coefficient alpha, we’ll take it from here. Psychological methods. 2018;23(3):412. pmid:28557467
47. Kalkbrenner MT. Alpha, omega, and H internal consistency reliability estimates: Reviewing these options and when to use them. Counseling Outcome Research and Evaluation. 2023;14(1):77–88.
48. Hashemian Moghadam A, Nemati-Vakilabad R, Imashi R, Yaghoobi Saghezchi R, Dolat Abadi P, Jamshidinia M, et al. The psychometric properties of the Persian version of the innovation support inventory (ISI-12) in clinical nurses: a methodological cross-sectional study. BMC Nursing. 2024;23(1):699. pmid:39342141
49. Streiner DL, Norman GR, Cairney J. Health measurement scales: a practical guide to their development and use: Oxford University Press, USA; 2015.
50. Spiliotopoulou G. Reliability reconsidered: Cronbach’s alpha and paediatric assessment in occupational therapy. Australian Occupational Therapy Journal. 2009;56(3):150–5. pmid:20854508
51. Clark LA, Watson D. Constructing validity: Basic issues in objective scale development. 2016.
52. Xue C, Yuan J, Lo GG, Chang AT, Poon DM, Wong OL, et al. Radiomics feature reliability assessed by intraclass correlation coefficient: a systematic review. Quantitative imaging in medicine and surgery. 2021;11(10):4431. pmid:34603997
53. Vinzi VE, Chin WW, Henseler J, Wang H. Perspectives on partial least squares. Handbook of partial least squares: Concepts, methods and applications: Springer; 2009. p. 1–20.
54. Cetin A. Organizational silence and organizational commitment: a study of Turkish sport managers. Annals of Applied Sport Science. 2020;8(2):0-.
55. Sürücü L, Maslakci A. Validity and reliability in quantitative research. Business & Management Studies: An International Journal. 2020;8(3):2694–726.
56. Tavakol M, Wetzel A. Factor Analysis: a means for theory and instrument development in support of construct validity. International journal of medical education. 2020;11:245. pmid:33170146
57. Heale R, Twycross A. Validity and reliability in quantitative studies. Evidence-based nursing. 2015;18(3):66–7. pmid:25979629
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2024 Mirzaei et al. This is an open access article distributed under the terms of the Creative Commons Attribution License: http://creativecommons.org/licenses/by/4.0/ (the “License”), which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Background
Healthcare professionals have a crucial responsibility to provide optimal patient care. However, maintaining silence within an organization can often lead to ethical dilemmas and negatively impact the quality of care. Healthcare professionals must speak up and ensure their concerns are heard and addressed to promote a safe and ethical healthcare environment. This study aimed to translate and assess the psychometric properties of the Persian version of the Organizational Silence Behavior Scale (OSBS-P) for clinical nurses.
Methods
A methodological study was conducted on 338 clinical nurses to evaluate the psychometric properties of the Persian version of the 32-item OSBS-P. A multilevel approach was used: forward-backward translation, face and content validity, construct validity (confirmatory factor analysis), and reliability (internal consistency and stability) of the Persian version of the OSBS-P were analyzed, respectively.
Results
The psychometric evaluation demonstrated that the Persian version of OSBS-P is not only understandable but also retains the conceptual integrity of the original English version. Confirmatory factor analysis validated its alignment with the proposed four-factor model, indicating robust construct validity. The OSBS-P’s reliability was excellent, with a Cronbach’s alpha coefficient of α = 0.969, and stability was confirmed with an intraclass correlation coefficient (ICC) of 0.951 (95% CI, 0.919–0.945).
Conclusion
The OSBS-P is a valid tool for evaluating the organizational silence behavior of clinical nurses.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer