Student engagement is the time and energy students devoted to educationally sound activities (Kuh, 2003), which is regarded as a key factor that facilitates school completion, and enhances motivation for achievements (Appleton et al., 2008). Student engagement is considered an important factor that leads to educational reform and evaluation, which mediates the influence of curricular policy and instructional reforms on students' achievements (Guthrie & Wigfield, 2000). Both Coates (2010) and Kuh (2003) consistently suggested that student engagement should be regarded as an indicator of institutional quality in higher education by encouraging students to participate in educationally purposeful activities (Groccia, 2018). The measurement of student engagement has been an important research topic as student engagement is never taken for granted and requires assessment and promotion. An increasing number of higher education institutes has emphasized the importance of advocating student engagement which aimed to improve students' achievement in a range of learning environments, and to tackle common problems in learning such as feelings of boredom, dropouts and attrition (Dixson, 2015; Walji et al., 2016).
With the advancement of technology in education, new learning environments have been created such as integrating online technology and face-to-face components. The blended learning environments such as the integration of e-learning (O'Neil et al., 2013), mobile learning (Pimmer et al., 2014) and simulation (Khalaila, 2014) are nowadays common in healthcare education. The online components are flexible and offer open education for students to complement practical skills training and problem-solving training based on current workplace scenarios. Thus such kind of blended learning has been favoured by medical schools and healthcare educational institutes in recent years (Goh & Sandars, 2020), which better meets individual learning needs with timely responses, and in which it is easier to design the teaching materials according to realistic working environments (McCowan, 2017). The e-learning environment allows flexibility for students to create a self-directed learning tool that can be adjusted according to their learning pace (Shorey et al., 2018) and can be integrated into the curricula to achieve improved learning outcomes (Kowitlawakul et al., 2017). Previous studies have shown that similar academic performance could be achieved in online and face-to-face learning modalities (Kemp & Grieve, 2014; Yen et al., 2018), and student engagement did not significantly differ between online and face-to-face learning (Brielmaier & Kuo, 2016; Butts et al., 2013).
BACKGROUNDIn the literature, there are a number of psychometric scales aiming at measuring student engagement, but the key constructs, theoretical perspectives and educational contexts of concern varied to a certain extent (Sinatra et al., 2015) and most of them suffered from methodological limitations in the scale development. Previous studies proposed a variety of factors or strategies which might raise students' engagement in learning such as interpersonal relationships, problem-based learning as a team (Amerstorfer & Freiin von Münster-Kistner, 2021), inclusiveness, active teaching strategies (Arjomandi et al., 2018), senses of involvement and expectation (Bowden et al., 2021). As for data analysis, a partial data-driven approach, that is, only conducting confirmatory factor analysis (CFA; Liem & Martin, 2012; Pintrich et al., 1993), or running exploratory factor analysis (EFA) separately on each predefined groups of items (Cho, 2012), might not identify distinct factors which have low correlations with others. Some other scales included items which assessed student's appraisal of the school, people and academic activities (Appleton et al., 2006; NSSE, 2021). Although these appraisals might have positive associations with student engagement, these items did not directly measure student's behavioural, emotional or cognitive involvement in educational activities. In the literature, there was an attempt of modifying a student engagement scale (Handelsman et al., 2005) to adapt it to measuring student engagement online (Dixson, 2015). However, the factor analysis procedure was unclear, and the factor of academic performance was disputable based on previous studies (Fredricks et al., 2004). Furthermore, there is no such an instrument that fits for measuring the student engagement both online and face-to-face learning.
Research questionOur study aimed to develop a generic student engagement scale, that is, GSES to measure the construct in a contemporary learning environment which leverages both face-to-face and online learning in a sample of students at the school of nursing and health studies. Our results will inform healthcare educators about the policy and practice to promote student engagement leading to positive educational outcomes.
THE STUDY DesignThis was a cross-sectional psychometric study testing reliability and validity of GSES questionnaire in a sample of adult students in their first or second year of higher diploma or bachelor's degree at the school of nursing and health studies, undertaking ordinary online and face-to-face learning at a local university in Hong Kong. Ordinary learning approaches for nursing students included lectures in lecture halls, tutorials in classrooms, nursing skill practice in laboratories with medical equipment, online lectures, online self-learning modules, etc. This study contained two phases. The first phase comprised instrument modification and examining validity. The second phase studied the instrument reliability and factor structure.
Method Phase 1 (four steps): Instrument modification and examining validity Step 1: Item modification and translationOur study developed from previous work on Distance Student Engagement Scale (DSES; Li & Yu, 2015) which was formulated after extensive literature review based on three domains including behavioural involvements, emotional states and cognitive strategies (Fredricks et al., 2004) which were well recognized in the literature (Sinatra et al., 2015). We obtained permission from the original author of DSES before modification. We carefully modified the items to increase adaptability, conciseness and comprehensibility (Liu et al., 2020) of GSES (Appendix S1); and translated Chinese contents into English by following an adopted protocol (Sousa & Rojjanasrirat, 2011). Forty-five items of DSES had been included among which 38 were modified.
Step 2: Examining content validitySeven experts, including academics with experiences in nursing (n = 3), higher education (n = 3) or social science (n = 1), used four-point ordinal scale (1 = not relevant to 4 = highly relevant) to evaluate the item relevancy and a dichotomous scale (Yes/No) to rate the scale adequacy. They were also invited to comment on items with unsatisfactory ratings. The cut-offs of item-content validity index (I-CVI) were set at 0.8 and scale level CVI (S-CVI) was 0.9 above which indicates good content validity (Polit et al., 2007; Portney & Watkins, 2015).
Step 3: Examining face validityTwenty undergraduate or higher diploma students from the school of nursing and health studies were recruited randomly to rate the item comprehensibility on a dichotomous scale (Yes/No), and rephrase each item in their own words (i.e. the interpretability; Lam, 2015; Lam et al., 2017; Streiner et al., 2014). The researcher rated the accuracy and appropriateness of the rephrased item by using an ordinal scale (1 = fully correct to 4 = totally wrong; Portney & Watkins, 2015). Items interpretable to students were formulated after a discussion between student and researcher.
Step 4: Standardizing terms useThe research team rephrased the terms in GSES in a uniform manner and revised the items into statements with similar usage and word order to facilitate comprehensibility among respondents. The major principles of item construction and design followed were (1) maintaining clarity, (2) preference for short statements, (3) avoidance of double negatives, (4) avoidance of double-barrelled statements and (5) avoidance of factual statements (Lam, 2015; Mishel, 1998; Streiner et al., 2014).
Phase 2: Instrument reliability and factor structureEach GSES item was measured with a five-point ordinal scale from “1 = not true at all” to “5 = absolutely true.” Scores of negatively keyed items were reversed in analysis, so higher global score indicated better student engagement. On average, each participant spent 20 min to complete the questionnaire at the university campus. Eligibility criteria of student recruitment were (1) full-time undergraduates or higher diploma students, (2) enrolled in the school of nursing and health studies and (3) who had online and face-to-face learning experiences. By convenience sampling, the targeted sample size was 633 according to similar study (Li & Yu, 2015). Data were collected between 2016 and 2018. The statistical significance level was set at .05. Analyses were done on R 4.0.3 (R Core Team, 2021) and RStudio 1.4.1103 (RStudio, 2021).
Beyond the sample for factor analysis, another sample of 70 students were randomly recruited to complete GSES at baseline and 4 weeks later (Lam et al., 2018). Two-way mixed model was used to compute intraclass correlation coefficient (ICC 3,1; Koo & Li, 2016), and the cutoff was set at .75 above which indicates good test–retest reliability (Portney & Watkins, 2015).
Normality, factorability and parallel analysisHalf of the sample were randomly drawn for EFA of the GSES (n = 225), another half was used for CFA (n = 226). The normality of EFA data was checked (RDocumentation, 2021f). A significant Bartlett's test result and Kaiser–Meyer–Olkin's measure of sampling adequacy (KMO's MSA) > 0.6 (RDocumentation, 2021d; Tabachnick & Fidell, 2007) indicate acceptable factorability of the correlation matrix. Parallel analysis was run with scree plot created (RDocumentation, 2021b). Spearman correlation matrix was analysed where eigenvalues for principal component analysis and principal axis factoring were computed with 500 simulated analyses.
Exploratory factor analysisModels were run for factoring methods of minimum residual (MINRES), unweighted least square (ULS), weighted least square (WLS) and principal axis factoring (PA) respectively (RDocumentation, 2021c). The rotation method was direct oblimin assuming that factors could be correlated measuring the same construct. Standardized loadings on pattern matrix were based on Spearman correlation analysed by ULS. Communalities, uniqueness and Hoffman's index of complexity were computed (Hofmann, 1978; Pettersson & Turkheimer, 2010). Variance explained, factor correlations, and factor score adequacy in terms of its correlation and multiple R2 with factors, and minimum correlation of possible factor scores, were evaluated.
Confirmatory factor analysisThe cutoff of the complexity index (Hofmann, 1978) was empirically determined as, with respect to each item, the number of factors having loadings larger than 0.3 became more than one or none, or having loadings ranging between 0.15 and 0.3 were multiple, when the complexity index reached 2.1 or above. On the other hand, at least three items per factor should be maintained (Marsh et al., 1998; Robinson, 2018). Therefore, items with a complexity index larger than 2.1 were excluded from analysis. Each item was assigned to the factor on which the loading was the highest across all factors. The model estimator was ULS (RDocumentation, 2021a, 2021e). The optimization method was nlminb. The exogenous latent variables were assumed to be correlated. Results were plotted with circular layout (Figure 1; RDocumentation, 2021h). Factor loadings larger than 0.3 were considered acceptable (Hon et al., 2013; Izquierdo et al., 2014; Maskey et al., 2018; Omondi Aduda et al., 2014; Wu, 2008; Yong & Pearce, 2013). The cut-offs of fit indices were that the comparative fit index (CFI) and Tucker–Lewis index (TLI) should be larger than 0.95 (Hu & Bentler, 1999); the standardized root mean squared residual (SRMR) should be smaller than 0.08 (Cho et al., 2020; Hooper et al., 2008; Hu & Bentler, 1999) and root mean square error of approximation (RMSEA) should be smaller than 0.08 (Browne & Cudeck, 1993; Lam et al., 2018, 2020; Liu et al., 2020; MacCallum et al., 1996) to indicate good model fit. A Cronbach's alpha above 0.7 (Tavakol & Dennick, 2011) and average variance extracted (AVE; RDocumentation, 2021g) above 0.4 (Fornell & Larcker, 1981) were considered acceptable.
FIGURE 1. Circular plot of standardized loadings, residual variances and covariances of exogenous variables, that is latent factors.
Initially, 665 students participated in the study with written informed consent. Then, 214 cases were removed after team discussions because of habitual responses and incomplete data. Hence the attrition rate was 32% and it left 451 cases for analysis. The sample age ranged between 18 and 33 years (mean = 21). Females accounted for 77% of the cases (n = 348). There were 163 and 91 higher diploma students who were studying in year 1 and 2 respectively. Regarding bachelor's degree, there were 112 and 91 students who were, respectively, studying in year 1 and 2.
Scale reliability and validityMost of the items had I-CVI ranged between 0.80 and 1.00 which were satisfactory. Six items scored 0.71 (item 12, 17, 23, 28, 34, 46) which was considered acceptable (Polit et al., 2007). Overall, the S-CVI was satisfactory (0.92). All items had got 100% ratings of comprehensibility and interpretability which supported face validity in the targeted population. The test–retest reliability was satisfactory (ICC = .88, 95% CI [0.81–0.92], p < .001).
Normality, factorability and number of factorsHalf of the cases (n = 225) were randomly drawn from the sample for EFA of the GSES, and the rest of the cases were analysed in CFA. For EFA data, the assumption of normality was not met (p < .001). The Bartlett's test statistics were significant (p < .001) for Kendall's correlation, Spearman's correlation as well as polychoric correlation although it was non-positive definite (NPD) with smoothing done (Lorenzo-Seva & Ferrando, 2021). The MSA of Kendall's correlation and Spearman's correlation were satisfactory (>0.85), but unsatisfactory for polychoric correlation (0.18). Parallel analysis suggested five factors and four components on scree plot.
Exploratory factor analysisIn the EFA of the GSES, the results between estimation methods were close. Generally, the root mean square of residuals, TLI and RMSEA were 0.046, 0.816 and 0.053 (90% CI = [0.048, 0.058]) respectively. Yet, the empirical chi square of WLS was the highest (945.96) and the values of MINRES and ULS were the lowest (945.46). The loadings are shown in Table 1. When the communalities were higher, the uniqueness was lower. However, the complexity index was neither correlated with communalities nor uniqueness. The higher the complexity index, the more obvious the cross-loading (Osborne et al., 2008). The GSES contains five factors where 1 = “self-regulated learning,” 2 = “cognitive strategy use,” 3 = “experienced emotion,” 4 = “teacher-student interaction,” and 5 = “enjoyment of school life.” The proportional variance explained by factor 1 to 5 were 10, 10, 7, 8 and 4 percent respectively. The factor correlations across factor 1 to factor 4 ranged between 0.32 and 0.43; however, their correlations with factor 5 were less than 0.3 (0.06–0.22). The factor score adequacy of the first four factors in terms of correlation of regression scores with factors (0.90–0.94), and multiple R2 of scores with factors (0.80–0.88), as well as minimum correlation of possible factor scores (0.61–0.75) were better than the findings of factor 5 which were 0.88, 0.77 and 0.53 respectively.
TABLE 1 Standardized loadings (pattern matrix) based upon correlation matrix in EFA
Item | Factor 1 | Factor 2 | Factor 3 | Factor 4 | Factor 5 | h2 | u2 | c |
1 | .19 | .81 | 2.4 | |||||
2 | .41 | .24 | .76 | 2.7 | ||||
3 | .32 | .19 | .81 | 1.7 | ||||
4 | .55 | .34 | .66 | 1.0 | ||||
5 | .37 | .33 | .36 | .64 | 2.6 | |||
6 | .55 | .46 | .54 | 1.7 | ||||
7 | .58 | .43 | .57 | 1.1 | ||||
8 | .61 | .41 | .59 | 1.0 | ||||
9 | .42 | .21 | .79 | 1.1 | ||||
10 | .32 | .36 | .64 | 2.9 | ||||
11 | .43 | .31 | .69 | 1.4 | ||||
12 | .15 | .85 | 2.5 | |||||
13 | .48 | .38 | .62 | 1.7 | ||||
14 | .55 | .39 | .61 | 1.1 | ||||
15 | .55 | .37 | .63 | 1.4 | ||||
16 | .47 | .49 | .51 | 1.9 | ||||
17 | .31 | .53 | .65 | .35 | 2.1 | |||
18 | .34 | .43 | .45 | .55 | 2.3 | |||
19 | .39 | .24 | .76 | 2.0 | ||||
20 | .41 | .59 | 3.5 | |||||
21 | .31 | .69 | 3.9 | |||||
22 | .27 | .73 | 3.8 | |||||
23 | .62 | .42 | .58 | 1.1 | ||||
24 | .33 | .29 | .71 | 2.1 | ||||
25 | .57 | .37 | .63 | 1.3 | ||||
26 | .49 | −.36 | .58 | .42 | 2.8 | |||
27 | .33 | −.33 | .33 | .67 | 3.4 | |||
28 | .68 | .48 | .52 | 1.1 | ||||
29 | .34 | .35 | .65 | 3.2 | ||||
30 | .50 | .51 | .49 | 2.0 | ||||
31 | .54 | .49 | .51 | 1.3 | ||||
32 | .75 | .64 | .36 | 1.0 | ||||
33 | .68 | .50 | .50 | 1.1 | ||||
34 | .42 | .34 | .66 | 1.5 | ||||
35 | .25 | .75 | 3.0 | |||||
36 | .39 | .61 | 3.0 | |||||
37 | .35 | .33 | .67 | 2.0 | ||||
38 | .40 | .32 | .41 | .59 | 2.6 | |||
39 | .37 | .39 | .61 | 2.9 | ||||
40 | .47 | .49 | .51 | 1.9 | ||||
41 | .60 | .50 | .50 | 1.3 | ||||
42 | .73 | .59 | .41 | 1.1 | ||||
43 | .59 | .48 | .52 | 1.2 | ||||
44 | .63 | .53 | .47 | 1.7 | ||||
45 | .65 | .45 | .55 | 1.1 |
Note: Loadings with absolute values smaller than 0.3 were suppressed. Bolded = the highest loading of an item where c ≤ 2.1; c, complexity of factor loadings; h2, communalities; u2, uniqueness.
Confirmatory factor analysisIn the CFA of the 29-item GSES, the parameter estimation ended after 59 iterations. The model chi-square to df ratio was 635.45 (df = 367), compared with the 9,781.24 (df = 406) of the baseline model, that is, null model estimating means and variances. The CFI, TLI, Bentler's SRMR and RMSEA were 0.971, 0.968, 0.078 and 0.057 (90% CI = [0.050, 0.064]; p = .062) respectively. Overall, the fit indices of the instrument were satisfactory. From the results, items related to emotion management and metacognition could be explained by factor 1 (Figure 1). Items associated with memory and summarization skills were indicators of factor 2. Items of resource management were found in the first two factors. In factor 3, it contained items of curiosity and tiresome feelings. Factor 4 comprised items of participation, persistence and interaction. Indicators associated with happiness and sense of belonging were explained by factor 5. The standardized factor loadings in the first three factors ranged from 0.55 to 0.77. The loadings of factor 4 ranged between 0.42 and 0.64. In factor 5, when the loading of the first indicator was constrained to 1, the standardized loadings of the other two items were 0.79 and 0.48 respectively. The loading on item 19 was the smallest. Regarding standardized covariance, the smallest magnitude was the one between factor 2 and 5, and the largest one was covariance between factor 1 and 4. As to variance, the result of item 16 was the smallest which was less than 0.1. The second smallest variance was from factor 4 (0.35). The variance of factor 5 was the largest which was above 0.9. The second largest variance came from item 7 (0.83). The reliability measures of the first three factors in the 29-item GSES were acceptable as Cronbach's alpha ranged between 0.83 and 0.88 (factor 1 to 3 = 0.88, 0.85, 0.83), and the AVE ranged from 0.45 to 0.53 (factor 1 to 3 = 0.48, 0.53, 0.45). The coefficients of factor 5 were good (Cronbach's alpha = 0.80, AVE = 0.61). However, the results of factor 4 were marginal according to the criteria (Cronbach's alpha = 0.76, AVE = 0.31).
DISCUSSION Evidence in contextOur study aimed at developing the GSES (Appendix 1) to measure student engagement in face-to-face and online learning in a sample of students at school of nursing and health studies. The factor “self-regulated learning” or reflection on learning progress was similar to the “Adaptive Behaviour” involving planning, task management and persistence in the Motivation and Engagement Scale (Liem & Martin, 2012). The factor “cognitive strategy use” was related to the educational indicators of “Reflective & Integrative Learning,” “Learning Strategies” and “Quantitative Reasoning” in the US National Survey of Student Engagement (NSSE, 2021). The indicators of “collaborative learning” and “student–faculty interaction” in NSSE (2021) were similar to the GSES factor “teacher–student interaction” associated with student's interaction with teachers and students. However, we also found items of reserving time for study and participating in optional learning activities in this factor. “Teacher–student interaction” was not explicit in other instruments tested on either adolescents (Appleton et al., 2006) or undergraduates (NSSE, 2021). Regarding emotional domain, we recognized factors of “experienced emotion” and “enjoyment of school life” which had the highest covariance among all factors. While these factors were not addressed in NSSE (2021), Student Engagement Instrument (Appleton et al., 2006) contained some items such as enjoying talking to teachers and students which might indicate enjoyment of school life but these items belonged to factors “teacher–student relationship” and “peer support for learning.” Liem and Martin (2012) included items which assessed student's feelings and emotions towards schoolwork, but these items were categorized into “adaptive cognition” or “impeding cognition” either facilitating or impeding learning in their non-data-driven theoretical framework of motivation. Compared with other student engagement scales, the GSES provides a more comprehensive assessment of student engagement addressing behavioural, cognitive and also emotional domains.
Added value and implicationsIn this study, we had balanced time and workload in designing item structure and response modality (Lam, 2015; Mishel, 1998), and obtained satisfactory results. We recognized factors such as “self-regulated learning” and “cognitive strategy use” which fall into behavioural and cognitive engagement domains. In GSES, the factor “teacher-student interaction” is not only behaviour but also attitude of proactiveness reflected in the time and resources used for study preparation. Previous study highlighted independence, autonomy and freedom of learning in student engagement (Christenson et al., 2012). Self-regulation and reflection should be focus areas in higher education to improve education quality (Chen et al., 2019; Hew, 2016). Our key findings will undergo careful translations into clear policy, educational practice and course units to improve learning in higher education. Subgroup analysis in terms of age, gender and other important characteristics could be future research directions. Concerning emotional factors, “experienced emotion” and “enjoyment of school life” have significance to achieving positive learning outcomes. Review paper reported that emotions could have impacts on academic achievement which might be mediated by cognitive process and school peer relationships (Carmona-Halty et al., 2021; MacCann et al., 2020; Valiente et al., 2012). A cross-sectional study showed that better emotional functioning was associated with higher academic achievement (Sadeghi Bahmani et al., 2018). Teaching in higher education, which incorporates information technology and equipment, has also been advanced (Goh & Sandars, 2020). COVID-19 pandemic and lockdown changed the dynamics of learning environment in which integration of online learning and traditional education becomes more popular (Basilaia & Kvavadze, 2020). This GSES allows cross-method comparison for both online and face-to-face learning, which fits the trend of blended-learning in healthcare education in the 21st century nowadays. With a new and better understanding on the students' engagement, policy should be formulated to seize new opportunities raising healthcare education quality in the new normal.
Future directionsIn this study, we offered explicit descriptions and justifications of item modifications which could facilitate adoption of the scale and multiple group comparisons in the future (Lam, 2015; Lam et al., 2017; Portney & Watkins, 2015). Moreover, a larger sample size and a higher number of categories on ordinal scale could favour the implementation of parametric methods and the interpretation of relevant fit measures which are commonly used (Jia & Wu, 2019; Li, 2016; Rhemtulla et al., 2012). If the research conditions allow random sampling, it could increase the representativeness of the sample and the generalizability of the results in the future.
LimitationsThirty-two percent of the recruited cases were excluded from analysis due to habitual responses and incomplete data. The reasons for missing data could not be explored because the survey was anonymous. Concerning generalisability, we recognized that the learning mode regarding online and face-to-face approaches as well as sample characteristics could vary between institutions and locations. As for sample size in factor analysis, although de Winter et al. (2009) reported that the sample size in EFA could be as small as 50, and Wolf et al. (2013) argued that there was no rule of thumb for sample size in CFA. We decided to divide the sample into two halves for possible comparison in terms of sample size and fit measures.
CONCLUSIONThis study developed 29-item GSES to measure student engagement in a sample of students at school of nursing and health studies. The EFA identified latent factors such as “self-regulated learning” and “cognitive strategy use” which are consistent with the literature, and factor “teacher–student interaction” which is new finding, and other factors “experienced emotion” and “enjoyment of school life” which had not been emphasized in relevant instruments previously. The construct validity of GSES was supported by good fit indices in CFA. Validity, factor reliability and test–retest reliability were satisfactory. The limitations of missing data and generalizability, as well as the implications of tailoring engagement promotion by focusing on factors identified in this study shed lights on new insights into future research directions.
AUTHOR CONTRIBUTIONS“REDACTED” conceived the study. “REDACTED” performed data analysis and drafted the manuscript. “REDACTED” collected the data. “REDACTED” helped to revise the manuscript. All authors have read and approved the final version of the manuscript. SCL, SL, SCNS, JYSC, HCYL, EYMT and KCL conceived the study. SCL, SL, SCNS and SWHK performed data analysis and drafted the manuscript. JYSC, HCYL and EYMT collected the data. SCL, KCL, JWYY and SWHK helped to revise the manuscript.
All authors have agreed on the final version and meet at least one of the following criteria [recommended by the ICMJE (
- substantial contributions to conception and design, acquisition of data or analysis and interpretation of data;
- drafting the article or revising it critically for important intellectual content.
This work was supported by the National Natural Science Foundation of China (NSFC) [Grant No. 61977011]. The funding sources have no role in the study.
CONFLICT OF INTERESTNo conflict of interest to declare.
DATA AVAILABILITY STATEMENTThe data that support the findings of this study are available from the corresponding author upon reasonable request.
ETHICS STATEMENTAll the research meets the ethical guidelines, including adherence to the legal requirements of the study country. Ethical approval was obtained from the Ethical Review Committee regarding Human Research, The Open University of Hong Kong (Ref: HE15Mar2016-URC201601). The study conforms to the recognized standard of the Declaration of Helsinki.
Please respond to the following questions based on your learning experience.
Note: “Teachers or students” in the following refers to people who come into contact during the learning process.
No. | Items | Not true at all | Not exactly true | Maybe | Somewhat true | Absolutely true |
1 | I can manage my study time well | 1 | 2 | 3 | 4 | 5 |
2 | When I feel a bit low emotionally during the learning process, I look for ways to regain my interest and enjoyment in learning | 1 | 2 | 3 | 4 | 5 |
3 | I always reflect on what I have learned and how I have grown during a learning process | 1 | 2 | 3 | 4 | 5 |
4 | At the beginning of a learning process, I always make a reasonable study plan | 1 | 2 | 3 | 4 | 5 |
5 | I try to follow my study plans with regular reviews on the progress | 1 | 2 | 3 | 4 | 5 |
6 | When learning does not go well, I reflect repeatedly on my learning targets and strategies to see if any adjustments need to be made | 1 | 2 | 3 | 4 | 5 |
7 | I always deduce conclusions about effective learning strategies | 1 | 2 | 3 | 4 | 5 |
8 | I regularly review my learning outcomes and analyse my learning problems | 1 | 2 | 3 | 4 | 5 |
9 | I always make use of memorization strategies to study during the course and review before exams (e.g. learning by rote, using images and mind maps) | 1 | 2 | 3 | 4 | 5 |
10 | In the learning process, I always try to connect what I have just learned with my existing knowledge | 1 | 2 | 3 | 4 | 5 |
11 | In the learning process, I try to find real-life examples to enhance my understanding of important concepts | 1 | 2 | 3 | 4 | 5 |
12 | In the learning process, I try to summarize the contents in my own words | 1 | 2 | 3 | 4 | 5 |
13 | I try to arrange for a comfortable environment for my studying | 1 | 2 | 3 | 4 | 5 |
14 | I always feel curious about the course contents that I'm going to learn. | 1 | 2 | 3 | 4 | 5 |
15 | I always look forward to the upcoming course activities | 1 | 2 | 3 | 4 | 5 |
16 | I am not too interested in the course contents* | 1 | 2 | 3 | 4 | 5 |
17 | I always feel bored by the course contents during the learning process* | 1 | 2 | 3 | 4 | 5 |
18 | I like learning through online platforms (e.g. Online Learning Environment) | 1 | 2 | 3 | 4 | 5 |
19 | I feel bored while doing the assignments* | 1 | 2 | 3 | 4 | 5 |
20 | I reserve enough time to complete the course learning tasks | 1 | 2 | 3 | 4 | 5 |
21 | Even when not required, I participate in activities that might be useful for the course (e.g. self quizzes, talks and peer discussions) | 1 | 2 | 3 | 4 | 5 |
22 | I still work hard to make sure I have enough time for my studies even during stressful times of my work and life | 1 | 2 | 3 | 4 | 5 |
23 | I share my views and resources with my teachers or students | 1 | 2 | 3 | 4 | 5 |
24 | I actively respond to questions and calls for help from teachers or students | 1 | 2 | 3 | 4 | 5 |
25 | I always discuss extra-curricular matters with my teachers or students | 1 | 2 | 3 | 4 | 5 |
26 | I participate actively in group learning activities (e.g. group discussion) | 1 | 2 | 3 | 4 | 5 |
27 | I feel happy when taking part in learning activities | 1 | 2 | 3 | 4 | 5 |
28 | I feel happy when sharing ideas with my classmates | 1 | 2 | 3 | 4 | 5 |
29 | I am willing to take part in student activities organized by the school or student organizations | 1 | 2 | 3 | 4 | 5 |
*Negatively keyed items. © “REDACTED”, reproduced with permission of the copyright owners. Authors retain the copyright of the GSES, and reproduction of GSES is available with authors' permission only.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2023. This work is published under http://creativecommons.org/licenses/by-nc/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Aim
Student engagement is an important factor to the success of higher education. This study aimed to develop a Generic Student Engagement Scale (GSES) for face-to-face and online learning.
Design
This was a cross-sectional psychometric study.
Methods
We tested the psychometric properties of GSES in 451 students at the school of nursing and health studies undertaking online and face-to-face learning at a local university in Hong Kong between 2016 and 2018.
Results
Content validity, face validity and test–retest reliability of GSES were satisfactory. The 29-item GSES contains five factors “self-regulated learning,” “cognitive strategy use,” “experienced emotion,” “teacher–student interaction,” and “enjoyment of school life” with the good model fit. The GSES is a reliable and valid psychometric instrument to measure student engagement in face-to-face and online learning among undergraduates and higher diploma students. Our results implied that student engagement can be assessed in routine or research by using our instrument.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details







1 School of Educational Technology, Faculty of Education, Beijing Normal University, Beijing, China
2 School of Nursing, The Hong Kong Polytechnic University, Hong Kong SAR, China
3 School of Arts and Social Sciences, Hong Kong Metropolitan University, Hong Kong SAR, China
4 School of Nursing, Tung Wah College, Hong Kong SAR, China
5 School of Nursing, The University of Hong Kong, Hong Kong SAR, China
6 Office for Advancement of Learning and Teaching, Hong Kong Metropolitan University, Hong Kong SAR, China
7 School of Open Learning, Hong Kong Metropolitan University, Hong Kong SAR, China