Although massive open online courses (MOOCs) have attracted much worldwide
attention, scholars still understand little about the specific elements that students
find engaging in these large open courses. This study offers a new original
contribution by using a machine learning classifier to analyze 24,612 reflective
sentences posted by 5,884 students, who participated in one or more of 18 highly rated
MOOCs. Highly rated MOOCs were sampled because they exemplify good practices or
teaching strategies. We selected highly rated MOOCs from Coursetalk, an open
user-driven aggregator and discovery website that allows students to search and review
various MOOCs. We defined a highly rated MOOC as a free online course that received an
overall five-star course quality rating, and received at least 50 reviews from
different learners within a specific subject area. We described six specific themes
found across the entire data corpus: (a) structure and pace, (b) video, (c) instructor,
(d) content and resources, (e) interaction and support, and (f) assignment and
assessment. The findings of this study provide valuable insight into factors that
students find engaging in large-scale open online courses.
Keyword: MOOCs, massive open online courses, engagement, text mining,
machine learning
Introduction
Online learning allows students to gain access to education despite spatial and
temporal restraints. One of the most significant developments of online learning is the
emergence of massive open online courses, or MOOCs for short. Taking a MOOC is
convenient, flexible and economic for any individual with an Internet connection.
According to the statistics presented by Class Central, a free directory of online
courses that helps users find and track MOOCs, there were more than 6,850 MOOCs being
offered by 700 universities as of December 25, 2016 (Shah, 2016). This data is
presented in Figure 1.
Figure 1. Growth of MOOCs as presented by Shah (2016). From "By the numbers:
MOOCs in 2016," D. Shah, 2016 (https://www.class-central.com/report/mooc-stats-2016/).
In the public domain.
Although the emergence of MOOCs has fueled much attention among researchers and
educators around the world, our understanding of student engagement in these large open
online courses is still limited (Anderson, Huttenlocher, Kleinberg, & Leskovec,
2014). Compared to conventional online courses, the task of engaging students in
large-scale open online learning environments is often more challenging (Hew, 2016). In
conventional online courses, learners usually share the same academic goals, are
familiar with one another, and are supervised closely by the teacher (Chiu & Hew,
2018). However in MOOCs, learners do not know most of their peers, are not supervised
by the teacher, and are under no expectation to complete the course (Chiu & Hew,
2018).
The main purpose of this study was to identify which aspects of MOOCs participants
found engaging, by analyzing a large data set of participant qualitative comments. More
specifically, we set out to offer a new contribution by testing a set of five machine
learning automatic classification models (k-Nearest Neighbors, Gradient Boosting Trees,
Support Vector Machines, Logistic Regression, and Naïve Bayesian). The best
performing model was employed to analyze 24,612 reflective sentences posted by 5,884
students, who participated in one or more of 18 highly rated MOOCs. To the best of our
knowledge, this is the first work that mobilized a repertoire of analytical and
technological resources in the fields of text data mining and machine learning to
analyze a large dataset of MOOC students' reflective comments. The scalable algorithmic
approach in machine learning freed up human labor, and enabled us to analyze large data
corpus at a scale that would be infeasible by human annotations.
Before explaining our use of the machine learning automatic classifier in detail, we
first describe the aspects of a technology-based environment, which may aid in student
engagement, by using the framework of engagement theory (Kearsley & Schneiderman,
1998). Second, in the Literature Review section, we provide a brief review of previous
MOOC research, followed by a discussion of the current research gaps regarding student
engagement in MOOCs. Third, in the Method section, we explain in detail how we selected
18 highly rated MOOCs, collected participant reflective comments of these MOOCs, and
analyzed the comments. Finally, we present the results, followed by the discussion and
conclusion.
Engagement Theory
Student engagement may take many forms, such as attending classes (behavioral
engagement), asking questions (cognitive engagement), and/or expressing enjoyment
towards the course activities or instructors (emotional engagement; Fredricks,
Blumenfeld, & Paris, 2014).
One frequently cited theory that serves as a useful conceptual framework to
understand teaching and learning in a technology-based environment is Engagement Theory
(Kearsley & Shneiderman, 1998). Engagement theory posits three primary elements to
accomplish student engagement: (a) Relating, (b) Creating, and (c) Donating. The role
of technology in this theory is to help facilitate engagement in ways that may be
difficult to achieve otherwise (Kearsley & Shneiderman, 1998).
The first element, Relating, emphasizes peer interaction whereby students
exchange ideas or opinions with other students, enabling learners from different
backgrounds to learn from one another (Kearsley & Shneiderman, 1998). The second
element, Creating, refers to the "application of ideas to a specific context"
(Kearsley & Shneiderman, 1998, p. 20), such as students discussing a case study on
a wiki (Hazari, North, & Moreland, 2009). The third element, Donating,
refers to the use of authentic learning environment that has strong connections to the
real world (Kearsley & Shneiderman, 1998). This principle is particularly valuable
for adult learners, who expect immediate application of knowledge learned in class. By
accomplishing authentic tasks, students can transfer in-class content and see the
immediate implementation of this knowledge (Kearsley & Shneiderman, 1998). The
authenticity of a task can boost students' satisfaction and motivation (Keller, 1987).
Since its inception, engagement theory has been referred to in a variety of
conventional online education contexts (Beldarrain, 2006; Bonk & Wisher, 2000;
Hazari et al., 2009; Knowlton, 2000; Sims 2003). However, hitherto, engagement theory
has not been used to analyze how MOOCs engage participants. As within a conventional
e-learning course, learning in MOOCs also happens online. However, as previously
explained, MOOCs and conventional e-learning courses are dissimilar in their nature.
MOOCs are characterized by free access, and massive open participation. Students can
choose to enroll or drop out of MOOCs at any time they wish without incurring any
penalty. Do the elements espoused in Engagement theory also apply to MOOC-specific
contexts? Which element (i.e., relate, create, donate), if any, is considered most
engaging by MOOC students? What additional elements are considered engaging to MOOC
students? Answers to these questions can help enrich our understanding of MOOC
engagement as well as extend our perspective of Engagement Theory.
Literature Review
In this section, we provide a brief review of previous MOOC studies. This is
followed by a discussion of the current knowledge gaps pertaining to student engagement
in MOOCs.
Currently, most previous research studies on MOOCs can be parsimoniously grouped
into five major categories: (a) impact of MOOCs on institutions, (b) student motives
for signing up for MOOCs and reasons for dropping out, (c) instructor motives and
challenges of teaching MOOCs, (d) click-stream analysis of log data, and (e) types of
MOOCs. Each of these categories will be briefly discussed in the following
paragraphs.
The advent of MOOCs has caused concerns to many universities and libraries. Gore
(2014), for example, examined the new challenges faced by librarians in light of MOOCs,
and discussed a number of challenges that librarians may face as MOOCs become more
widespread. These challenges include licensing and copyright, and delivering remote
services. Lombardi (2013) examined the types of decisions undertaken by Duke
University, an institution which attempted to capitalize on the opportunities and
challenges presented by MOOCs.. Decisions revolved around questions such as how well
does partnership with Coursera (a MOOC platform provider) align with the University's
academic goals, and how might this partnership promote a sustainable model to advance
open education as a social good?
Hew and Cheung (2014) reviewed 25 studies to understand the motivation and
challenges of instructors' and students' use of MOOCs. Their study revealed four main
reasons of student motives for signing up for a MOOC, including the wish to learn
something new, to expand existing knowledge reservoir, to challenge themselves, and to
get a MOOC completion certificate. Reasons for students dropping out include difficulty
in understanding the subject material, insufficient support, and having other
priorities over the course.
Instructors' motives for teaching MOOCs include the desire to enhance their
professional reputation, to provide opportunity for students around the world to access
their courses (Kolowich, 2013). Challenges of teaching MOOCs include lack of student
feedback, lack of online forum participation, and heavy burden of time and effort in
developing and implementing the MOOCs (Hew & Cheung, 2014). In a study by Baxter
and Haycock (2014) regarding MOOC online forum participation, most students only posted
intermittently in the online forum. Those who considered themselves frequent
contributors only accounted for about 6% of 1,000 randomly selected students (Baxter
& Haycock, 2014).
Other studies used click-stream data to investigate student online activities during
MOOCs. For example, previous studies found that more students watched videos than
worked on course assignments, and that the number of student participation deteriorated
steadily as the weeks progressed (Coffrin, de Barba, Corrin, & Kennedy, 2014).
Previous studies also found that frequency of forum postings and quiz attempts
positively correlated with student MOOC grades (Coetzee, Fox, Hearst, & Hartmann,
2014; de Barba, Kennedy, & Ainley, 2016). Other studies attempted to propose models
or methods to predict MOOC dropout (Kloft, Stiehler, Zheng, & Pinkwart, 2014).
Finally, other scholars focused on examining the different types of MOOCs.
Essentially, MOOCs can be parsimoniously classified into either xMOOCs or cMOOCs.
xMOOCs follow a cognitive-behavioral approach (Conole, 2013), while cMOOCs are modeled
after the notion of connectivism (Daniel, 2012; Rodriguez, 2012). xMOOCs typically come
with a syllabus, a course content that consists of readings, discussion forums,
assignments (e.g., quizzes, projects), and pre-recorded instructor lecture videos (Hew
& Cheung, 2014). The syllabus, course content, readings, forums and assignments in
xMOOCs are predefined by the instructors before the commencement of the course (Hew
& Cheung, 2014). In cMOOCs, however, students define the actual course contents as
the course progresses, and there is no fixed syllabus (Rodriguez, 2012). Participants
organize their own learning according to different learning goals, and to interact with
others, while an emphasis is placed on personalized learning through a personal
learning environment (Conole, 2013; Rodriguez, 2012). This could result in more than
one topic being examined concurrently (Hew & Cheung, 2014).
Research Gaps
Although the aforementioned studies have provided us with a useful understanding of
MOOCs, they fall short of explaining the reasons why participants find a course or
certain parts of a course engaging. Many researchers have begun questioning the
validity of using traditional metric such as completion or dropout rate to measure
whether a MOOC is engaging or not. As previously described, MOOCs are open courses that
are usually offered free of charge, and learners who sign up are under no obligation
whatsoever to complete the course. Due to time constraints and work commitment,
learners may only complete certain course activities (Kizilcec, Piech, & Schneider,
2013) and still find the activities engaging.
Therefore, in order to understand which aspects of MOOCs students find engaging, we
need to analyze students' reflective comments about the MOOCs. So far to the best of
our knowledge, only two studies were found that focused specifically on student
reflection data of MOOCs. Hew (2015) analyzed guidelines of improving the quality of
online teaching and learning from four professional councils, as well as qualitatively
analyzed 839 participants' comments of two highly rated MOOCs using the grounded
approach. By synthesizing the policy guidelines and actual opinions of online learners,
Hew (2015) proposed a rudimentary model of engaging online students, covering six
dimensions: course information, course resources, active learning, interaction,
monitoring of learning, and making meaningful connections. The findings of this study
show us an emerging picture of what is valued by MOOC students.
Another grounded approach study analyzed 965 course participants' reviews on three
top-rated MOOCs in the subjects of literature, arts and design, and programming
language to find out what factors about the course engaged students, and contributed to
their favorable consideration of the online learning experience (Hew, 2016). Five
factors were listed: problem-centric learning, instructor accessibility and passion,
active learning, peer interaction, and helpful course resources. This study set a
foundation of what are worthwhile factors to consider when instructors prepare and
deliver a MOOC.
The limitation of these two study lies in the fact that only two or three MOOCs were
inspected respectively, which, as the author noted, is not sufficient to warrant strong
conclusions (Hew, 2016). Nevertheless, the results of these two studies provide a
useful conceptual basis for other researchers to conduct studies to examine students'
reflection comments.
Method
This study aims to answer the following question: What elements pertaining to the
course design or the instructor did students find enjoyable, helpful in learning the
materials, or motivational (motivating them to take part in the activities)? In this
section, we explain how we chose the 18 highly rated MOOCs, collected the MOOC
participants' reflective comments, and analyzed the comments. An overview of the whole
data collection, processing, and analysis procedure is shown in Figure 2.
Figure 2. Overview of the method.
Data Collection
Highly rated MOOCs were sampled because the exemplified good practice or teaching
strategies. We selected "highly rated" MOOCs from Coursetalk, an open
user-driven aggregator and discovery website that allows students to search and review
various MOOCs. We defined a highly rated MOOC as a course that received an overall
five-star course quality rating, and received at least 50 reviews (from different
learners) within a particular subject discipline. Using reviews from many different
participants provided data triangulation, which promotes trustworthiness of the ratings
(Shenton, 2004).
Coursetalk was chosen because it is considered the largest platform
connecting learners to courses (Business Wire, 2014). When this study was conducted, it
listed more than 50,000 courses Visitors to the website can rate a course from three
dimensions: course content, course instructor, and course provider. One to five stars
can be given, with five stars being the highest quality. The website will calculate an
individual user's rating across the three categories, and further aggregate an overall
rating among all participants. Therefore, a course will have a general rating score, as
well as a detailed list of ratings from specific participants. In addition,
participants can write down review comments.
Using data from Coursetalk, we searched all 282 subject disciplinary areas,
and applied the following selection criteria to identify eligible student reviews: free
of charge, rated five-star, and with more than 50 reviews. As of July 20, 2017, 18
highly rated MOOCs were identified. We provide an overview of the 18 MOOCs in Table
1.
One of the authors wrote a web crawler to automatically download all the reviews of
the 18 MOOCs. Given the URLs of the 18 course pages, the crawler was launched at 9:46
p.m. on July 20th, 2017 and obtained a total of 5,884 learners' reflective
posts, with 24,612 sentences generated on or before that day. Of the 5,884 learners,
90.9% completed one or more MOOCs, 8.3% were currently taking a MOOC, and 0.8% were
drop outs. Once the entire data had been downloaded, another researcher randomly
selected 20 of the downloaded reviews and checked them against the original reviews
posted on CourseTalk. This procedure served to establish the reliability of
the data collection process. The percent agreement was 100%.
Table 1
MOOCs Reviewed in This Study
Subject discipline area
MOOC title
University
Ratings and number of reviews
Purpose of the course
1
Computer sciences
An Introduction to Interactive Programming in Python
Rice University
5⋆, 3077 reviews
Develops simple interactive games (e.g., Pong) using Python.
2.
Social sciences
The science of everyday thinking
The University of Queensland
5⋆, 1167 reviews
Explores the psychology of our everyday thinking, such as why
people believe weird things, and how we can make better decisions.
3.
Natural sciences
The science of the solar system
Caltech
5⋆, 303 reviews
Explores Mars, the outer solar system, planets outside our solar
system, and habitability in our neighborhood and beyond.
4.
Environmental sciences
Introduction to environmental science
Dartmouth College
5⋆, 233 reviews
Surveys environmental science topics at an introductory level,
ultimately considering the sustainability of human activities on the planet.
5.
Design
Design: creation of artifacts in society
University of Pennsylvania
5⋆, 217 reviews
Focuses on the basic design process: define, explore, select, and
refine. Weekly design challenges test student ability to apply those ideas to solve
real problems.
6.
Literary art
Modern and contemporary American poetry
University of Pennsylvania
5⋆, 171 reviews
Introduces modern and contemporary U.S. poetry, with an emphasis
on experimental verse, from Dickinson and Whitman to the present.
7.
Cybersecurity
Cybersecurity fundamentals
Rochester Institute of Technology
5⋆, 168 reviews
Introduces essential techniques in protecting systems and network
infrastructures, analyzing and monitoring potential threats and attacks, devising
and implementing security solutions.
8.
Statistics
The analytics edge
Massachusetts Institute of Technology
5⋆, 149 reviews
Explores the use of data and analytics to improve a business or
industry.
9.
Management
Finance: time value of money
n.a.
5⋆, 90 reviews
Introduces the logic of investment decisions and familiarize with
compounding, discounting, net present value, and timeliness.
10.
Computer sciences
HTML5 coding essentials and best practices
The World Wide Web Consortium (W3C)
5⋆, 84 reviews
Explains the new HTML5 features to help create great Web sites and
applications in a simplified but powerful way.
11
Management
u.lab: leading from the emerging future
Massachusetts Institute of Technology
5⋆, 82 reviews
Introduces a method called Theory U, developed at MIT, for leading
changes in business, government, and civil society contexts worldwide.
12.
Art and culture
Drawing nature, science and culture: natural history illustration
101
The University of Newcastle, Australia
5⋆, 81 reviews
Introduces essential skills and techniques that form the base for
creating accurate and stunning replications of subjects from the natural
world.
13
Biology
Introduction to biology-the secret of life
Massachusetts Institute of Technology
5⋆, 75 reviews
Explores the mysteries of biochemistry, genetics, molecular
biology, recombinant DNA technology and genomics, and rational medicine.
14.
Literary art
Comic books and graphic novels
University of Colorado Boulder
5⋆, 75 reviews
Presents a survey of the Anglo-American comic book canon and of
the major graphic novels in circulation in the United States today.
15.
Social sciences
Justice
Harvard University
5⋆, 68 reviews
Explores critical analysis of classical and contemporary theories
of justice, including discussion of present-day applications.
16.
Computer sciences
Mobile computing with app inventor-CS principles
Trinity College
5⋆, 51 reviews
Explains the open development tool, App Inventor, to
program on Android devices, and some of the fundamental principles of computer
science.
17
Social sciences
International human rights law
Université catholique de Louvain
5⋆, 51 reviews
Examines a wide range of topics including, religious freedom in
multicultural societies, human rights in employment relationships, etc.
18
Environment
Climate change: the science
University of British Columbia
5⋆, 50 reviews
Introduces climate science basics such as flows of energy and
carbon in Earth's climate system, how climate models work, climate history, and
future forecasts.
Data Processing
To train the machine learning algorithm, we first developed a coding scheme
consisting of the following six themes based on previous grounded approach studies that
specifically analyzed MOOC students' engagement using participant review comments (Hew,
2015; 2016). These themes, which are not listed in order of importance or priority
included:
Theme 1: Structure and pace,
Theme 2: video,
Theme 3: instructor attributes,
Theme 4: content and resources,
Theme 5: interaction or community, and
Theme 6: assignment and assessment.
Table 2
Themes and Descriptors
Theme
Descriptors
Structure and Pace
Clear objective, duration, structure and syllabus
Video
Videos, captions, choice of speed variation, video-integrated
quiz
Instructor attributes
Instructor knowledge, instructor passion, instructor humor
Content and resources
Examples or case studies that relate to the real world,
problem-solving centric, relevant and up-to-date course content and resources,
Availability of transcript and pdf documentation, slide notes
Interaction and support
Student-student interaction, instructor-student interaction,
course support
Assignment and assessment
Use of active learning strategies such as mini-projects,
exercises, quizzes, questions, feedback
To fulfill the training purpose, an annotated "instructional" dataset is required to
form the training materials, consisting of positive and negative cases, on which the
automatic machine classifiers "learn" through adapting their model parameters to fit
the dataset. A sample of sentences (259) were randomly extracted from the newly
collected dataset, and human raters were recruited to manually label these texts using
the aforementioned theme labels. One human rater independently labeled the texts using
the labels. This was then independently examined by another human rater. Discrepancies
among the raters were resolved through discussion.
For example, 25 positive cases of Theme 6 "Assignment and Assessment" were annotated
by the human rater, and they were used to train the automatic classifiers. In the
training process, the computer might automatically learn that positive cases of this
theme may contain certain linguistic cues such as: "(+) quiz" or "(+) task." When the
machine analyzed the data independently afterwards, if a case contained "(+) quiz" or
"(+) task", most likely the computer would classify it into "Assignment and
Assessment". For example, the comment "the quizzes are detailed and complete and
require a bit of extra programming and the mini-project and peer reviews require a few
hours of extra effort" was classified as positive within this theme. In the meanwhile,
other cases without these cues were possibly considered negative.
To simplify the problem in our scenario, we leveraged the one-versus-the-rest
treatment (Bishop, 2006, p. 182) for multi-label classification task, and decomposed
the problem into six binary classification tasks, each of which corresponds to a
specific theme. To explain, using the one-versus-the-rest treatment, we trained six
independent classifiers for the six themes. One classifier could only decide whether a
case was positive or negative under the corresponding theme. In other words, it only
focused on one theme and did not consider other themes. The six classifiers were
adopted independently to process the same data corpus. Note that in our case, the theme
labels were not exclusive, hence we were free from the issue of ambiguous
classification regions proposed in Bishop (2006, pp. 182-183).
However, regarding a single theme, the fact that all its non-positively-labeled
cases were taken as negative cases greatly enlarged the proportion of
negative-to-positive data ratios, resulting in data unbalance. To counteract the data
unbalance, we conducted data expansion for positive samples in each theme class.
Specifically, we extracted all sentences in the rest of the corpus containing the
theme-dependent cue terms in a term list (summarized by the annotators). Immediately
after that, the annotators checked the new data and removed wrong samples. The sampling
and validation procedure repeated until the new cases were all valid. The statistics
for the dataset for each theme before and after data expansion are summarized in Table
3. As a result of the data expansion procedure, the sub corpus for each theme was
balanced.
Table 3
Annotated Corpus Statistics Before and After Data Expansion
Theme 1
Theme 2
Theme 3
Theme 4
Theme 5
Theme 6
Positive samples before expansion
37
28
68
78
36
25
Negative samples before expansion
222
231
191
181
223
234
Corpus size after expansion
444
462
382
362
446
468
Data Analysis and Testing of Models
The machine learning goal in our scenario can be formalized as follows. The
objective is to fit a series of binary classifiers
f=[f1, ..., fN]
(where fi corresponds to the ith theme, and N =6
is the number of themes) on the annotated dataset D, and then apply
f to predict theme labels on the rest texts. Given a text
x and the trained classifiers f ,
the themes of x are predicted as y
= f(x) =
[f1(x), ...,
fN(x)], where y
∈ {1,0}N is a N-dimensional vector of zeros and ones
with the ith dimension indicating whether the ith theme is assigned
to the text. Values 1 and 0 in y designate true and false
respectively.
As the data representation building block, the design matrix (Murphy, 2012, p. 2) in
our context is implemented with TF-IDF features, i.e., a matrix of which columns and
rows represent terms and documents respectively, and each cell in the matrix records a
value computed on the frequency of the term in the document (measuring its degree of
popularity within the document), weighted by the term's inversed document frequency
(measuring its degree of rarity in the whole corpus) (Wu, Luk, Wong, & Kwok, 2008).
TF-IDF is a common technique of text representation and can filter out stop words and
keep the most discriminant terms (Robertson, 2004).
On the other hand, due to what has been claimed in the no free lunch
theorem (Box & Draper, 1987, p. 424), we could not guarantee a best performed
classification model type beforehand, since there is not a best model which can
outperform all the other models on all problems. As a result, we prepared a set of
candidate model types and expected to find the best performing one. The candidate
classifiers included:
1. k-Nearest Neighbors (KNN) (Altman, 1992),
2. Gradient Boosting Trees (GBT) (Friedman, 2001),
3. Support Vector Machines (SVM) (Cortes & Vapnik, 1995),
4. Logistic Regression (LR) (Cox, 1958), and
5. Naïve Bayesian (NB) (Murphy, 2012).
To evaluate the classification performance in a comprehensive stance, we adopted
five metrics. In these metrics, the first four measurements are calculated on the
confusion matrix (Stehman, 1997), which records predicted labels and ground truths and
is primarily for measuring correctness from different angles. The kappa value (Cohen,
1960) measures the classification consistency between the learning classifiers and the
human annotators:
1. Accuracy: the proportion of correct predictions (both positive and negative) in
all cases;
2. Precision: the proportion of correct predictions in all cases predicted as
positive.
3. Recall: the proportion of true positive cases predicted as positive in all true
positive cases;
4. F1: the harmonic mean of precision and recall, which balances the measurements of
2) and 3)
5. Cohen's kappa: the agreement between two raters (the machine trained classifier
and human annotators in our scenario (Cohen, 1960).
We implemented the machine learning and evaluation experiment using Python
programming language. The classifiers were implemented with the Scikit-learn
packagei, and the texts were segmented, tokenized and cleaned with the
spaCyii natural language processing toolkit, before the design matrix was
constructed via the Scikit-learn TF-IDF feature extraction tool.
The experiment was conducted on an Ubuntu16.04 system equipped with Intel Core
i5-4460 3.20GHz CPU and 16GB memory. We arranged five-fold cross-validation (Bishop,
2006, p. 33) for each classifier on each theme, so that each classifier on each theme
obtained five test scores on each metric. Finally, the metric scores were averaged, and
the best performed classifiers on each metric and each theme identified (Table 4).
Table 4
Best Performing Automatic Machine Learning Classifiers on Each Theme
Theme 1
Theme 2
Theme 3
Theme 4
Theme 5
Theme 6
accuracy
GBT
GBT
GBT
GBT
GBT
SVM
precision
NB
SVM
LR
LR
LR
NB
recall
KNN
KNN
KNN
KNN
KNN
KNN
f1
GBT
GBT
GBT
GBT
GBT
SVM
kappa
GBT
GBT
GBT
GBT
GBT
GBT
In general, the candidate classifiers performed differently in different metrics and
for different themes. Whereas, the GBT classifier performed better than all the other
classifiers on kappa metrics, demonstrating its superb consistency with the human
annotators. The same classifier was also good in f1 scores, meaning that it was
relatively balanced in precision and recall. Finally, GBT was very accurate in making
positive and negative predictions, with the accuracy score outperforming most of the
other classifiers except for Theme 6. Due to its stable and superb performance, we
selected GBT as the prediction model for later use. Table 5 presents the performances
of GBT.
The GBT achieved sound values in accuracy, f1, precision and recall. In addition,
its kappa values were all above 0.61, indicating its substantial agreement with human
annotators (McHugh, 2012). All the metrics demonstrate the promising quality of the
classifier.
Table 5
Performance Metrics for GBT
Theme 1
Theme 2
Theme 3
Theme 4
Theme 5
Theme 6
accuracy
0.8079
0.9630
0.8289
0.8056
0.8467
0.8991
precision
0.7868
0.9820
0.8773
0.8625
0.8967
0.9148
recall
0.8464
0.9435
0.7684
0.7333
0.7836
0.8845
f1
0.8152
0.9623
0.8171
0.7894
0.8361
0.8980
kappa
0.6158
0.9261
0.6579
0.6111
0.6934
0.7982
Results
Full Dataset Prediction and Analysis
Adopting the GBT classification model, we trained six GBT classifiers for each of
the six themes, and used them to classify the remaining texts according to the
appropriate theme categories. Figure 3 shows how frequently each theme was found in the
participants' reflective sentences; the more frequent a certain theme was mentioned in
the participants' reflective comments, the bigger the circle size of the theme. Figure
3 also indicates how likely each theme co-occurred with other theme (mentioned together
by the participants); the closer the distance between the themes are shown in Figure 3,
the more likely they are mentioned together by the participants. The distances between
the themes shown in Figure 3 were not designed a priori to the analysis. To explain how
the distance between themes was computed, we provide the following illustration.
Suppose we have the following text: "The exercises are related to the videos and
allow the student to progress week after week". This text was inferred correctly by the
automatic classifiers to contain Theme 2 (Video) and Theme 6 (Assignment and
assessment). The results were then represented with a vector with six dimensions (6-D)
indicating the existence or nonexistence of six themes. In this case, the example text
would be marked with a 6-D vector (0, 1, 0, 0, 0, 1), representing Theme 1 (leftmost)
to Theme 6 (rightmost), where each position represents whether the theme is on (using
1) or off (using 0). In our example, the 2nd and 6th dimensions
of the vector were turned on (using 1), indicating the text contained Theme 2 and Theme
6. The other themes (i.e., Themes 1, 3, 4, 5) were turned off (using 0). Now assuming
we have the following results from Texts 1 to 4 (Table 6):
Table 6
Examples of Texts and Themes
Text
Theme1
Theme2
Theme3
Theme4
Theme5
Theme6
Text 1
1
1
1
0
0
0
Text 2
0
1
0
0
0
0
Text 3
0
1
1
1
0
0
Text 4
1
1
1
0
0
0
We find that Themes 2 and 3 are more likely to co-occur (they co-occurred in Texts
1, 3, 4), so we could consider Themes 2 and 3 as more inclined to be semantically
closer than the other themes. Moreover, we can find that the closeness could be
measured by the similarity of the columns of the themes in the above matrix. The Theme
2 column, which is (1,1,1,1) is only one bit different from Theme 3 column, i.e.,
(1,0,1,1).
Below we give an example showing how we can calculate theme distances via their
column vectors. We denote a, b to be any
column vectors of two themes in Table 6, then using the formula of Euclidean distance
metric (which can be used to compute the distance of two vectors), the distance of
a, b can be computed as:
where are the ith elements of the column vectors
a and b respectively, is the square root function, and N is
the dimension of the column vectors (N = 4 in our example, since we have only
4 texts). We can compute the distance between Theme 2 and Theme 3, and other themes,
using the above formula, and obtain for example:
Therefore, we can see that the vector of Theme 2 is closer to Theme 3 than to Themes
1 and 5. So when two themes are more likely to co-occur (mentioned together in the
participants' comments), their column vectors are closer.
To make it possible to visualize the high-dimensional theme vectors, we conducted
Principle Component Analysis (PCA) (Wold, Esbensen & Geladi, 1987) for the matrix
and managed to reduce and project the column vectors into two-dimensional space, as
presented in Figure 3. Note that PCA not only projects the vectors into the 2-D space,
but also maximally keeps the vector closeness information in the original high
dimensional space. With the 2-D coordinates yielded by PCA, we are able to draw the
themes on x-y axis and generate Figure 3.
Figure 3. Visualization of the relatedness of the themes. Circle size is
proportional to the percentage of the theme in the corpus, and the distances between
circles indicate their relatedness in terms of co-occurrence.
It is beyond the length of this paper to list out every single student comment
related to each of the seven themes. We therefore provide some representative examples
to provide the reader with a closer look at each of the six themes (see Table 7). The
main findings can be summarized as follows:
1. The most frequently mentioned theme was instructor attributes. The other commonly
mentioned themes were course content and resources, assignment and assessment, and
structure and pace.
2. Two particular instructor attributes stood out among the many student comments:
instructor's passion about the subject as well as teaching it, and instructor's sense
of humor.
3. Students enjoy course content and resources that emphasized real-world
application or problem-solving.
4. Students desire moderately challenging courses assignments that require them to
apply the contents learned. Easy assignments or questions that merely test factual
recall are disliked. Assignments that are fun and enjoyable (e.g., building simple
games) make the tasks more engaging to students.
5. Student prefer short lecture videos ranging from about five to ten minutes long.
Videos consisting of different instructors were perceived to be more engaging. Guest
speakers' appearances in videos are generally welcomed and appreciated by
students.
6. Use of in-video quizzes helps sustain students' attention to lecture
content.
7. Students prefer a course structure that builds on each lesson progressively from
simple to difficult, and provides options for students to do either the minimum or go
deeper into the subject.
Table 7
Themes and Examples of Students' Comments
Discussion
It is interesting to note that all 18 MOOCs can be classified as xMOOCs. To briefly
recall, xMOOCs typically come with a syllabus, a course content that consists of
readings, discussion forums, assignments (e.g., quizzes, projects), and pre-recorded
instructor lecture videos (Hew & Cheung, 2014). The syllabus, course content,
readings, forums and assignments in xMOOCs are predefined by the instructors before the
commencement of the course (Hew & Cheung, 2014). The results of the present study
demonstrate that xMOOCs, which come with a clear instructor-defined course structure,
are perceived more positively than cMOOCs by students. Students appreciate a course
structure with a well thought progression of pace from easy to more difficult as the
weeks advance.
Overall, the most frequently mentioned themes that participants perceived as
engaging were instructor attributes. The most common means for an instructor to present
course materials in a MOOC is through lecture videos (Young, 2013). Even though
instructors may feel awkward being on videos, they can still engage students if they
are excited or enthusiastic about the subject matter (Young, 2013). An instructor's
enthusiasm for teaching the subject can help break the boredom of watching videos, and
even motivate students to complete the activities, as indicated in student feedback:
"His [the instructor] enthusiasm is contagious, and it helped to motivate me to
complete the quizzes." (Student)
As indicated in student feedback, students also value an instructor's humor because
it helps break the boredom and make the lesson more enjoyable, as shown in the
following student comments. Use of humor can help arouse students' attention, increase
student liking of the professor, establish a positive rapport with students, and
motivate students to participate in the course (Wanzer, 2002): "Their geek sense of
humor helped the classes be very interesting and enjoyable" (Student) and "I also liked
the added humor in the lectures to avoid the material from becoming dull"
(Student).
Course content and resources that emphasized real-world application or
problem-solving were other themes that participants perceived as engaging. As outlined
by Dillahunt, Wang, and Teasley (2014), a majority of MOOC learners are adult learners
who have at least a bachelor's degree and are employed. Adult learners will be more
engaged in learning when new content that is presented is applicable to real-life
situations (Knowles, Holton, & Swanson, 2011). This finding implies that
instructors should emphasize application of content to real-world practices over mere
transmission of information. Instructional strategies such as showing real-life
problems to which the principles or solutions taught in the course can be applied, and
presenting practical tips are particularly useful because these elements provide
valuable add-on insights to students' learning. In addition, instructors should provide
text-based resources such as video transcripts, video captions, and pdf documentations
to help students review the course content.
Interestingly, despite the commitment-free nature of MOOCs (e.g., no actual course
credit or course fees), students still desire moderately challenging courses
assignments that require them to think or apply the concepts or principles learned. One
possible explanation for this may be offered by the achievement goal theory. According
to achievement goal theorists, there are two main types of goals: (1) the performance
goal, which focuses on exhibiting ability in comparison with other people; and (2) the
mastery goal, which focuses on developing competence in a particular topic or area
(Ames, 1992; Meece, Blumenfeld, & Hoyle, 1988). Since learners in a MOOC do not
know most of their peers, it is unlikely that they are motivated by performance goals.
We posit therefore that learners in MOOCs are more likely to be motivated by mastery
goals. Adopting a mastery goal is believed to produce a desire for moderately
challenging tasks, a positive stance toward learning, and enhanced task enjoyment
(Elliot & Church, 1997). The implication here is that instructors should avoid
simple assignments that merely test factual recall. Instead, instructors should employ
strategies such as asking students to apply the concepts learned to solve some
real-world problems. The activities should also have varying levels of difficulty so
that students can choose an activity that matches their personal ability, while
simultaneously providing them with an opportunity to accomplish more difficult tasks in
order to master a particular topic or skill.
Other themes that participants perceived as engaging included interaction and
support, and video lectures. As reported by participants, interaction and support from
the tutors (instructors and/or teaching assistants) helped foster cognitive engagement,
which can assist student learning of the topic. Not all the 18 MOOCs have the same
degree of interaction and support. MOOCs that had relatively more participant comments
about interaction and support such as the Interactive Python Programming and American
Poetry courses used one or more of the following strategies:
a. Providing an opportunity for students to interact with peer raters regarding
their submitted assignments. For example, the instructor of the American Poetry MOOC
provided an opportunity for students to interact with peer reviewers (e.g., seek
clarification) regarding their submitted assignments. Students in the Interactive
Python Programming MOOC could discuss the quiz problems with other students with the
only rule being that they could not post explicit answers in the forum (Warren,
Rixner, Greiner, & Wong, 2014).
b. Organizing live online sessions for instructor and fellow students to exchange
answers and ideas. For example, observation of the American Poetry MOOC revealed that
the teaching staff of the said MOOC offered live webcasts every week for students
around the world to join the live discussions through the telephone, Facebook, and
course forums.
c. Hold weekly one-hour virtual office hours. For example, the teaching staff of the
American Poetry MOOC organized one-hour virtual office hours every week in the
discussion forum to answer student questions. The Python MOOC used a professional
help desk service (http://helpscout.net) to manage student email inquiries. Specifically,
this desk service routed student help requests to a course website called "Code
Clinic" so that the teaching staff could respond to them (Warren et al., 2014). The
professional help desk service provides several useful features such as collision
detection to prevent duplicate replies, an auto reply to let students know their
email request has been received, auto tracking of student requests, and customizable
stock responses to common questions (Warren et al., 2014; http://helpscout.net). These features
helped the instructors managed the students' email inquiries (Warren et al.,
2014).
A variety of video production styles were used in the 18 MOOCs. These video
production styles may be categorized under one or a combination of the following labels
(Guo, Kim, & Rubin, 2014, p. 44):
a. Slides: PowerPoint slide presentation with voice-over.
b. Code: video screencast of the instructor writing code in a text-editor.
c. Khan-style: video of instructor drawing freehand on a digital tablet.
d. Classroom: video captured from a live classroom lecture.
e. Studio: instructor recorded in a studio with no audience.
f. Office desk: close-up shots of instructor's head filmed at an office desk.
Despite the different video production styles, the following three findings should
be noted:
a. First, students prefer short videos. Findings from a study that analyzed 6.9
million video watching sessions across four MOOCs revealed that the median video
watching time was six minutes (Guo et al., 2014). This therefore suggests that
instructors should segment videos into short chunks, shorter than six minutes (Guo et
al., 2014).
b. Second, to help minimize student mind wandering during video watching instructors
should embed short-recall quiz into the video. For example, the instructors of the
Interactive Python Programming and Science of the Solar System MOOCs interrupted the
videos with in-video quizzes. Several psychological research studies (Szpunar, Khan,
& Schacter, 2013; Szpunar, Jing, & Schacter, 2014) have found that
participants who answered short-recall questions after each short video segment
(interpolated tests), retained more information at the end of the lecture as compared
to the participants who did not receive any recall questions, or the those who merely
watched the questions-with-given-answers. Compared to participants in the control
group who were not provided interpolated tests, mind wandering occurred less (half as
much) for participants who answered the interpolated tests (Szpunar et al.,
2013).
c. Third, lecture videos that featured different instructors' faces (e.g.,
appearances of guest speakers) were perceived to be more engaging because it helped
break the boredom of watching the same instructor throughout all the sessions.
Conclusion
Despite the worldwide attention attributed to MOOCs, scholars still understand
little about student engagement in these large open online courses. This study offers a
new original contribution by analyzing the reflective comments posted by 5,884 students
who participated in one or more of 18 highly rated MOOCs in order to identify the
reasons why participants find a MOOC or certain parts of a MOOC engaging. These 18
highly rated MOOCs were chosen from a pool of 282 subject disciplinary areas, having
successfully fulfilled the following selection criteria: free-of-charge, rated
five-star, and with more than 50 reviews. In this section, we discuss several
implications for distance education theory, research, and practice. We conclude by
describing the limitations of the present study.
First, the theoretical contribution of this paper lies in its examination of the
elements espoused by Engagement Theory, as well as extending our current perspective of
Engagement Theory in the context of large-scale fully online courses such as MOOCs that
have no requirement for face-to-face attendance. We found that the elements of
Creating, and Donating in Engagement Theory (which refer to the application of ideas,
and to the use of real-world contexts respectively) (Kearsley & Shneiderman, 1998)
were two of the themes found in the MOOCs participants' reflective comments. Many
participants of the 18 highly rated MOOCs reported that the use of moderately
challenging assignments that require them to apply the concepts or principles learned,
instead of merely asking them to recall factual information, helped students learn the
subject material better. Participants also reported that the use of content and
resources focusing on real-world examples or problems made the course material very
relevant. This helped bring tangible meaning to the concepts or principles taught,
which sustained students' interest, and enabled them to learn the material more easily
because they could see how the principles or theories learned might be applied in
real-life.
Contrary to expectation, the Engagement Theory element of Relating, which emphasizes
peer interaction (Kearsley & Shneiderman, 1998), was one of the least mentioned
themes found in the MOOCs participants' reflective comments. This implies that MOOC
students do not seem to attach much importance with respect to the need for peer
interaction in large-scale open online courses when compared to traditional online or
face-to-face classes. It is likely that the anonymous nature of MOOCs, along with job
or family responsibilities diminishes student expectations of course interaction with
their peers.
The present findings suggested that MOOC student engagement is promoted when certain
instructor attributes are present, namely the instructor's ability to show enthusiasm
when talking about the subject material, and the instructor's ability to use humor.
These instructor attributes, which formed the most frequently mentioned theme perceived
as engaging by MOOC participants, extend our current perspective of Engagement Theory
in the context of large-scale fully online courses, which rely primarily on an
instructor presenting the subject materials through videos. Although the inclusion of
an instructor's face can give a more personal and intimate feel to the video lecture
(Kizilcec et al., 2014), how an instructor projects himself or herself (e.g., by
showing interest in teaching the material) seems to play a more important role than
merely putting a face in the video.
Second, this study contributes to distance education research by proposing and
testing five scalable algorithmic models. This is the first work, to our knowledge,
that mobilized a repertoire of analytical and technological resources in the fields of
machine learning and text data mining to analyze a large dataset of MOOC students'
reflective comments. Specifically, we found the Gradient Boosting Tree algorithm
(Friedman, 2001), to be the best performing model. The detail technical procedure
provided in this study will be of great interest to other researchers who are similarly
keen in this type of research methodology.
Third, this study contributes to distance education practice by highlighting several
practical solutions to other instructors of large open online courses, as well as those
teaching traditional e-learning classes. For example, the various strategies to support
instructor-student interactions, and practical tips of using video lectures (as
described in the Discussion section) can offer possible solutions for traditional
e-learning courses that might otherwise be overlooked.
We conclude the present article by highlighting three limitations. First, it should
be noted that highly rated courses may not necessarily be the most effective ones. It
is beyond the scope of this study to examine causal effect between course effectiveness
(e.g., learning performance) and user ratings. Second, this study did not examine the
participants' disaffection of using MOOCs. Exploring student disaffection may offer
information that can complement our overall understanding of student engagement. We
therefore invite other researchers to conduct this investigation. Third,
CourseTalk did not provide any indication on which student comments came from
students taking the MOOC for course credit or from students who were taking it for
other reasons. This precludes an investigation of how the comments from these students
may differ. Despite the aforementioned limitations, we believe that the findings of
this study provide valuable insight on the specific elements that students find
engaging in large-scale open online courses.
Notes
a. http://scikit-learn.org/stable/
b. https://spacy.io/
Acknowledgment
This research was supported by a grant from the Research Grants Council of Hong Kong
(Project reference no: 17651516).
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2018. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and conditions, you may use this content in accordance with the terms of the License.
Abstract
Although massive open online courses (MOOCs) have attracted much worldwide attention, scholars still understand little about the specific elements that students find engaging in these large open courses. This study offers a new original contribution by using a machine learning classifier to analyze 24,612 reflective sentences posted by 5,884 students, who participated in one or more of 18 highly rated MOOCs. Highly rated MOOCs were sampled because they exemplify good practices or teaching strategies. We selected highly rated MOOCs from Coursetalk, an open user-driven aggregator and discovery website that allows students to search and review various MOOCs. We defined a highly rated MOOC as a free online course that received an overall five-star course quality rating, and received at least 50 reviews from different learners within a specific subject area. We described six specific themes found across the entire data corpus: (a) structure and pace, (b) video, (c) instructor, (d) content and resources, (e) interaction and support, and (f) assignment and assessment. The findings of this study provide valuable insight into factors that students find engaging in large-scale open online courses.
Keyword: MOOCs, massive open online courses, engagement, text mining, machine learning
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer