| Article ID | Journal | Published Year | Pages | File Type |
|---|---|---|---|---|
| 4938821 | The Internet and Higher Education | 2017 | 38 Pages |
Abstract
This study addresses overload and chaos in MOOC discussion forums by developing a model to categorize threads based on whether or not they are substantially related to course content. A linguistic model was built based on manually coded starting posts in threads from a statistics MOOC, and tested on the second offering of the course, another statistics MOOC, a psychology MOOC, a physiology MOOC, and a test set of reply posts. Results showed that content-related starting posts had distinct linguistic features that appeared unrelated to the domain. The model demonstrated good reliability for all starting posts in statistics and psychology as well as for reply posts (accuracy ranged from 0.80 to 0.85). Reliability for starting posts in physiology was lower but still provided reasonably good predictive ability (accuracy was 0.73). The classification model was useful across all time segments of the courses; the number of views and votes threads received were not helpful.
Keywords
Related Topics
Social Sciences and Humanities
Social Sciences
Education
Authors
Alyssa Friend Wise, Yi Cui, WanQi Jin, Jovita Vytasek,
