Thesis Sample
Thesis Sample
Thesis Sample
IN ONLINE EDUCATION
by
A DISSERTATION
TUSCALOOSA, ALABAMA
2012
Copyright Jill Mask Simpson 2012
ALL RIGHTS RESERVED
ABSTRACT
The purpose of this study was to examine student perceptions of quality and satisfaction
in regionally accredited online courses. This study intended to answer the following four
research questions (1) Is there a significant difference in levels of student satisfaction between
online courses that have undergone a systematic faculty peer review process and online courses
that have not undergone a systematic faculty peer review process?; (2) Is there a significant
difference in levels of student satisfaction between online courses that have not undergone a
systematic faculty peer review process but are affiliated with a peer review program and online
courses that have no affiliation with a peer review program?; (3) Which factors of quality
instruction most directly relate to increased levels of student satisfaction in online courses that
have been faculty peer reviewed?; and (4) Which factors of quality instruction most directly
relate to increased levels of student satisfaction in online courses that have not been faculty peer
reviewed but are affiliated with a peer review program? A total of 157 responses (out of 1,774
solicited) were obtained from the student satisfaction questionnaire, which is a 9% student
response rate. A total of 7 responses (out of 54 solicited) were obtained from the course designer
questionnaire, which is a 13% course designer response rate. Results of the study confirmed with
statistical significance that students who were more comfortable with distance learning reported
higher satisfaction with their online course. Results of the study also indicated that online
courses that have undergone a formal peer review may lead to higher student satisfaction in the
course. Surprisingly, however, results also indicated that courses that have not undergone a
formal peer review but that are offered at institutions that are subscribed to a peer review
ii
program may lead to lower student satisfaction of the course. Both of these results were non-
significant. The researcher recommends striving for statistical significance in future research by
using a larger sample size. Additionally, the researcher recommends future research regarding
the correlation of student satisfaction to peer review subscriptions, even without a formal peer
review.
iii
ACKNOWLEDGMENTS
I am overwhelmed at the many colleagues, friends, and family who have supported me in
this educational goal. I appreciate Angela Benson, the chairman of this dissertation, for her
encouragement throughout every step of this journey. Her expertise in online education is
invaluable. I would also like to thank all of my committee members, Doug Barrett, Elizabeth
Fisher, Margaret Rice, and Vivian Wright for their contribution and support of this dissertation.
inspiration. I also want to thank my mother-in-law, Brenda Simpson, for going out of her way to
keep me on track in my doctoral studies. I want to thank my sister, Lesa Mask, who gave me the
confidence to succeed. Most of all, I want to thank my husband, Jon Simpson, for his endless
love and support; for without him, I would not be the person I am today. Finally, I want to thank
my son, Bay Simpson, for his patience. Throughout this doctoral journey he has shown a great
deal of tolerance and maturity as schoolwork monopolized a great deal of my time. He has
grown into a special young gentleman, and I am so proud to have him in my life.
iv
CONTENTS
ABSTRACT .................................................................................................................................... ii
ACKNOWLEDGMENTS ............................................................................................................. iv
1 INTRODUCTION .....................................................................................................................1
Theoretical/Conceptual Framework.....................................................................................1
Definition of Terms..............................................................................................................5
Introduction ..........................................................................................................................8
QM ...............................................................................................................................12
v
U21G ............................................................................................................................13
Sloan-C Satisfaction.....................................................................................................15
TAM.............................................................................................................................16
Accessibility.................................................................................................................28
vi
Conclusion .........................................................................................................................35
3 METHODS ..............................................................................................................................36
Introduction ........................................................................................................................36
Participants .........................................................................................................................37
Instrumentation ..................................................................................................................40
4 RESULTS ................................................................................................................................50
Introduction ........................................................................................................................50
Demographics ....................................................................................................................51
Gender ..........................................................................................................................52
Age ...............................................................................................................................52
t Test Overall Satisfaction: Peer Reviewed versus Not Peer Reviewed ......................57
vii
t-Test Mean of QM Factor Means: Peer Reviewed versus Not Peer Reviewed ..........58
Introduction ........................................................................................................................70
Demographics ..............................................................................................................72
viii
Conclusions ........................................................................................................................77
Limitations .........................................................................................................................79
REFERENCES ..............................................................................................................................84
APPENDICES:
ix
LIST OF TABLES
15 Levene’s Test for Equality of Variances for Overall Satisfaction in Peer Reviewed
Courses .....................................................................................................................................57
18 Levene’s Test for Equality of Variances for Mean of QM Factor Means in Peer
Reviewed Courses ....................................................................................................................59
x
20 Linear Regression Model Summary of Peer Review to Overall Satisfaction ..........................60
xi
CHAPTER 1
INTRODUCTION
Allen and Seaman (2010) suggested growth in online course enrollment is increasing
more rapidly than in higher education courses offered overall. As online education continues to
expand, institutions of higher education are determined to find ways of ensuring quality in online
course development (Chua & Lam, 2007; Dringus, 2000; Puzziferro & Shelton, 2008; Ross,
Batzer, & Bennington 2002). Puzziferro and Shelton (2008) insisted that developing an online
course is a “complex and multifaceted process” (p. 119) that cannot be completed by one person
alone. They also suggested that online course development needs a “common framework for
consistency, design, pedagogy and content” (Puzziferro & Shelton, 2008, p. 119)
Theoretical/Conceptual Framework
The Quality Matters (QM) framework is a faculty-centered, peer review process designed
to ensure quality in online and hybrid courses. QM’s peer-based approach for continuous
improvement has received national recognition in online education (Quality Matters, 2010). The
QM rubric was originally developed by MarylandOnline, through The Fund for the Improvement
of Postsecondary Education (FIPSE) grant and is continuously reviewed and improved upon
framework focusing on course design rather than course delivery. This framework is structured
around eight general standards: course overview and introduction, learning objectives
1
(competencies), assessment and measurement, instructional materials, learner interaction and
Evaluate) framework of instructional design by working closely with faculty to create the
standards included in their rubric (MarylandOnline, 2010). “The ADDIE framework is a cyclical
process that evolves over time and continues throughout the instructional planning and
implementation process” (Peterson, 2003, p. 228). The analyze phase allows the designers to
determine the learning needs of their audience. The design phase allows course designers to
research and plan the design of the course. The development phase allows course designers to
use the information gained in the first two phases to physically construct a course. The
development phase gathers formative evaluations to help designers improve a course before
implementation. The implementation phase allows designers to work with faculty and students to
analyze, redesign, and enhance the course. The final phase, evaluation, can actually occur
throughout the previous four phases, as course designers continually evaluate and improve upon
The purpose of this study was to examine student perceptions of quality and satisfaction
in regionally accredited online courses. A review of the literature suggested there may be
specific factors in online courses that increase students’ perceived quality of an online course
(Ross et al., 2002). This study reviewed the existing body of literature to determine which factors
were significant with regard to student perceived quality and satisfaction in online education,
which led the researcher to Aman’s (2009) dissertation “Improving student satisfaction and
2
retention with online instruction through systematic faculty peer review of courses.” Aman’s
recommendations for future research included a time-lag study and a replication study for other
types of online courses outside of the community college general education courses. This study
intended to satisfy both of those recommendations, as well as build upon Aman’s research by
further exploring the QM framework as significant with regard to student satisfaction in online
graduate courses that have not been faculty peer reviewed by QM but who subscribe to the
program. Aman found no statistical significance in the diffusion of treatment between courses
that have not been formally peer reviewed by QM but that are offered at institutions where
faculty peer reviewed courses are also offered and courses that have no affiliation to QM.
However, this led the researcher to an interesting question: Could there be a significant
difference in levels of student satisfaction between online courses that have not undergone a
systematic faculty peer review process but that are affiliated with a peer review program and
online courses that have no affiliation with a peer review program? The Quality Matters website
identifies 435 institutions that are current subscribers to QM, but that are not recognized by QM
as being peer reviewed. Is it possible that simply being subscribed to the QM framework may
enhance course design, leading to higher student satisfaction with an online course even without
Traditional educators have doubted the quality control in distance education (Casey,
2008). Accrediting bodies continue to search for ways to assess quality in online course design
(Council for Higher Education Accreditation, 2002). The Quality Matters (QM) program rubric
was created based on the “Best Practices for Electronically Offered Degree and Certificate
3
Programs,” which is the same set of principles implemented by the eight regionally accrediting
bodies in the United States (Legon, 2006). The QM program is similar to traditional accreditation
processes in that they are both essentially peer review processes. Traditional accreditation
processes, however, are focused on evaluating overall programs and institutions; while the QM
program is focused on evaluating the quality of individual online courses (Legon, 2006). The
significance of this study lies in supporting online education accreditation standards for peer
Research Questions
that have undergone a systematic faculty peer review process and online courses that have not
that have not undergone a systematic faculty peer review process but are affiliated with a peer
review program and online courses that have no affiliation with a peer review program?
3. Which factors of quality instruction most directly relate to increased levels of student
4. Which factors of quality instruction most directly relate to increased levels of student
satisfaction in online courses that have not been faculty peer reviewed but are affiliated with a
4
Assumptions of the Study
Survey participants were assumed to be 19 years of age or over, per the letter of informed
consent. It was assumed that participants would answer the survey questions honestly and
completely. The sample of students who participated in this study was assumed to represent the
Limitations of this study included the possibility of unreliable data collection. Using a
web-based survey posed the potential problem of incomplete data and duplicate submissions
(Schmidt, 1997). It is also possible that there was a great variance in the courses offered between
participating institutions.
Definition of Terms
Mean of QM factor means--the combined mean of all QM factors included in the study to
with respect to orientation, learning outcomes, services, and instructor and peer interaction
5
Recognized--institutions have undergone a formal QM peer review by three QM-certified
peer reviewers (including one from an institution outside of the one being reviewed and one who
is a content area specialist) and are recognized as including all standards deemed by QM as
essential in addition to obtaining a minimum of 85% of the possible 95 rubric points required to
meet the QM course design level of acceptance (Legon & Runyon, 2007; MarylandOnline,
2011).
subscription allows those institutions access to the automated, web-based tools, trainings, and
Non-affiliated--institutions have not undergone a formal QM peer review and do not hold
a subscription to QM.
Chapter Summary
conceptual framework used for the study, statement of the purpose, significance of the problem,
and research questions. Additionally, the assumptions and limitations of the study were
discussed, as well as the operational definition of terms. Chapter 2 of this dissertation outlines
the review of literature, which includes a review of regional accreditation, quality in online
education, satisfaction in online education, the assumption that quality equals satisfaction,
retention in online education, the conceptual framework of Quality Matters, and a review of
Aman’s research.
Chapter 3 describes the methods used to determine student satisfaction in online courses
that are (1) faculty peer reviewed; (2) not faculty peer reviewed but offered at institutions that are
6
subscribed to peer review tools, trainings, and materials; and (3) not faculty peer reviewed and
offered at institutions that have no affiliation with a peer review program. It also addresses the
setting of the study, participants, instrumentation, research questions, data collection and data
analysis.
Chapter 4 analyzes the response rates, demographics, and research questions associated
with this study. Qualitative data obtained from the QM Institutional Representative and the
Chapter 5 recaps the purpose, framework, and research questions of this study. In
addition, this chapter summarizes a discussion of the findings, the researcher’s conclusions, and
recommendations for practice. Limitations of the study and recommendations for future research
are also included. The chapter ends with the researcher’s concluding remarks.
7
CHAPTER 2
REVIEW OF LITERATURE
Introduction
This chapter outlines the review of literature, which includes a review of regional
accreditation, quality in online education, satisfaction in online education, the assumption that
quality equals student satisfaction, and retention in online education. Additionally, this chapter
reviews the conceptual framework of Quality Matters, as well as, a review of Aman’s research.
Regional Accreditation
legitimacy” (Casey, 2008, p. 49). Page, Bradford, and Canova (1997) defined accreditation as the
“official certification that a model is acceptable for use within the context of a specific objective”
(p. 396). Dill et al. (1996) defined accreditation as a way to govern that “an institution or a
program meets threshold quality criteria and therefore certifies to the public the existence of
minimum educational standards” (p. 20). Hayward (2006) defined accreditation as “a process of
self-study and external quality review used in higher education to scrutinize colleges, universities
and higher education programs for quality assurance and quality improvement” (p. 5).
improvement of courses and educational programs. The United States Department of Education
(DOE) uses regional accrediting agencies as proxies for enforcing DOE educational
8
requirements (Schmadeka, 2011). The Southern Association of Colleges and Schools (SACS)
Commission on Colleges (2011) is “the regional body for the accreditation of degree-granting
higher education institutions in the Southern states” (para. 1). Their mission is to certify the
Mississippi, North Carolina, Tennessee, Texas, and Virginia. To achieve SACS accreditation,
institutions must comply with all standards and requirements set forth by the accrediting agency.
When the American College of Surgeons (ACS) began offering regional accreditation to
who was already working with simulation, proposed a new center entitled Institute for Surgical
and Interventional Simulation. Their goal was to promote their School of Medicine with ACS
accreditation. In 2008, Sachdeva, Pellegrini, and Johnson conducted a case study on UW’s
journey to accreditation. This study reinforced the importance of accreditation by showing that
the establishment of this new center and the achievement of accreditation for the center
generated growth in both student enrollment and faculty recruitment at the institution.
program initiative to provide distance learning to undergraduate degrees in areas where shortages
had been declared. Those who participated in this qualitative study emphasized that meeting
education. Ternus, Palmer, and Faulk (2007) created a rubric to evaluate the quality of an online
course, which included structure (context, organization, and environment), content (presentation
9
of information), processes (relationships and interactions), and outcomes (mastery of content and
course evaluation) (p. 51). McGorry (2003) emphasized that a quality online course should
interaction, perceived usefulness and ease of use of technology, technical support, and student
satisfaction. Means, Toyama, Murphy, Bakia, and Jones (2010) found that including multiple
media was related to quality online instruction when the student was able to control the media.
Their study also found that student reflection was critical for student success in an online
learning environment. Herrington, Herrington, Oliver, Stoney, and Willis (2001) found that
pedagogies, resources, and delivery strategies were crucial for quality in online education.
published by various authors in the U.S. and found that the most commonly cited standards for
quality assurance in online education during that time were executive commitment, technological
infrastructure, student services, design and development, instruction and instructor services,
program delivery, financial health, legal and regulatory requirements, and program evaluation.
Phipps and Merisotis (2000) argued that institutional support, course development, teaching and
learning, course structure, student support, faculty support, and evaluation and assessment are
essential to quality in online education. Deubel (2003) indicated that “instructor’s attitude,
motivation, and true commitment toward instruction delivery via distance education programs
affect much of the quality of instruction” (para. 5). Deubel also stressed that social presence,
course materials, and technical support were related to online course quality. Bickle and Carroll
(2003) suggested several items that should be included in online education to ensure a quality
relevancy, learning objectives, and technical support. Sonwalkar (2007) insisted that quality in
10
an online course could be measured by its medial elements, learning styles, and interactivity.
Wiesenberg and Stacey (2005) identified teaching support, learning support, and administrative
Best Practices
Interstate Commission for Higher Education (WICHE) is one framework that targets quality
assurance in distance education (Johnstone & Krauth, 1996). WICHE drafted the “Best Practices
for Electronically Offered Degree and Certification Programs” (Western Cooperative for
Educational Telecommunications, 2001), which was later established by the eight regional
accrediting agencies in response to the increased demand for distance education in institutions of
higher education. The “Best Practices” are not new standards; they simply aim to clarify how the
existing accreditation standards are applicable to the emerging online environments. Eaton
(2001) agreed that classroom accreditation standards are not appropriate for the online learning
environment and that those standards need to be modified specifically for distance education.
The five standards identified by “Best Practices” include the following: institutional context and
commitment, curriculum and instruction, faculty support, student support, and evaluation and
assessment (WCET, 2009). The practices are periodically reviewed and revised as needed, with
the latest update being implemented in 2009. WCET joined with the State Higher Education
Executive Officers (SHEEO, 2003) to examine the “goals, functions, challenges, and outcomes”
of distance education across the United States (p. 1). This study included a survey sent to all
state-wide and system-wide virtual colleges and universities, as well as, multiple interviews with
system administrators. The results of this study concluded that successful institutions were those
11
that focused on the up-front learner support. Additionally, institution consortia were found to be
Sloan-C Quality
The Sloan Consortium (Sloan-C) is another framework that strives for quality assurance
and student satisfaction in online education (Sloan-C, 2011). Sloan-C is a community that
“shares techniques, strategies, and practices in online education that have worked for them”
(Sloan-C, 2011, para. 1). All practices identified as effective are peer reviewed to ensure quality.
The five pillars identified by Sloan-C include learning effectiveness, cost effectiveness and
institutional commitment, access, faculty satisfaction, and student satisfaction (Moore, 2005).
Zhao (2003) conducted a study to explore issues that may affect the quality of online education.
Sloan-C was incorporated as the framework for this study but was extended by emphasizing
course effectiveness, which was considered the “core of the overall quality of online higher
education” (Zhao, 2003, p. 218). After surveying participants of this study, Zhao concluded that
universities should implement their own quality assurance plan specific to their institution, using
the framework as a guide. More research was needed to determine significance of the framework
QM
Quality Matters (QM) is another framework that has developed a rubric for online course
design that sets a standard for measuring quality assurance, as well as accreditation standards, in
online education (Legon, 2006; Lewis et al., 2011; Shattuck, 2010). This rubric can be used to
design online higher education courses, as well as hybrid/blended higher education courses, with
12
the intention of increasing student learning, retention, and satisfaction (Legon, 2009). The QM
framework is structured around eight broad standards: course overview and introduction,
learning objectives, assessment and measurement, instructional materials, learner interaction and
engagement, course technology, learner support, and accessibility. In 2008, Ralston-Berg and
Nath surveyed students to find out how much they valued the QM rubric standards in their online
course. Students were informed that by value the researchers meant “how much they wanted the
feature in an online course” (Ralston-Berg & Nath, 2008, p. 2). Students were asked to rank the
features according to their perspective of importance. Analyses showed that students ranked 40
out of 41 standards of the QM rubric (Quality Matters, 2010) as somewhere between 2 to 3 (very
important to essential), disregarding the standard for netiquette. The researchers performed
correlations between student satisfaction, as perceived by the student, and all QM features. This
analysis indicated that students who claimed high satisfaction in their online course were
correlated with those who ranked the QM features as more valuable; whereas the students who
claimed low satisfaction in their online course were correlated with those who ranked the QM
features as less valuable. Artino (2011) also conducted a study designed to explore students’
perceptions of quality in their online courses that were designed with the QM framework. The
results of this study indicated that students viewed almost all of the QM standards as valuable to
U21G
Chua and Lam (2007) investigated the use Universitas 21 Global (U21G) as a method of
universities across four continents. The purpose of creating U21G was to allow all participating
13
institutions to pool their resources and offer high quality online education via collaboration
between the institutions. The quality assurance processes measured through this program include
recruitment, pedagogy, and delivery. All course content is peer reviewed through a double-blind
review process. Online courses offered through this program contain “textual material, graphics,
interactive exercises, assignment questions, test questions, online library and links to other online
resources” (Chua & Lam, 2007, p. 135). At the end of the course, students were given a 5-point
Likert-type survey to assess their perception of course quality. Chua and Lam concluded that the
quality assurance measures implemented by U21G greatly improved the overall quality of the
Opinions on what constitutes student satisfaction vary across the discipline. Lee (2010)
claimed that timely feedback from instructors is essential to student satisfaction in an online
learning environment. Social presence is another factor emphasized as leading to higher student
satisfaction in online education (Abdous & Yen, 2010; Richardson & Swan, 2003). Support
services have also been characterized as a predictor for student satisfaction in online courses
(Lee, 2010). McGorry (2003) indicated that student satisfaction is affected by the flexibility in a
course, social presence, technical support, and course technology. Lorenzo and Moore (2002)
declared that student satisfaction is a product of responsive, timely, personalized services and
support; high-quality learning outcomes; academic and administrative support services; and
14
Seven Principles
Babb, Stewart, and Johnson (2010) studied student perceptions in a hybrid course using
the framework by Chickering and Gamson (1983): Seven Principles for Good Practice in
among students, active learning, prompt feedback, emphasis on time on task, communication of
high expectations, and respect of diverse talents and ways of learning. The study conducted by
Babb et al. (2010) found that students who were active in their learning were likely to be more
satisfied with their online experience. By allowing students to bring personal experiences into
discussions, it brings a more personal atmosphere altogether into the online course. Students
should be encouraged to participate in active learning, peer reviews, and team-building problem-
Sloan-C Satisfaction
Fredericksen, Pickett, Shea, Pelz, and Swan (2000) collected data from a student
satisfaction survey in 1999 based on the Sloan-C framework for quality and satisfaction in online
education. The results of this study indicated that students were more satisfied with their course
when they had higher levels of interaction with their teacher and their classmates. However,
students also stated higher levels of satisfaction with their online course when they were satisfied
with the reliability of the technology and when they were satisfied with the level of technical
support they received during technical difficulties. Students who were taking the online course
for its flexibility reported higher levels of satisfaction than those who were taking it only because
15
TAM
Arbaugh (2000) conducted a study using the Technology Acceptance Model (TAM)
framework, which focuses on the usefulness of technology and the perceived ease of use of
technology as predictors for user satisfaction. Arbaugh concluded that the usefulness of
positively related to student satisfaction in their online course. Flexibility in the course was also a
positive factor in student satisfaction. Swan’s (2001) study found that students were more
satisfied with their online course when the course structure was consistent and easy to use.
Participants in this study also reported higher levels of satisfaction when they were satisfied with
the level of instructor/student, student/student, and student/content interaction. Carr and Hagel
(2008) conducted a study in which students were surveyed to assess their perceptions of quality
in online and face-to-face courses in Australia. This study found that student satisfaction
Eom, Wen, and Ashill (2006) conducted a study to assess factors related to student
learning and satisfaction in online education. They used the conceptual framework Determinants
of Learning Effectiveness, created by Piccoli, Ahmad, and Ives (2001), which points to both
human and design factors as essential to online course quality. In this framework, human factors
include students and instructors; whereas, design factors include technology, learner control,
course content, and interaction. More specifically, Eom et al. predicted that student satisfaction
would be related to student self-motivation, student learning style, instructor knowledge and
facilitation, instructor feedback, interaction, and course structure. The results of this study
16
concluded that all six of these factors were significant with regard to student satisfaction in their
online course.
ADL Network
Artino (2008) conducted a study aimed at predicting characteristics that would contribute
to student satisfaction of their online course. Artino used the Advanced Distributed Learning
(ADL) network as a framework for the study. Online courses developed by the U.S. Navy were
evaluated in this study. An anonymous, Likert-type survey was distributed to U.S. service
academy students at the completion of their online course. Results of this study indicated that
student’s self-reported “task value, self-efficacy beliefs and perceptions of instructional quality”
(Artino, 2008, p. 266) were significant predictors of overall student satisfaction of the online
course.
Sun et al. (2008) conducted a study using a framework designed specifically for the
study, which included six dimensions based on a review of literature: learner, instructor, course,
technology, design, and environment. Validity of the framework was established through a series
of interviews with experienced e-learners and then reliability was established via a pilot test of e-
learners. The final questionnaire included a Likert-type scale that was distributed to e-learners
who had not already participated in the pilot test. The purpose of this study was to examine
factors that influenced student satisfaction in online education. The study concluded that
computer anxiety, instructor attitude, flexibility, course quality, usefulness, ease of use, and
17
diversity in assessments were the factors that most influenced student satisfaction with their
online course.
Hong (2002) conducted a study using the observational case study design (Bogdan &
Biklen, 1998) to evaluate student satisfaction in their problem-based online course. Hong
interacted with students participating in the research. Data were collected by distribution of
questionnaires, faculty records, and interviews with the students. The results of this study
indicated that gender, age, and learning styles showed no relation to student perceptions of
satisfaction in their online course; however, students who were more skilled with computers
In reviewing the literature for online course quality and student satisfaction, it appeared
that the same factors were identified as promoting both quality and satisfaction. It also appeared
that an online course that included quality standards would lead to higher student satisfaction of
that course. Sloan Consortium (2011) identifies effective online education as including the
element of student satisfaction. Rodriguez, Ooms, and Montanez (2008) conducted a study of
student perceptions regarding online quality instruction and found that student satisfaction was
related to perceived quality of course delivery. Sun et al. (2008) agreed that online course quality
was a critical factor in student satisfaction. Zhao (2003) also agreed that student satisfaction is
standard in quality online education. Based on the literature, the researcher assumed that an
18
online course designed with a reputable quality framework would result in higher student
Rovai (2003) insisted that student retention was an important factor for determining
quality assurance and student satisfaction in higher education. Rovai claimed that retention rates
in higher education are “strongly related to the ability of educational programs to satisfy adult
needs” (p. 2). Tinto (1975) claimed that insufficient social presence may result in lower retention
rates in online education. He also believed that teamwork and collaboration supported greater
retention rates in higher education. Workman and Stenard (1996) concluded that the degree to
which distance learner’s needs are met influenced their retention rates. These needs included
“detailed information about the school, educational programs, and courses,” “an online student
manual that covers the e-learning system in detail,” the “names, e-mail addresses, and phone
numbers of online instructors, advisers, and technicians,” “prompt feedback and replies,”
support services” (Workman & Stenard, 1996, pp. 10-11). Morris, Wu, and Finnegan (2005)
suggested that retention rates in higher education distance education may be related to student
demographic characteristics, such as high school GPA, educational background, age, and gender.
Lorenzo and Moore (2002) indicated that retention rates in online education were related to
student satisfaction.
Nash’s (2005) study of student retention in online education suggested that better course
development, instruction, management, and timely feedback might increase student retention.
Herbert (2006) researched student satisfaction and retention in online courses and found that
19
students who claimed to be satisfied with the factors of their online course tended to be related to
the completion rate of the online course; however, a model for predicting student retention was
never produced as significant. Rovai (2003) disagreed with Hebert and further claimed that
student persistence, and therefore retention, is generally based on both external and internal
student characteristics, such as hours of employment, family responsibilities, study habits, stress,
etc. Nichols’ (2010) study agreed with Rovai by showing that most students who dropped out of
their online course named personal reasons for their withdrawal. Eom et al. (2006) agreed that
Fisher, and Han (2007-2008) specifically studied the use of the QM framework on student
retention. Although this study showed a 95% retention rate in all courses that were designed
using the QM rubric, there was no scientific comparison to show whether or not the QM
framework aided in student retention rates because the study did not include courses that were
Quality Matters
Although there are varying definitions of what constitutes quality and satisfaction in
online education, and varying frameworks for evaluating these characteristics in online courses,
this study focused on the Quality Matters framework due to the shortage of literature
documenting quality and satisfaction in online education using QM. One case study was
documented in 2006, and a few case studies were found from 2009 and thereafter, but the
literature on QM is still thin. This research expected to add to that body of literature.
Quality Matters has received “national recognition for their research-based rubric and
inter-institutional peer review processes” (Puzziferro & Shelton, 2008, p. 125). The values that
20
initiated the creation of the rubric include those established by The Guiding Principles for
Distance Learning in a Learning Society and the Distance Learning Evaluation Guide (Shattuck,
2010). Quality Matters routinely collects feedback from QM Rubric subscribers and peer
reviewed institutions and uses that feedback to revise the rubric as needed. This rubric can be
used to design online higher education courses, as well as hybrid/blended higher education
courses, with the intention of increasing student learning, retention, and satisfaction (Legon,
2009).
Quality Matters is a faculty-centered, peer review process that certifies the quality of
online courses that are built upon the QM framework. This framework is structured around eight
broad standards: course overview and introduction, learning objectives, assessment and
learner support, and accessibility. Within these broad standards, there are 41 specific standards
important. According to the QM rubric, all essential standards with a 3-point rating must be
present in an online course. Furthermore, out of 95 possible rubric points, a course must obtain a
minimum of 85% (or 81 points) to meet the QM course design level of acceptance. Peer review
of a course includes three QM-certified reviewers, one of which must be from an institution
outside of the one being reviewed, and one of which must be a content area specialist (Legon &
recommendations for existing course improvement, as well as, quality thresholds for new
courses. Additionally, course reviews provide institutions with quality process indicators for
21
Course Overview and Introduction
Conrad (2002) suggested that students feel anxiety when starting a new course. He
suggested giving a thorough orientation at the beginning of each online course to relieve some of
this apprehension. Conrad also identified students as preferring an organized course structure,
introduction from the instructor, and clear expectations of the course. Achtemeier, Morris, and
Finnegan (2003) agreed that giving students clear expectations at the beginning of the course, as
well as structured deadlines to keep students on pace, is preferred by students. Workman and
Stenard (1996) suggested that students should be given a detailed online student manual at the
beginning of a course to assist in learning the policies, procedures, and technical requirements
for that online course. Shea, Pickett, and Pelz (2003) indicated that an effective online course
will include setting the curriculum, designing methods, establishing time parameters, utilizing
the medium effectively, and establishing netiquette at the start of the course. Setting the
curriculum was defined as the instructor clearly communicating important course outcomes and
course topics. Designing methods was defined as the instructor providing clear instructions on
participation expectations. Establishing time parameters was defined as the instructor clearly
communicating the course schedule and deadlines. Utilizing the medium effectively was defined
as the instructor providing clear instructions regarding how to use the course media. Establishing
netiquette was defined as the instructor providing students with clear instructions and examples
Learning Objectives
22
sphere” (Artino, 2011). Providing learning objectives to students at the beginning of a course
(whether online or face-to-face) is important for a clear understanding of what is expected from
that course. Students should be told what outcomes are expected of them and any assessment
taken by that student should measure those expected outcomes (Achtemeier et al., 2003;
Hatziapostolou & Paraskakis, 2010; Palomba & Banta, 1999). Workman and Stenard (1996) also
claimed that identifying clear and measurable learning objectives is critical to student success in
technological problems during exams, and learner anxiety are disadvantages of taking online
assessments (Yates & Beaudrie, 2009). In an effort to offset these disadvantages, students should
be assessed using a variety of methods (Rovai, Ponton, & Baker, 2008; Palomba & Banta, 1999),
by Thurmond, Wambach, Connors, and Frey (2002) found that the strongest predictor of student
satisfaction was being assessed by a variety of methods. Another predictor of student satisfaction
included prompt feedback from instructors. Feedback on assessments should be prompt and
specific to each student (Chickering & Gamson, 1983; Hatziapostolou & Paraskakis, 2010;
Palomba & Banta, 1999), as many students are motivated by the satisfaction or repercussion of
grades (Rovai et al., 2008). Effective feedback should be “timely, constructive, motivational,
personal, manageable and directly related to assessment criteria and learning outcomes”
(Hatziapostolou & Paraskakis, 2010, p. 111). Soon, Sook, Jung, and Im (2000) participated in a
study to assess student satisfaction in an online course via a questionnaire at the end of the
23
course and found negative responses in relation to inadequate feedback from professors.
Workman and Stenard (1996) claimed that students need a sense of self-esteem to be successful
in an online course. He stated that students’ self-esteem can be nurtured by timely feedback.
Instructional Materials
Achtemeier et al. (2003) suggested that when planning an online class, it is important to
assign tasks where resources for that task are easily accessible to every student in the class. With
the expansion of online education, these resources may largely be electronic and available
through a digital library or course management system (Dong, Xu, & Lu, 2009). Dong et al.
found that students reported lower rates of satisfaction with their online education when certain
factors, such as inadequate resources, were identified. Deubel (2003) agreed that appropriate use
of media and interactivity with content contributed to course quality and student satisfaction.
Deubel advised that instructional materials be centrally located for ease of access and that a list
of additional resources (such as supplementary texts and websites) may be helpful for students
seeking further exploration of course content. Deubel (2003) also indicated that instructor
websites should include online help and technical assistance, frequently asked questions, plug-
ins needed to view course content, online libraries and databases, and a list of supplementary
websites related to course content, and that graphics and other visuals should be relevant for
course content and sized appropriately to reduce student frustration from long download times.
Moore (1989) introduced us to the concept of transactional distance, which is the measure
of distance one feels between himself and his peers or between himself and his instructor. Tomei
24
(2006) believed that the most important thing in an online course is teacher-student interaction.
Grandzol and Grandzol (2010) supported this in their study, which found learner to instructor
interaction showed the strongest correlation with student perceptions of learning. Walther,
Anderson, and Park, as quoted in Casey (2008, p. 50), claimed, “When cues are filtered out,
communication becomes more ‘task oriented, cold and less personal than face-to-face
interaction with their instructor leads to student dissatisfaction with their online course, as well
as higher withdrawal rates from that course (Abdous & Yen, 2010; Steinman, 2007). To reduce
transactional distance between students and instructors, Steinman (2007) suggested that
instructors set virtual office hours for students, which is most effective if the student can both see
Steinman (2007) indicated that discussion boards and chat tools help facilitate
communication between students, which helps to reduce transactional distance between students
to other students. Baturay and Bay (2010) studied problem-based learning (PBL) and learned that
students were more satisfied in PBL courses due to the social interaction increasing students’
perceptions of “connectedness” (p. 51). Rovai (2002) suggested using team-building activities to
reduce transactional distance between students, although Baturay and Bay’s study contradicted
this by showing that students working in team environments showed poorer achievement levels,
Baturay and Bay (2010) concluded that problem-based learning minimizes the effects of
transactional distance between students and their content. Swan (2002) showed that students who
had higher interaction with their instructor, other students, and course content combined reported
25
higher levels of satisfaction with their online course. Interestingly, Grandzol and Grandzol
(2010) found that higher levels of interaction in an online course environment appeared to have a
relationship with decreased student retention rates in those courses. This is supported by Rourke,
Anderson, Garrison, and Archer (2001) who found that although social presence is important for
an online learning environment, too much interaction may actually lead to lower achievement.
Grandzol and Grandzol (2010) also found that even though student-instructor interaction showed
a positive correlation with student perceptions of learning in their online course, there was no
Course Technology
Augustsson (2010) stated that online education has now become more of a collaborative
student learning process, rather than just a way for instructors to deliver instruction. Within this
(Cohen, Carbone, & Beffa-Negrini, 2011). This may include Virtual Lecture Halls (VLH) such
as Tegrity, Camtasia, and Captivate (Cramer, Collins, Snider, & Fawcett, 2006); e-mail; web
sites; discussion boards; chats; wikis; podcasts; videocasts; social bookmarking; social
networking; blogs; and document sharing services (An, 2010; Arnold & Paulus, 2010;
Augustsson, 2010; Halic, Lee, Paulus, & Spence, 2010; Kear, Woodthorpe, Robertson, &
Hutchison, 2010; Preston, Phillips, Gosper, McNeill, Woo, & Green, 2010). Web 2.0 software,
such as social networking and wikis, allow higher levels of interaction between students,
Cohen et al. (2011) noted that “using technology for the sake of technology” (p. 84) may
decrease student learning in an online course. Additionally, Schroeder et al. (2010) found that
26
incorporating technological features into a course is not sufficient unless the instructor facilitates
the use of those tools. Course management systems (CMS), such as Angel, Blackboard, Moodle,
etc., were designed to incorporate technologies into a manageable course environment (Arnold &
Learner Support
Rovai et al. (2008) suggested that online students expect the same type of and
convenience of access to support services that is offered to on-campus students. This may
include enrollment support, instruction support, and particularly technical support. Workman and
Stenard (1996) agreed that students need “ready access to support services such as bookstores,
library, financial aid offices, and advisers” (p. 11). They also suggested that students in an online
learning environment need access to tutoring, study preparation, and technology training.
Additionally, they suggested using tools that support multiple learning styles to accommodate
students with varying needs. Lee (2010) agreed that students need support services such as
assistance with registration, financial aid, and technical support. McGorry (2003) stated that
students in an online course need access to institutional resources as well as electronic resources
that keep the student engaged in their course. Shea and Armitage (2002) argued that many
institutions have neglected support services for their distance education students due to lack of
resources and flexibility. They further indicated that online education will never see the same
level of success or retention as campus-based education until online students have access to all
student services made available to on-campus students. Shea and Armitage insisted that services
that cannot be offered online be offered in some format that is an accessible and affordable
27
alternative to students, such as e-books or mail-order textbooks, online library resources, test
Accessibility
“Online courses can inadvertently erect barriers for students and instructors with
disabilities” (Burgstahler, Corrigan, & McCarter, 2004, p. 234). The Americans with Disabilities
Act (ADA) requires that online courses be compatible with assistive technologies. Section 504,
which is an extension of the Americans with Disabilities Act of 1990, prohibits institutions from
discriminating against students with disabilities regardless of whether they are federally funded
(Burgstahler et al., 2004). Section 508 of the Rehabilitation Act requires that information and
data be accessible nationwide by individuals with disabilities (United States Government, 2011).
The Assistive Technology Act, enacted by the United States Congress (1998) requires states
(including collegiate institutions) to comply with Section 508. Pearson and Koppi (2001) created
a set of guidelines entitled Guidelines for Accessible Online Courses that could be used to assist
course developers in creating online courses that are accessible. These guidelines include the
following:
1. Ensure consistent and appropriate use of graphics, icons and other visual cues,
2. Ensure all graphics, figures, and other illustrations include a text equivalent,
3. Organize content to take account of the transition to an online environment,
4. Use the features provided by WebCT [or other learning management systems] to
organize and structure your course content,
5. Make PDF and other read only file formats accessible,
6. Be aware of the limitations of screen readers in interpreting unusual text, characters,
and abbreviations,
7. Ensure that tables are carefully and appropriately used,
8. Ensure appropriate use of colors and contrasts in screen design,
9. Provide alternative sources of information for video or audio or other moving images,
and
10. Use style sheets to format text and control layout. (Pearson & Koppi, 2001, pp. 7-19)
28
Additional QM Case Studies
In 2011, Swan, Matthews, Bogle, Boles, and Day conducted a study that implemented
both the QM and the Community of Inquiry (COI) frameworks to redesign one online course,
hoping to enhance student learning. The COI framework is primarily a subjective, constructivist
framework embedded in course implementation (Swan et al., 2011). The idea was to redesign a
course to meet QM design standards and then tweak it based on COI ratings. Results of the study
suggested that using the two frameworks simultaneously may positively improve student
outcomes.
Legon (2006) conducted a study to compare the “Best Practices for Electronically
Offered Degree and Certificate Programs”--which was adopted in 2001 by the eight regional
accrediting bodies across the United States--with the QM rubric to see if an online course
designed upon the QM framework would meet the accreditation guidelines for online education.
Legon provided a one-to-one correspondence with the Best Practices for Electronically Offered
Degree and Certificate Programs and QM principles. Although there were a few Best Practices
principles that could not be paralleled with QM principles, those principles were not found to be
in conflict with QM. Overall, Legon (2006) found the QM rubric to be “fully consistent with
published accreditation standards for online education” (p. 9) and he suggested the rubric could
Little (2009) conducted a study using two sets of online course standards: the 2006
College of Public Health (COPH) Online Course Standards and the 2005 Quality Matters (QM)
Peer Course Review Rubric. The purpose of the study was to observe a convenience sample,
program. In the end, the study suggested that either instrument could provide accurate
29
evaluations of online courses, so deciding which tool to use should be based on other measures,
such as user-friendliness, as the reviewers declared that the COPH standards were “time
consuming and difficult to use” (Little, 2009, p. 413). The peer reviewers for this study
suggested adoption of the QM rubric based on its “ease of use, content validity, and national
McFerrin, Duchardt, Furr, Horton, and Gentry (2009) studied the design and
using the QM rubric as the basis for designing these online courses. To measure the effectiveness
of each online course, faculty participated in peer evaluations for each course via a survey with
both closed and open-ended questions. Additionally, students enrolled in each course completed
a survey at the end of their term giving feedback for the course. Finally, an external evaluation
was completed by an objective third-party that analyzed both the peer and the student feedback.
While the courses were still being developed and tweaked at the time of this study, there is no
cumulative data to show the effectiveness of the QM rubric in designing these courses.
In 2009, Pollacia, Russell, and Russell engaged in a study that created an online minor in
Computer Information Systems (CIS), using the QM rubric as the basis for development. As of
the publication date, there were no data to show whether or not this effort increased enrollment
Aman’s Research
Aman (2009) began his research by determining whether online courses that were faculty
peer reviewed would increase student satisfaction and retention rates, compared to online courses
that were not faculty peer reviewed. He also studied courses that were not faculty peer reviewed,
30
but that were offered by institutions in which faculty peer reviewed courses were offered, to see
if there was a diffusion of treatment between faculty at those institutions. He categorized these
His study also looked at which specific quality standards were most directly related to student
satisfaction and retention rates. The research questions for Aman’s (2009) study included the
following:
courses that have undergone a systematic faculty peer review process and non-peer
reviewed courses?
2. Is there a significant difference in course retention rates between online courses that
have had a systematic faculty peer review process and non-peer reviewed online
courses?
3. Which factors of quality instruction most directly relate to increased levels of student
4. Which factors of quality online instruction are most important in terms of increased
Participants in this study included 455 students, among 41 courses. To gather data on student
scale survey administered to students enrolled in online courses. Data on student retention were
31
obtained at the course level and compared student enrollment during the end of the first week of
class with student enrollment at the end of the course for each course included in the study.
Aman (2009) participated in QM training to become a certified peer reviewer for the process
of implementing online courses through the institution from which he was employed. Being
familiar with the QM rubric, Aman chose this as his peer reviewed framework. The rubric
2. Learning Objectives,
4. Instructional Materials,
6. Course Technology,
8. Accessibility,
Aman (2009) found only seven of those standards supported in his theoretical research:
32
He then concluded that two of the standards, “Course Overview and Introductions” and “Learner
Support” were not significantly supported theoretically because (1) Course Overview and
Introductions was not discussed in the majority of his research, and (2) Learner Support was
typically found to be out of the faculty member’s control. Therefore, Aman (2009) proceeded
with his research assessing only five of the QM standards (see Table 1).
Table 1
Based on his review of the literature, Aman (2009) explored the following demographic
variables, to see if they provided any correlation to student satisfaction: (1) student age, (2)
sudent gender, (3) number of prior online courses taken, and (4) student level of comfort with
online technology.
After an extensive review of literature, Aman (2009) defined student satisfaction in two
ways. First, students were questioned about their levels of satisfaction with QM factors of quality
found to be present in their online peer reviewed course. The assumption was that courses
containing factors of quality would stimulate higher satisfaction among students than courses
that did not contain the same factors of quality. Second, student retention rates were compared
within their online course and other similar courses that had not been peer reviewed. The
assumption was that students would drop an online course they felt did not satisfy their needs.
33
Aman’s Findings
Aman’s study found that comfort with distance learning, age, and gender may be related
to student satisfaction in their online courses. A review of the literature supported that students’
prior level of computer skills may be conducive to success and satisfaction with their online
course (Abdous & Yen, 2010; Koroghlanian & Brinkerhoff, 2008; Lim, 2001; Menchaca &
Bekele, 2008; Sahin & Shelley, 2008; Thurmond et al., 2002). Richardson and Swan (2003),
however, found that age was not a significant factor in student satisfaction, which is supported
by Thurmond et al. (2002), Hong (2002) and Lim (2001). The review of literature found mixed
reviews regarding gender. Kim and Moore (2005) and Richardson and Swan (2003) found
gender to be a significant factor in student satisfaction with online courses; whereas, Hong
(2002) and Lim (2001) found no statistical significance with gender in relation to student
Aman’s (2009) research found no significance in the number of prior online courses
taken, which is supported by Richardson and Swan (2003), whose research showed that the
number of college credits earned online showed no significant relationship to student satisfaction
in online courses. Thurmond et al. (2002) also showed that the number of previous online
courses taken was not a predictor of student satisfaction in an online learning environment.
Arbaugh (2001) found that prior student experience with online courses was actually negatively
associated with student satisfaction of their course. Arbaugh and Duray’s (2002) study, however,
disagreed by showing that students with more experience in online learning reported higher
Diffusion of treatment was found to have no statistical significance for courses called
“non-reviewed;” therefore, for the remainder of Aman’s (2009) study he combined the “non-
34
reviewed” and “no-review” courses into a category called “all other.” The last question, which
asked for student overall satisfaction, was not found to be significant with regard to faculty peer
reviewed courses; so Aman calculated a mean of QM factor means, based on the combined
means calculated for all of the QM standards included in the study. The mean of QM factor
means did result in a significant difference in satisfaction between courses that were categorized
as “reviewed” and “all other.” None of the variables identified in Aman’s research showed a
statistical significance to student retention. A review of literature supported that identifying best
practices for student retention is difficult to quantify and is non-consistent (McGivney, 2004;
Nichols, 2010; Patterson & McFadden, 2009; Tait, 2004; Woodley, 2004).
Conclusion
To further validate Aman’s (2009) questionnaire, the five QM factors identified by Aman
as significant to student perceptions of quality and satisfaction will be included in this study.
This consisted of learning outcomes or objectives, student assessment and measurement, learning
resources and materials, learner interactions (instructor, student, content), and course technology.
The demographic variables, comfort with distance learning, age, and gender, showed a
possible relationship with student satisfaction in Aman’s research. The review of literature,
however, did not show consistency pertaining to significance of these variables. Number of prior
online courses taken did not show a significant relationship with student satisfaction in Aman’s
study, and the review of literature also dismissed the significance of this variable. However, to
maintain the validity and reliability of the survey instrument, all four demographic variables
35
CHAPTER 3
METHODS
Introduction
This chapter describes the methods used to determine student satisfaction in online
courses that are (1) faculty peer reviewed; (2) not faculty peer reviewed but offered at
institutions that are subscribers to peer review tools, trainings, and materials; and (3) not faculty
peer reviewed and are offered at institutions that have no affiliation with a peer review program.
It also addresses the setting of the study, participants, instrumentation, research questions, data
The setting of this study included senior institutions accredited by the Southern
Association of Colleges and Schools (SACS) in which upper-level and/or graduate courses were
offered in an online format. Institutions asked to participate in this study included some that
offered online courses peer reviewed by the QM rubric; some that offered online courses not
peer reviewed by the QM rubric but offered at institutions that were subscribed to QM tools,
trainings, and materials; and some that offered online courses not peer reviewed by the QM
rubric and offered at institutions that had no affiliation with QM. To avoid a potential diffusion
of treatment, institutions that were identified in one category were excluded from the other
categories.
36
Participants
This study included upper-level and/or graduate online classes offered at senior
institutions accredited by SACS. The participants identified in this study included three groups of
participants over the age of 19: students who were enrolled in online courses at regionally
accredited institutions that had been peer reviewed using the QM rubric; students who were
enrolled in online courses at regionally accredited institutions that had not been peer reviewed
using the QM rubric but that were offered at institutions who were subscribed to QM tools,
trainings, and materials; and students who were enrolled in online courses at regionally
accredited institutions that had not been peer reviewed using the QM rubric and that had no
affiliation with QM. For the remainder of this study, the three categories were known as follows:
2. Subscribed--have not undergone a formal peer review through QM but are subscribed
3. Non-affiliated--have not undergone a formal peer review through QM and are not
subscribed to QM.
Member, Candidate, and Applicant List (2011), there were 805 institutions regionally accredited
through SACS. A comparison of the SACS Member, Candidate, and Applicant List; the QM
website; and the individual websites of each of the institutions included in these two references
determined that 10 institutions met criteria for the “recognized” category, 72 institutions met
criteria for the “subscribed” category, and 453 institutions met criteria for the “non-affiliated”
category. The remaining 270 institutions did not meet research criteria for inclusion in this study.
37
To maintain the confidentiality of data, institutional names of participants have not been
disclosed. Upon receiving Institutional Review Board (IRB) approval from the researcher’s
home institution (Appendix A), the QM Institutional Representative was contacted at each of the
10 institutions that met research criteria for the “recognized” category to request support for this
study. Of the 10 institutions who were contacted, 4 agreed to participate in this study, 3 declined
participation, and 3 did not respond. IRB approval was granted from the 4 institutions who
agreed to participate. A total of 334 students were enrolled in the 8 online courses recognized by
QM across these 4 institutions, and were solicited for participation in this study. Those who
declined participation gave the following reasons for not participating: (1) institution is closing
for restructure, (2) solicitation of students is prohibited, (3) students are already participating in
another study and the institution does not want to burden them with a second request. Non-
respondents were sent follow-up requests anywhere from one to four times, in 2-week intervals
between April 26, 2012 and July 30, 2012. Research invitations and follow-ups are further
select 10 of the 72 institutions that met research criteria for the “subscribed” category. Of the 10
selected, IRB approval was granted from 3 of those institutions, IRB approval was denied from 2
of the institutions, and 5 of the institutions did not respond. Of the institutions who agreed to
participate, online courses were chosen that were similar to the courses studied from other
participating institutions. A total of 764 students were enrolled in the 26 online courses chosen
from these 3 institutions, and were solicited for participation in the study. Those who declined
IRB approval gave the following reasons for not participating: (1) solicitation of students is
prohibited, (2) institution is in transition with implementing QM and therefore not a good
38
candidate for this study. Non-respondents were sent follow-up requests anywhere from one to
four times, in 2-week intervals between April 26, 2012 and July 30, 2012. Research invitations
and follow-ups are further discussed under the section entitled Data Collection.
The researcher used Research Randomizer to randomly select 10 of the 453 institutions
that met research criteria for the “non-affiliated” category. Of the 10 selected, IRB approval was
granted from 3 of those institutions, IRB approval was denied from 3 institutions, and 4
institutions did not respond. Of the institutions who agreed to participate, online courses chosen
that were similar to the courses studied from the other participating institutions. A total of 676
students were enrolled in the 24 online courses chosen from these 3 institutions, and were
solicited for participation in this study. Those who declined IRB approval gave the following
reasons for not participating: (1) it is not a good time to solicit student participation as it would
conflict with other data collection and activities currently underway, (2) student solicitation is
prohibited, (3) students are already involved in another study and a second request would be
burdensome. Non-respondents were sent follow-up requests anywhere from one to four times, in
2-week intervals between April 26, 2012 and July 30, 2012. Research invitations and follow-ups
Students enrolled in the courses solicited were emailed a research invitation requesting
institutional representative (where applicable) or the course designer for each course included in
this study was sent a research invitation requesting their participation in an online qualitative
survey to assess if and how they use peer review in the design of the courses. Institutional reps
(or course designers) were sent research invitations anywhere from one to four times, in 2-week
39
intervals between April 26, 2012 and July 30, 2012. Research invitations and follow-ups are
Instrumentation
With Dr. Aman’s permission (Appendix B), this study used Aman’s Satisfaction
Instrument, which included a 5-point Likert-type scale ranging from 1 (strongly disagree) to 5
(strongly agree), with a neutral midpoint. This questionnaire began by asking a few questions
pertaining to each of the five QM standards that could be directly related to student satisfaction
in their online course: learning outcomes, assessment and measurement, learning resources and
technology. Each question attempted to establish the importance of that standard to the student,
as well as the student’s satisfaction with that particular standard in their online course.
Importance was defined as “a general, student reported, benchmark for the value placed on each
of the factors,” while satisfaction was defined as “the student’s level of approval for that specific
related to student satisfaction with their online course: comfort level with technology, gender,
number of prior online courses completed, and age. Questions 1, 27, 28, and 29 were used to
measure the demographic variables. This questionnaire also assessed background information
about the online course, including course introduction, technology support, and student support,
which were measured in questions 2, 3, and 4, respectively. One final question (question 30) was
used to determine students’ overall satisfaction with their online course. The independent
40
variable identified in this research was faculty peer review in online courses. Table 2 shows the
Table 2
Aman (2009) used his review of literature to create construct validity of the questionnaire
by designing the questionnaire around the QM framework. He tested this questionnaire for
content validity by asking a panel of online learning experts and QM experts to review the
questionnaire. After eight revisions, the questionnaire was pilot-tested on two treatment and two
control groups. During this pilot test, the participants were surveyed twice during a semester,
once at midterm and once at the end of the semester (prior to final exams). Factor analyses
showed a strong positive correlation between the QM factors chosen for this study and the
satisfaction questions that were created based on the QM framework. A mean Cronbach’s alpha
41
of .84 confirmed reliability of the questionnaire. The final questionnaire used to survey
participants in Aman’s study included 30 questions (Appendix C). This questionnaire was
peer reviewed and all other online courses. Student enrollment was obtained from the course
category was sent a research invitation (Appendix D) by email that asked questions regarding the
level of QM implementation across their institution. The QM Institutional Representative (or the
course designer if the QM Representative was unavailable) of the institutions in the “subscribed”
category was sent a research invitation by email that asked questions regarding how many of the
QM resources were implemented in their courses and which specific tools were used. The course
designer of the institutions in the “non-affiliated” category was sent a research invitation by
email that asked questions to determine if the institution was using another type of peer review
evaluation (other than QM) and, if so, was it being used institution-wide or just in particular
courses. Also, questions were asked of this category to determine what specific peer review
The qualitative questions asked of the QM Institutional Representative and/or the course
Recognized
42
a. Rubric
b. Faculty development
d. QM peer-review
e. Self-study
Subscribed
Rubric
Faculty development
QM peer-review
Self-study
Both
Select Yes or No
43
Non-affiliated
1. Are you subscribed to any type of peer-review quality assurance program for your
online courses/programs?
Select Yes or No
If yes,
Research Questions
that have undergone a systematic faculty peer review process and online courses that have not
that have not undergone a systematic faculty peer review process but are affiliated with a peer
review program and online courses that have no affiliation with a peer review program?
3. Which factors of quality instruction most directly relate to increased levels of student
4. Which factors of quality instruction most directly relate to increased levels of student
satisfaction in online courses that have not been faculty peer reviewed but are affiliated with a
44
Data Collection
Upon receiving IRB approval, the researcher requested student email addresses of the
students enrolled in courses meeting research criteria from each institution’s registrar. In a few
cases, when email addresses were not released, the QM Institutional Representative or the course
instructor (where applicable) was asked to forward the research invitation to their students on the
researcher’s behalf. Student research invitations (Appendix F) were sent anywhere from one to
four times, in 2-week intervals between April 26, 2012 and July 30, 2012, as shown in Table 3.
The institution who only received one invitation (with zero follow-ups) insisted that follow-ups
would not be allowed for their students. This study included data from the Spring 2012 term and
data from the Summer 2012 term. Students were contacted between the midterm and end of their
Table 3
45
SurveyMonkey was used to distribute the student satisfaction questionnaire to
participants of this study due to its anonymous collection of data. Participation was voluntary. A
letter of informed consent (Appendix G) was posted on the opening page of the survey. If a
student agreed to participate, he was asked to click “Continue” at the bottom of the screen and
proceed answering the questions that followed. If a student chose not to participate, he was asked
to “Exit” the survey and no further action was required of that student. This survey did not
collect internet protocol (IP) addresses, or any other identifiable information, of its participants.
The survey did not require a response for every question; therefore, the participant had the option
to not answer some or all of the questions. The participant was able to withdraw participation at
Institutions in the “subscribed” and “non-affiliated” categories were sent research invitations to
peer review being implemented in online course design, if any. These research invitations were
sent were sent anywhere from one to four times, in 2-week intervals between April 26, 2012 and
study due to its anonymous collection of data. Participation was voluntary. A letter of informed
consent (Appendix H) was posted on the opening page of the survey. Participants were asked to
click “Continue” at the bottom of the screen and proceed to answering the questions that
followed. Anyone who chose not to participate was asked to “Exit” the survey and no further
46
action was required of that individual. The survey did not collect internet protocol (IP) addresses,
or any other identifiable information, of its participants. The survey did not require a response
for every question; therefore, the participant had the option to not answer some or all of the
Data Analysis
Student satisfaction was measured by (1) students’ overall satisfaction with their online
course, and (2) students’ satisfaction with each of the five QM framework factors included in this
study: learning outcomes, assessment and measurement, learning resources and materials, learner
interactions, and course technology. Demographic variables included in this study were student
age, student gender, number of prior online courses completed, and student level of comfort with
online technology. Background data was collected to determine the whether the online course
included a course introduction, technology support, and student support. Finally, data was
collected to determine the level of peer review that was included in the design of each online
course.
any of the demographic variables and overall student satisfaction in their online
course.
course introduction, technology support, and student support were included in the online course.
The first research question, “Is there a significant difference in levels of student
satisfaction between online courses that have undergone a systematic faculty peer review process
47
and online courses that have not undergone a systematic faculty peer review process?,” was
A t-test analysis determined whether the satisfaction score means of the two groups
satisfaction based on a course that has been peer reviewed. Significance was
The second research question, “Is there a significant difference in levels of student
satisfaction between online courses that have not undergone a systematic faculty peer review
process but are affiliated with a peer review program and online courses that have no affiliation
with a peer review program?,” was analyzed using the following procedures:
A t-test analysis determined whether the satisfaction score means of the two groups
satisfaction based on a course that has not been peer reviewed but is offered from an
The third research question, “Which factors of quality instruction most directly relate to
increased levels of student satisfaction in online courses that have been faculty peer reviewed?,”
relationship between peer review and student satisfaction with each of the five QM
48
The fourth research question, “Which factors of quality instruction most directly relate to
increased levels of student satisfaction in online courses that have not been faculty peer reviewed
but are affiliated with a peer review program?,” was analyzed using the following procedure:
relationship between being subscribed to peer review tools, trainings, and materials
(even without a formal peer review) and student satisfaction with each of the five QM
It was hypothesized that online courses that had undergone a formal peer review, as well
as courses that had not undergone a formal peer review but were offered at institutions that were
subscribed to peer review tools, trainings, and materials, would receive higher student
satisfaction ratings than online courses that had not undergone a formal peer review and were
Course designer data was collected as part of a descriptive research analysis pertaining to
the level of peer review used in the design of each online course.
Chapter Summary
Aman’s (2009) satisfaction questionnaire was used to obtain data regarding student
satisfaction in online courses that had been faculty peer reviewed and online courses that had not
been faculty peer reviewed. Institutions were contacted requesting participation and those who
agreed to participate were included in the sample for this study. The hypotheses were tested
through a one-way ANOVA, a t-test analysis, a simple linear regression analysis, and a Pearson
correlation analysis.
49
CHAPTER 4
RESULTS
Introduction
This chapter analyzes the response rates, demographics, and research questions
associated with the study. Qualitative data obtained from the QM Institutional Representatives
Response Rates
The “recognized” category included participant solicitation from 334 students enrolled in
eight online courses across four institutions. Of those, 22 students chose to participate and were
included in the sample for the “recognized” category. The “subscribed” category included
participant solicitation from 764 students enrolled in 26 online courses across three institutions.
Of those, 48 students chose to participate and were included in the sample for the “subscribed”
category. The “non-affiliated” category included participant solicitation from 676 students
enrolled in 24 online courses across three institutions. Of those, 87 students chose to participate
and were included in the sample for the “non-affiliated” category. This comes to a total of 157
responses out of 1,774 solicited, which is a 9% student response rate. This low response rate
chose to participate in the qualitative course designer survey for the “recognized” category.
Three course designers (out of 26 solicited) chose to participate in the qualitative course designer
50
survey for the “subscribed category. Three course designers (out of 24 solicited) chose to
participate in the qualitative course designer survey for the “non-affiliated” category. This comes
to a total of 7 responses out of 54 solicited, which is a 13% course designer response rate. The
low participant rate reduces the level of significance for this study.
Demographics
The following demographic variables were included in this study: age, gender, comfort
with distance learning, and number of prior online courses taken. As shown in Table 4, the age
range of participants was 20 years old to 58 years old, with a mean age of 33.87. The gender was
split between 107 female respondents and 49 male respondents, with one respondent’s gender
unknown. Using a Likert-type scale of 1 to 5 (with one being very uncomfortable and 5 being
very comfortable), the reported comfort level with distance education was varied with a mean of
3.54. This diversity may be attributed to the range of previous online courses taken, which was
Table 4
Demographic Statistics
51
Gender
A one-way ANOVA determined if there was a mean difference between gender and
overall student satisfaction in their online course. Table 5 shows that overall satisfaction was
reported slightly higher in females (M = 3.98, SD = 1.258) than in males (M = 3.75, SD = 1.102).
The differences between the means were non-significant at the .05 level (p = .300), as shown in
Table 6.
Table 5
Gender Descriptives
Table 6
Age
A one-way ANOVA determined if there was a mean difference between age and overall
student satisfaction in their online course. The results of this factor were categorized into six
groups: the first group included results where age was not reported, the second group included
52
results reported as less than 25 years old, the third group included results reported as between 25-
35 years old, the fourth group included results reported as between 35-45 years old, the fifth
group included results reported as between 45-55 years old, and the sixth group included results
reported as 55 years of age or older. Data in Table 7 indicate that overall satisfaction was
reported higher in the age groups 25 to 35 years (M = 4.21, SD = .925) and 55 years or older (M
= 4.20, SD = .447) than in the other age groups: less than 25 years (M = 3.73, SD = 1.202), 35 to
3.29, SD = 1.704). Table 8, however, shows the differences between the means were non-
Table 7
Age Descriptives
Table 8
53
Prior Online Courses Taken
A one-way ANOVA determined if there was a mean difference between number of prior
online courses taken and overall student satisfaction in their online course. The results of this
factor were categorized into two groups: the first group included results reported as less than 10
previous online courses; and the second group included results reported as 10 or more previous
online courses. Overall satisfaction was reported slightly higher from students who had taken 10
or more previous online courses (M = 3.94, SD = 1.282) than from students who had taken less
than 10 previous online courses (M = 3.88, SD = 1.150), as shown in Table 9. Table 10 indicates
the differences between the means were non-significant at the .05 level (p = .744).
Table 9
Table 10
54
Comfort with Distance Learning
A one-way ANOVA determined if there was a mean difference between comfort with
distance learning and overall student satisfaction in their online course. The results of this factor
were categorized into two groups: the first group included results reported as very
uncomfortable, uncomfortable, and neutral; and the second group included results reported as
comfortable and very comfortable. Table 11 showed overall satisfaction was reported higher
from students who were reportedly very comfortable and comfortable with distance learning (M
= 4.06, SD = 1.131) than from students who were neutral, uncomfortable, or very uncomfortable
with distance learning (M = 3.58, SD = 1.323). The differences between the means were
Table 11
95% Confidence
Interval for Mean
Lower Upper
N M SD Std. Error Bound Bound Minimum Maximum
Very 45 3.58 1.323 .197 3.18 3.98 1 5
Uncomfortable,
Uncomfortable,
Neutral
Comfortable, Very 96 4.06 1.131 .115 3.83 4.29 1 5
Comfortable
Table 12
55
Background Data
Descriptive statistics were calculated to assess background information about the online
course, including course introduction, technology support, and student support, which were
measured in questions 2, 3, and 4. According to the reported results in Table 13, most
participants agreed that a clear course introduction was available at the beginning of the course
(M = 4.19, SD = .956), technology support was available for using the online features of the
course (M = 3.96, SD = .992), and student support was available in using the online format of the
Table 13
N Minimum Maximum M SD
Q2 141 1 5 4.19 .956
Q3 139 1 5 3.96 .992
Q4 140 1 5 3.89 1.025
Research Question 1
The first research question asked, “Is there a significant difference in levels of student
satisfaction between online courses that have undergone a systematic faculty peer review process
and online courses that have not undergone a systematic faculty peer review process?” A t-test
analysis determined if the satisfaction score means of the peer reviewed group (“recognized”)
were different from the two groups that had not been peer reviewed (“subscribed” and “non-
affiliated”).
56
t Test Overall Satisfaction: Peer Reviewed versus Not Peer Reviewed
Statistical significance of student overall satisfaction between the peer reviewed group
and the non-peer reviewed group was determined with a two-tailed independent sample t test
with a significance measured at p < .05. Peer review was identified as the independent variable,
while overall satisfaction was defined as the dependent variable. Using a Likert-type scale of 1-5
(with 1 being strongly disagree and 5 being strongly agree) the mean overall satisfaction for the
peer reviewed group was higher (M = 4.14, SD = .910) than the mean overall satisfaction for the
two groups that were not peer reviewed (M = 3.87, SD = 1.256), as shown in Table 14. Because
the Levene’s test for equality of variances, shown in Table 15, showed a sig. value of .201,
which is greater than .05, equal variances are assumed (Holcomb, 2006). The difference between
the means was determined to be non-significant at the .05 level (p = .337), as shown in Table 16.
Table 14
Table 15
Levene’s Test for Equality of Variances for Overall Satisfaction in Peer Reviewed Courses
F Sig.
Q30 Equal variances assumed 1.651 .201
Equal variances not assumed
57
Table 16
t-Test Mean of QM Factor Means: Peer Reviewed versus Not Peer Reviewed
Because the t test for overall satisfaction was non-significant, an additional t-test analysis
was calculated based on the mean of QM factor means. Statistical significance of the mean of
QM factor means between the peer reviewed group and the non-peer reviewed group was
determined with a two-tailed independent sample t test with a significance measured at p < .05.
Peer review was identified as the independent variable, while the mean of QM factor means was
defined as the dependent variable. Using a Likert-type scale of 1-5 (with 1 being strongly
disagree and 5 being strongly agree) the mean of QM factor means for the peer reviewed group
was higher (M = 4.2440, SD = .56130) than the mean of QM factor means for the two groups
that were not peer reviewed (M = 4.1266, SD = .75574), as shown in Table 17. Equal variances
were assumed, as shown in Table 18. Table 19 showed the difference between the means was
58
Table 17
Table 18
Levene’s Test for Equality of Variances for Mean of QM Factor Means in Peer Reviewed
Courses
F Sig.
Mean of QM Factor Means Equal variances assumed .465 .497
Equal variances not assumed
Table 19
59
Simple Linear Regression Analysis: Significance of Peer Review in Predicting
Overall Student Satisfaction
A simple linear regression was calculated to predict overall student satisfaction based on
a course that has been peer reviewed. Peer review was identified as the independent variable,
while overall satisfaction was identified as the dependent variable. Tables 20, 21, and 22 showed
the regression equation was not significant (p = .337), which indicated that peer review cannot be
Table 20
Table 21
Table 22
Standardized
Unstandardized Coefficients Coefficients
Model β Std. Error β t Sig.
1 (Constant) 4.143 .265 15.654 .000
Peer Reviewed -.276 .287 -.081 -.963 .337
60
Research Question 2
The second research question asked: “Is there a significant difference in levels of student
satisfaction between online courses that have not undergone a systematic faculty peer review
process but are affiliated with a peer review program and online courses that have no affiliation
with a peer review program?” A two-tailed independent sample t test with a significance
measured at p < .05 determined if the satisfaction score mean of the group that had not
undergone a systematic faculty peer review process, but was affiliated with a peer review
program (“subscribed”), was different from the group that had no affiliation with a peer review
program (“non-affiliated).
Statistical significance of student overall satisfaction between the group that has not
undergone a systematic faculty peer review process but is affiliated with a peer review program
(“subscribed”) and the group that has no affiliation with a peer review program (“non-affiliated”)
was determined with a two-tailed independent sample t test with a significance measured at p <
.05. Subscription to a peer review program was identified as the independent variable, while
overall satisfaction was defined as the dependent variable. Using a Likert-type scale of 1-5 (with
1 being strongly disagree and 5 being strongly agree), the mean overall satisfaction for the group
that is subscribed to a peer review program was actually lower (M = 3.73, SD = 1.245) than the
mean overall satisfaction for the group that had no affiliation with a peer review program (M =
3.94, SD = 1.264), as shown in Table 23. Equal variances were assumed, as shown in Table 24.
Table 25 showed the difference between the means was non-significant at the .05 level (p =
.399).
61
Table 23
Table 24
Levene’s Test for Equality of Variances for Overall Satisfaction in Subscribed Courses
F Sig.
Q30 Equal variances assumed .019 .891
Equal variances not assumed
Table 25
Equal variances not -.851 82.173 .397 -.205 .241 -.684 .274
assumed
62
t-Test Mean of QM Factor Means: Subscribed versus Non-Affiliated
Since the t test for overall satisfaction was non-significant, an additional t-test analysis
was calculated based on the mean of QM factor means. Statistical significance of the mean of
QM factor means between the subscribed group and the non-affiliated group was determined
with a two-tailed independent sample t test with a significance measured at p < .05. Having a
subscription to peer review tools, trainings, and materials was identified as the independent
variable, while the mean of QM factor means was defined as the dependent variable. Using a
Likert-type scale of 1-5 (with 1 being Strongly Disagree and 5 being Strongly Agree), data in
Table 26 indicated the mean of QM factor means for the subscribed group was actually lower (M
= 4.1085, SD = .61886) than the mean of QM factor means for the non-affiliated group (M =
4.1369, SD = .82758). Equal variances were assumed, as shown in Table 27. Table 28 showed
the difference between the means was non-significant at the .05 level (p = .849).
Table 26
Table 27
Levene’s Test for Equality of Variances for Mean of QM Factor Means in Subscribed Courses
F Sig.
Mean of QM Factor Means Equal variances assumed .695 .406
Equal variances not assumed
63
Table 28
A simple linear regression was calculated to predict overall student satisfaction based on
a course that is offered at an institution that is subscribed to peer review tools, trainings, and
materials; although the course has not been formally peer reviewed. Subscription was identified
as the independent variable, while overall satisfaction was identified as the dependent variable.
Tables 29, 30, and 31 showed the regression equation was not significant (p = .399), which
indicated that subscribing to a peer review program cannot be used to predict overall student
satisfaction.
Table 29
64
Table 30
Table 31
Standardized
Unstandardized Coefficients Coefficients
Model β Std. Error β t Sig.
1 (Constant) 3.732 .196 18.995 .000
Subscribed .205 .242 .078 .847 .399
Research Question 3
The third research question asked, “Which factors of quality instruction most directly
relate to increased levels of student satisfaction in online courses that have been faculty peer
reviewed?” A correlation analysis determined if there was a relationship between peer review
and student satisfaction with each of the five QM factors included in this study.
A Pearson correlation coefficient was calculated for the relationship between peer review
and student satisfaction with each of the five factors included in this study: learning outcomes,
assessment and measurement, learning resources and materials, learner interactions, and online
course technology. Table 32 showed that all correlations were non-significant at the .05 level (p
65
= .485, p = .221, p = .415, p = .787, p = .564), indicating that peer review is not related to student
Table 32
Learning Online
Learning Assessment Resources Learner Course Peer
Outcomes Measurement Materials Interactions Technology Reviewed
Learning Pearson 1 .815** .821** .699** .713** -.060
Outcomes Correlation
Sig. .000 .000 .000 .000 .485
(2-tailed)
N 139 135 137 139 139 139
Assessment Pearson .815** 1 .786** .673** .641** -.106
Measurement Correlation
Sig. .000 .000 .000 .000 .221
(2-tailed)
N 135 136 134 136 136 136
Learning Pearson .821** .786** 1 .679** .758** -.070
Resources Correlation
Materials Sig. .000 .000 .000 .000 .415
(2-tailed)
N 137 134 138 138 138 138
Learner Pearson .699** .673** .679** 1 .611** -.023
Interactions Correlation
Sig. .000 .000 .000 .000 .787
(2-tailed)
N 139 136 138 140 140 140
Online Pearson .713** .641** .758** .611** 1 -.049
Course Correlation
Technology Sig. .000 .000 .000 .000 .564
(2-tailed)
N 139 136 138 140 140 140
Peer Pearson -.060 -.106 -.070 -.023 -.049 1
Reviewed Correlation
Sig. .485 .221 .415 .787 .564
(2-tailed)
N 139 136 138 140 140 157
66
Research Question 4
The fourth research question asked, “Which factors of quality instruction most directly
relate to increased levels of student satisfaction in online courses that have not been faculty peer
reviewed but are affiliated with a peer review program?” A correlation analysis determined if
there was a relationship between being subscribed to peer review tools, trainings, and materials
and student satisfaction with each of the five QM factors included in this study.
A Pearson correlation coefficient was calculated for the relationship between peer review
and student satisfaction with each of the five factors included in this study: learning outcomes,
assessment and measurement, learning resources and materials, learner interactions, and online
course technology. Table 33 showed that all correlations were non-significant at the .05 level (p
= .565, p = .866, p = .950, p = .913, p = .997), indicating that subscribing to a peer review
program is not related to student perceived satisfaction with each of the five factors.
67
Table 33
Learning Online
Learning Assessment Resources Learner Course
Outcomes Measurement Materials Interactions Technology Subscribed
Learning Pearson Correlation 1 .815** .821** .699** .713** .054
Outcomes Sig. (2-tailed) .000 .000 .000 .000 .565
N 139 135 137 139 139 118
Assessment Pearson Correlation .815** 1 .786** .673** .641** .016
Measurement Sig. (2-tailed) .000 .000 .000 .000 .866
N 135 136 134 136 136 116
Learning Pearson Correlation .821** .786** 1 .679** .758** .006
Resources Sig. (2-tailed) .000 .000 .000 .000 .950
Materials N 137 134 138 138 138 117
Learner Pearson Correlation .699** .673** .679** 1 .611** .010
Interactions Sig. (2-tailed) .000 .000 .000 .000 .913
N 139 136 138 140 140 119
Online Pearson Correlation .713** .641** .758** .611** 1 .000
Course Sig. (2-tailed) .000 .000 .000 .000 .997
Technology N 139 136 138 140 140 119
Subscribed Pearson Correlation .054 .016 .006 .010 .000 1
Sig. (2-tailed) .565 .866 .950 .913 .997
N 118 116 117 119 119 135
Qualitative Data
Results of the “recognized” course designer survey indicated that the one institution that
responded had online courses formally recognized by the QM program and used the rubric,
faculty development, the course development checklist, and QM peer review tools. None of the
seven institutions reported using the self-study tool. Results of the “subscribed” course designer
survey indicated that all three of the institutions that are subscribed to the Quality Matters
program used the QM rubric to guide course design, while two of those (66.7%) used the course
development checklists, and one (33.3%) used the faculty development tool. Zero participants
indicated using the self-study tool. When asked how the QM rubric was being used, all seven
participants reported using the rubric for improvement in existing courses, while only one
(33.3%) reported using the rubric as a blueprint for designing new courses. While two (66.7%) of
68
the institutions agreed that a subscription to the Quality Matters Program required all courses at
their institutions to use the available tools, one (33.3%) declared that these tools were not
required of every online course at their institution. Results of the “non-affiliated” course designer
survey indicated that none of the three institutions were subscribed to any type of peer review
program.
Chapter Summary
This chapter provided a description of the response rate and demographics of participants
in the study. The first two research questions were analyzed via a t-test analysis and a simple
linear regression analysis, while the last two research questions were analyzed via a Pearson
correlation analysis. Finally, the qualitative data collected from the QM Institutional
69
CHAPTER 5
Introduction
This chapter concludes the study summarizing the purpose, framework, and research
recommendations for practices, limitations of the study, and recommendations for future
The purpose of this study was to examine student perceptions of quality and satisfaction
in regionally accredited online courses. A review of the literature suggested there may be
specific factors in online courses that increase students’ perceived quality of that course (Ross et
al., 2002), which, in turn, would increase students’ perceived satisfaction with the course
(McGorry, 2003; Sloan Consortium, 2011; Sun et al., 2008; Zhao, 2008). The factors identified
as promoting quality vary depending on the researcher, including frameworks from Best
Practices (WCET, 2009), Sloan-C (Sloan-C, 2011), QM (Quality Matters, 2010), and U21G
(Chua & Lam, 2007). The factors identified as promoting satisfaction also vary, including
frameworks from Seven Principles (Chickering & Gamson, 1983), Sloan-C (Sloan-C, 2011),
TAM (Arbaugh, 2000), Determinants of Learning Effectiveness (Piccoli, Ahmad, & Ives, 2001),
ADL Network (Artino, 2008), Sun, Tsai, Finger, Chen, and Yeh Framework (Sun et al., 2008),
70
The Quality Matters (QM) program was chosen as the framework for this study. QM is a
peer-based approach structured around eight general standards: course overview and
materials, learner interaction and engagement, course technology, learner support, and
QM focuses on evaluating the quality of individual online courses, rather than overall programs
that have undergone a systematic faculty peer review process and online courses that have not
that have not undergone a systematic faculty peer review process but are affiliated with a peer
review program and online courses that have no affiliation with a peer review program?
3. Which factors of quality instruction most directly relate to increased levels of student
4. Which factors of quality instruction most directly relate to increased levels of student
satisfaction in online courses that have not been faculty peer reviewed but are affiliated with a
71
Discussion of Findings
Demographics
Age, gender, comfort with distance learning, and previous online courses taken were
included in this study as demographic variables. The descriptive statistics showed that the mean
age of participants was 33.87, while the gender was primarily female (68%). Students’ comfort
with distance learning was varied, possibly because the number of previous online courses
ranged from 0-42. A one-way ANOVA was calculated for each of these demographic variables
to determine if there was a mean difference between the variable and overall satisfaction
reported. The analyses showed that gender, age, and prior online courses taken were non-
significant to student perceptions of overall satisfaction of their online course. Student comfort
with distance learning, however, was shown to have a significant relationship with student
The significant results for the comfort with distance learning factor were supported by the
literature. Abdous and Yen (2010) agreed that comfort with distance learning was a useful
predictor of overall satisfaction, although it should be noted that the significance of the
contribution was weak. Lim (2001) also found that students who had more years of experience
with computer use, especially in an online class, reported higher levels of satisfaction with their
online course. Thurmond et al. (2002) also found a positive relationship between student comfort
The review of literature supported the researcher’s finding by indicating that age was
non-significant to student satisfaction. Richardson and Swan (2003) found that age was not
significant to students’ perceptions of an online course. Thurmond et al. (2002) found that age
does not help predict a student’s level of satisfaction. Hong (2002) also found no relationship
72
between student age and student satisfaction in an online course. The review of literature
regarding gender as a predictor for student satisfaction in an online course was varied.
Richardson and Swan (2003) found gender to be positively related to student satisfaction in an
online course. Neither Hong (2002) nor Lim (2001) found a statistical significance between
gender and student satisfaction in an online course. The literature mostly indicated that the
number of prior online courses taken was non-significant to student satisfaction with online
education (Richardson & Swan, 2003; Thurmond et al., 2002); although, Arbaugh and Duray
(2002) argued that students with more experience in online learning showed a positive
Research Question 1
satisfaction between online courses that have undergone a systematic faculty peer review process
and online courses that have not undergone a systematic faculty peer review process?” A two-
tailed independent samples t-test analysis determined that the overall satisfaction mean of the
peer reviewed group (“recognized”) was higher (M = 4.14, SD = .910) than the overall
satisfaction mean of the two groups that had not been peer reviewed (“subscribed” and “non-
affiliated”) (M = 3.87, SD = 1.256). This finding was determined to be non-significant at the .05
level (p = .337). Next, a t-test analysis was calculated based on the mean of QM factor means,
which included the means from all questions included in the five factors of this study: learning
outcomes, assessment and measurement, learning resources and materials, learner interactions,
and course technology. A two-tailed independent samples t-test analysis determined that the
mean of QM factor means for the peer reviewed group was higher (M = 4.2440, SD = .56130)
73
than the mean of QM factor means for the two groups that were not peer reviewed (M = 4.1266,
SD = .75574). This finding was determined to be non-significant at the .05 level (p = .509).
Although non-significant, these results are supported by Ralston-Berg and Nath (2008)
who found that students who claimed higher perceived satisfaction in an online course was
positively correlated with all Quality Matters features embedded in the course design: learning
engagement, course technology, learner support, and accessibility. Artino’s (2011) study also
found that students viewed almost all of the QM standards as valuable to their online course.
Swan et al. (2011) and Aman (2009) also found peer review to be significantly related to student
A simple linear regression analysis was also calculated to predict overall student
satisfaction based on a course that has been peer reviewed. The regression analysis was non-
significant at the .05 level (p = .509), indicating that peer review cannot be used to predict
overall student satisfaction. This contradicts the researcher’s hypothesis that courses that have
undergone a formal peer review will lead to higher student satisfaction than courses that have not
undergone a formal peer review. This also contradicts Aman’s (2009) study, which found that
undergoing a formal peer review would lead to higher student perceived satisfaction in an online
course, and Swan et al. (2011) who found that a formal peer review would lead to higher student
Research Question 2
satisfaction between online courses that have not undergone a systematic faculty peer review
74
process but are affiliated with a peer review program and online courses that have no affiliation
with a peer review program?” A two-tailed independent samples t-test analysis determined that
the overall satisfaction mean of the group that had not undergone a systematic faculty peer
review process but were subscribed to a peer review program’s tools, trainings, and materials
(“subscribed”) was actually lower (M = 3.73, SD = 1.245) than the overall satisfaction mean of
the group with no affiliation to a peer review program (“non-affiliated”) (M = 3.94, SD = 1.264).
The difference was determined to be non-significant at the .05 level (p = .399). Next, a t-test
analysis was calculated based on the mean of QM factor means, which included the means from
all questions included in the five factors of this study: learning outcomes, assessment and
measurement, learning resources and materials, learner interactions, and course technology. A
two-tailed independent samples t-test analysis determined that the mean of QM factor means for
the subscribed group was lower (M = 4.1369, SD = .82758) than the mean of QM factor means
for the non-affiliated group (M = 4.1369, SD = .82758). The difference was determined to be
The findings did not support the hypothesis for this study. The findings also contradicted
Ralston-Berg and Nath (2008), Artino (2011), Swan et al. (2011), and Aman (2009), who found
peer review tools, trainings, and materials to be positively related to increased student
Finally, a simple linear regression analysis was calculated to predict overall student
satisfaction based on a course offered from an institution that is subscribed to a peer review
program, although the course had not undergone a formal peer review process. The regression
analysis was non-significant at the .05 level (p = .849), indicating that subscription to peer
review tools cannot be used to predict overall student satisfaction. This contradicted the
75
researcher’s hypothesis that being subscribed to a peer review program, even without undergoing
a formal peer review, would lead to higher student satisfaction with their online course. The
findings also contradicted Ralston-Berg and Nath (2008), Artino (2011), Swan et al. (2011), and
Aman (2009), who found peer review tools, trainings, and materials to be positively related to
Research Question 3
Research Question 3 asked, “Which factors of quality instruction most directly relate to
increased levels of student satisfaction in online courses that have been faculty peer reviewed?”
A Pearson correlation coefficient was calculated to determine the relationship between peer
review and each of the five factors included in this study: learning outcomes, assessment and
measurement, learning resources and materials, learner interactions, and online course
technology. All correlations were non-significant at the .05 level (p = .485, p = .221, p = .415,
p = .787, p = .564), indicating that peer review is not related to student perceptions of satisfaction
with any of the five factors. This contradicts previous literature regarding peer review and online
course quality. Ralston-Berg and Nath (2008) and Artino (2011) both found all five of these
factors to be positively related to increased student satisfaction in an online course. Aman (2009)
actually found learning outcomes, assessment and measurement, learning resources and
materials, learner interactions, and online course technology to all be positively related to
increased student satisfaction with an online course, with learning resources and materials being
76
Research Question 4
Research Question 4 asked, “Which factors of quality instruction most directly relate to
increased levels of student satisfaction in online courses that have not been faculty peer reviewed
but are affiliated with a peer review program?” A Pearson correlation coefficient was calculated
to determine the relationship between subscription to peer review tools, trainings, and materials
and each of the five factors included in this study: learning outcomes, assessment and
measurement, learning resources and materials, learner interactions, and online course
technology. All correlations were non-significant at the .05 level (p = .565, p = .866, p = .950, p
= .913, p = .997), indicating that a subscription to peer review tools is not related to student
perceptions of satisfaction with any of the five factors. This contradicts Ralston-Berg and Nath
(2008), Artino (2011), Swan et al. (2011), and Aman (2009), who found peer review programs to
Conclusions
Three major conclusions can be drawn from this study. First, the results indicated that a
formal peer review of an online course may lead to higher student satisfaction of that course. The
results of this study were non-significant, most likely due to the low response rate of participants.
However, since the mean overall satisfaction was higher in the groups that had been peer
reviewed, this is still a good sign that peer review of online course design may lead to higher
student satisfaction in a course. More research is needed to validate this finding. Second, results
of the t-test analyses indicated higher satisfaction rates of those in the group with no affiliation to
a peer review program (“non-affiliated”) than of those in the group affiliated with a peer review
program, although no systematic faculty peer review had been completed (“subscribed”).
77
Although these results were non-significant, this poses a supplementary question concerning
whether or not subscribing to peer review program tools, trainings, and materials (without a
formal peer review) provides any benefit to the quality of online course design. Third, the
Pearson correlation analyses showed that all correlations were non-significant, indicating that (1)
peer review is not related to student perceived satisfaction with each of the five quality factors
and (2) subscribing to faculty peer review tools, trainings, and materials without a formal peer
review is not related to student perceived satisfaction with each of the five quality factors. This is
a contradiction of the literature, which showed that peer review was a predictor for higher levels
The high level of statistical insignificance found reduces the value of the researcher’s
findings. Because so many of the findings were determined non-significant, the researcher
concluded that the sample size was insufficient and that further research is needed (Johnson,
1999).
formal peer review of online courses being developed because this research showed that the
mean overall student satisfaction of a peer reviewed online course was higher than the mean
overall student satisfaction for the two groups that had not been peer reviewed. Although this
finding was non-significant, it still indicated that a formal peer review of an online course may
be beneficial to student satisfaction of the course. This study did not find a subscription to peer
review tools, trainings, and materials to be beneficial to student satisfaction in an online course.
Therefore, it is recommended that those who are subscribed to peer review tools, trainings, and
78
materials follow through with a formal peer review process to maximize the quality and student
Limitations
Limitations of this study should be taken into consideration with regard to the findings
and conclusions. There were four limitations identified as significant to this study, which are
The first limitation of the study was the population from which participation was invited.
The setting of this study included only senior institutions accredited by the Southern Association
of Colleges and Schools in which upper-level and/or graduate courses were offered in an online
format. Additionally, only 10 schools were selected from each study category (recognized,
The second limitation of the study was the sample size of participants. Although 1,774
students were solicited for participation, only 157 chose to respond. Additionally, out of 54 QM
Institutional Representatives and course designers solicited for participation, only 7 chose to
The third limitation of the study was in the specifications of the survey. The online
survey did not collect IP addresses, names, or any other identifiable information of its
participants. Although follow-up research invitations were sent out at 2-week intervals, it was
impossible to personalize each follow-up invitation due to the anonymity of the survey data
collected.
79
Finally, the fourth limitation was identified as the structure of the student satisfaction
survey. Because all questions asked in this study were closed-ended questions, it was impossible
One recommendation for future research includes expanding the study to include all QM
“recognized” courses and all institutions who are currently “subscribed” to QM, which, in turn,
would call for a greater population from the “non-affiliated” category. Expanding the population
from which participants were solicited may result in a larger sample size.
A second recommendation for future research includes using a different technique for
was implemented for data collection in this study. Because responses were collected
participants. Wright (2005) suggested offering some type of financial incentive to increase
response rates, although implementing an incentive would require the questionnaire to obtain
identifiable information from its participants. Shannon and Bradshaw (2002) indicated that
response rates from electronic surveys are typically lower due to participants’ concern of
anonymity and confidentiality. Evans and Mathur (2005) suggested that low response rates to
online surveys may be a result of individuals perceiving the research invitation as internet spam.
Aman (2009) suggested that securing faculty acceptance of the study might lead to higher
student response rates of online surveys, as faculty may encourage their students to participate.
A third recommendation for future research is for a modification of the student survey
80
open-ended questions, such as which online course the survey responses represent. This
recommendation came from an anonymous survey participant who contacted the researcher with
significance of being subscribed to peer review tools, trainings, and materials. The researcher
expected to find that courses offered from institutions who were “subscribed” to peer review
tools, trainings, and materials (even without a formal peer review) would result in higher
satisfaction rates than courses offered from institutions that have no affiliation with a peer review
process. The analysis showed otherwise, although it should be noted that the analysis was
determined non-significant.
Concluding Remarks
The purpose of this study was to determine if a formal peer review process, or if
subscription to peer review tools, trainings, and materials (even without a formal peer review),
would affect student perceptions of quality and satisfaction in online education. Using data
collected through a Likert-type scale online questionnaire, a t-test analysis determined that online
courses that had undergone a formal peer review resulted in a higher mean overall satisfaction
than online courses that had not undergone a formal peer review; however this difference was
81
found to be non-significant. Additionally, a simple linear regression analysis determined that
On the other hand, a t-test analysis determined that online courses that had not undergone
a formal peer review but that were subscribed to peer review tools, trainings, and materials,
resulted in a lower mean overall satisfaction than online courses that had no affiliation with a
peer review program. This difference, too, was found to be non-significant. A simple linear
regression analysis determined that subscription to peer review tools, trainings, and materials
This study also considered which factors of quality instruction were related to overall
student satisfaction in both courses that have undergone a formal peer review and in courses that
have not undergone a formal peer review but that are offered at institutions with a subscription to
peer review tools, trainings, and materials. A Pearson correlation determined that satisfaction
with the five factors included in this study were not significantly related to peer review or to a
Finally, a one-way ANOVA determined whether there was a relationship between four
demographic variables (age, gender, comfort with distance learning, and number of previous
online courses taken) and overall student satisfaction in their online course. This analysis showed
no statistical significance in age, gender, or number of previous online courses taken. However, a
statistical significance was found between overall satisfaction and student comfort with distance
learning. This analysis showed that students who were reportedly comfortable or very
comfortable with distance learning were more satisfied overall with their online course.
82
Implications from this study were difficult to project because most of the results were
formal peer review of an online course would lead to higher student satisfaction of that course.
population/sample size, a revision of the solicitation procedures, and a revision of the survey
instrument to include open-ended questions. Also, it is recommended that further research take
place to determine whether being subscribed to a peer review program (without undergoing a
83
REFERENCES
Abdous, M., & Yen, C. (2010). A predictive study of learner satisfaction and outcomes in face-
to-face, satellite broadcast, and live video-streaming learning environments. Internet and
Higher Education, 13, 248-257.
Achtemeier, S. D., Morris, L. V., & Finnegan, C. L. (2003). Considerations for developing
evaluations of online courses. Journal for Asynchronous Learning Networks, 7(1), 1-13.
Allen, I. E., & Seaman, J. (2010). Learning on demand: Online education in the United States,
2009. Needham, MA: Babson Survey Research Group.
Alstete, J. W., & Beutell, N. J. (2004). Performance indicators in online distance learning
courses: A study of management education. Quality Assurance in Education, 12(1), 6-14.
Aman, R. R. (2009). Improving student satisfaction and retention with online instruction through
systematic faculty peer review of courses. Dissertation. Oregon State University.
An, Y. (2010). Scaffolding wiki-based, ill structured problem solving in an online environment.
Journal of Online Learning and Teaching, 6(4), 1-12.
Arbaugh, J. B. (2000). Virtual classroom characteristics and student satisfaction with internet-
based MBA courses. Journal of Management Education, 24(1), 32-54.
Arbaugh, J. B. (2001). How instructor immediacy behaviors affect student satisfaction and
learning in web-based courses. Business Communication Quarterly, 64(4), 42-54.
Arbaugh, J. B., & Duray, R. (2002). Technological and structural characteristics, student learning
and satisfaction with web-based courses: An exploratory study of two on-line MBA
programs. Management Learning, 33(3), 331-347.
Arnold, N., & Paulus, T. (2010). Using a social networking site for experiential learning:
Appropriating, lurking, modeling and community building. Internet and Higher
Education, 13, 188-196.
Augustsson, G. (2010). Web 2.0, pedagogical support for reflexive and emotional social
interaction among Swedish students. Internet and Higher Education, 13, 197-205.
84
Babb, S., Stewart, C., & Johnson, R. (2010). Constructing communication in blended learning
environments: Students’ perceptions of good practice in hybrid courses. Merlot Journal
of Online Learning and Teaching, 6(4), 839-851.
Baturay, M. H., & Bay, O. F. (2010). The effects of problem-based learning on the classroom
community perceptions and achievement of web-based education students. Computers &
Education, 55, 43-52.
Benson, A. D. (2003). Dimensions of quality in online degree programs. The American Journal
of Distance Education, 17(3), 145-159.
Bickle, M. C., & Carroll, J. C. (2003). Checklist for quality online instruction: Outcomes for
learners, the profession and the institution. College Student Journal, 37.
Billings, D. M. (2000). A framework for assessing outcomes and practices in web-based courses
in nursing. Journal of Nursing Education, 39(2), 60-67.
Billings, D. M., Connors, H. R., & Skiba, D. J. (2001). Benchmarking best practices in web-
based nursing courses. Advancees in Nursing Science, 23(3), 41-52.
Bogdan, R. C., & Biklen, S. K. (1998). Qualitative research for education: An introduction to
theory and methods (3rd ed.). London: Allyn and Bacon.
Burgstahler, S., Corrigan, B., & McCarter, J. (2004). Making distance learning courses
accessible to students and instructors with disabilities: A case study. The Internet and
Higher Education, 7, 233-246.
Burton, J., Lockee, B., & Potter, K. (2010). Examining standards for distance education systems.
11th International Conference on Education Research. Virginia Tech, VA.
Carmichael, P., & Burchmore, H. (2010). Social software and academic practice: Postgraduate
students as co-designers of Web 2.0 tools. Internet and Higher Education, 13, 233-241.
Carr, R., & Hagel, P. (2008). Students’ evaluations of teaching quality and their unit online
activity: An empirical investigation. Retrieved from Proceedings ascilite Melbourne:
http://www.ascilite.org.au/conferences/melbourne08/procs/carr-r.pdf
Carr-Chellman, A., & Duchastel, P. (2000). The ideal online course. British Journal of
Educational Technology, 31(3), 229-241.
Chickering, A. W., & Ehrmann, S. C. (1996, October). Implementing the seven principles:
Technology as lever. Retrieved October 16, 2011, from The Learning Technology (TLT)
Group: http://www.tltgroup.org/programs/seven.html
85
Chickering, A. W., & Gamson, Z. F. (1983). Seven principles for good practice in
undergraduate education. Retrieved January 13, 2009, from http://honolulu.hawaii.edu/
intranet/committees/FacDevCom/guidebk/teachtip/7princip.htm
Chikering, A. W., & Gamson, Z. F. (1987). Seven principles for good practice in undergraduate
education. AAHE Bulletin, 39(7), 3-6.
Chua, A., & Lam, W. (2007). Quality assurance in online education: The universitas 21 global
approach. British Journal of Educational Technology, 38(1), 133-152.
Cohen, N. L., Carbone, E. T., & Beffa-Negrini, P. A. (2011). The design, implementation, and
evaluation of online credit nutrition courses: A systematic review. Journal of Nutrition
Education and Behavior, 43(2), 76-86.
Council for Higher Education Accreditation. (2002). Accreditation and assuring quality in
distance learning. Washington, DC: Author.
Cramer, K. M., Collins, K. R., Snider, D., & Fawcett, G. (2006). Virtual lecture hall for in-class
and online sections: A comparison of utilization, perceptions, and benefits. Journal of
Research on Technology in Education, 38(4), 371-381.
Deubel, P. (2003). Learning from reflections: Issues in building quality online courses. Retrieved
from http://www.ct4me.net/building_online_courses.htm
Dietz-Uhler, B., Fisher, A., & Han, A. (2007-2008). Designing online courses to promote student
retention. J. Educational Technology Systems, 36(1), 105-112.
Dill, D. D., Massy, W. F., Williams, P. R., & Cook, C. M. (1996, Sep/Oct). Accreditation &
academic quality assurance: Can we get there from here? Council for Higher Education
Accreditation, 28(5), 16-24.
Dong, S., Xu, S., & Lu, X. (2009). Development of online instructional resources for earth
system science education: An example of current practice from China. Computers and
Geosciences, 35, 1271-1279.
Dringus, L. P. (2000). Towards active online learning: A dramatic shift in perspective for
learners. The Internet and Higher Education, 2(4), 189-195.
Eaton, J. S. (2001). Distance learning: Academic and political challenges for higher education
accreditation. Retrieved from Council for Higher Education Accreditation:
http://www.chea.org/
86
Eom, S. B., Wen, H. J., & Ashill, N. (2006). The determinants of students’ perceived learning
outcomes and satisfaction in university online education: An empirical investigation.
Decision Sciences Journal of Innovative Education, 4(2), 215-235.
Epper, R. M., & Garn, M. (2003). Virtual college and university consortia. Retrieved from
WCET: http://wcet.wiche.edu/wcet/docs/resources/Virtual_College_University.pdf
Evans, J. R., & Mathur, A. (2005). The value of online surveys. Emerald Group Publishing
Limited, 15(2), 195-219.
Fox, J. (1998). Distance education: is it good enough? The University Concourse, 3(4), 3-5.
Fredericksen, E., Pickett, A., Shea, P., Pelz, W., & Swan, K. (2000). Student satisfaction and
perceived learning with on-line courses: Principles and examples from the SUNY
Learning Network. Retrieved April 23, 2009, from Sloan Consortium:
http://www.aln.org/alnweb/journal/Vol4 issue2/le/
Grandzol, C. J., & Grandzol, J. R. (2010, Summer). Interaction in online courses: More is not
always better. Online Journal of Distance Learning Administration, 8(2), 1-13.
Halic, O., Lee, D., Paulus, T., & Spence, M. (2010). To blog or not to blog: Student perceptions
of blog effectiveness for learning in a college-level course. Internet and Higher
Education, 13, 206-213.
Harasim, L. (2000). Shift happens: Online education as a new paradign in learning. The Internet
and Higher Education, 3, 41-61.
Hatziapostolou, T., & Paraskakis, I. (2010). Enhancing the impact of formative feedback on
student learning through an online feedback system. Electronic Journal of E-Learning,
8(2), 111-122.
Head, R. B., & Johnson, M. S. (2011). Accreditation and its influence on institutional
effectiveness. New Directions for Community Colleges, 153, 37-52.
Herbert, M. (2006). Staying the course: A study in online student satisfaction and retention.
Online Journal of Distance Learning Administration, 9(4), 1-9.
Herrington, A., Herrington, J., Oliver, R., Stoney, S., & Willis, J. (2001). Quality guidelines for
online courses: The development of an instrument to audit online units. In G. Kennedy,
M. Keppell, C. McNaught, & T. Petrovic (Ed.), Meeting at the crossroads: Proceedings
of ASCILITE 2001 (pp. 263-270). Melbourne: The University of Melbourne.
87
Hollandsworth, R. J. (2007). A hybrid approach for online lectures in the business classroom.
TechTrends, 51(4), 39-44.
Hong, K. S. (2002). Relationships between students’ and instructional variables with satisfaction
and learning from a web-based course. Internet and Higher Education, 5, 267-281.
Johnson, S. D., Aragon, S. R., Shaik, N., & Palma-Rivas, N. (2000). Comparative analysis of
learner satisfaction and learning outcomes in online and face-to-face learning
environments. Journal of Interactive Learning Research, 11(1), 29-49.
Johnstone, S. M., & Krauth, B. (1996). Balancing quality and access: Some principles of good
practice for the virtual university. Change, 28(2), 38-41.
Junco, R., Heibergert, G., & Loken, E. (2010). The effect of Twitter on college student
engagement and grades. Journal of Computer Assisted Learning, 1-14.
Kassop, M. (2003, May/June). Ten ways online education matches, or surpasses, face-to-face
learning. Retrieved January 13, 2009, from The Technology Source:
http://66.102.1.104/scholar?hl=en&lr=&q=cache:8JWZOtftfscJ:pegmarc.com/csusb/pdf/
10WaysOnlineEducation.pdf+ten+ways+online+education+matches,+or+surpasses,+face
-to-face+learning
Kear, K., Woodthorpe, J., Robertson, S., & Hutchison, M. (2010). From forums to wikis:
Perspectives on tools for collaboration. Internet and Higher Education, 13, 218-225.
Kim, K. S., & Moore, J. L. (2005, November 7). Web-based learning: Factors affecting
students’ satisfaction and learning experience., 10(11). Retrieved October 14, 2011, from
First Monday: Peer Reviewed Journal on the Internet: http://www.firstmonday.org/htbin/
cgiwrap/bin/ojs/index.php/fm/article/view/1294/1214
Koroghlanian, C. M., & Brinkerhoff, J. (2008). Online students’ technology skills and attitudes
toward online instruction. Journal of Educational Technology Systems, 36(2), 219-244.
Lee, J. (2010). Online support service quality, online learning acceptance, and student
satisfaction. Internet and Higher Education, 13, 277-283.
Legon, R. (2009). The quality matters program. Baltimore, MD: IGI Global.
88
Legon, R., & Runyon, J. (2007). Research on the impact of the quality matters course review
process. 23rd Annual Conference on Distance Teaching and Learning (pp. 1-5).
Milwaukee: Board of Regents, University of Wisconsin.
Lewis, K. O., Baker, R. C., & Britigan, D. H. (2011, Spring). Current practices and needs
assessment of instructors in an online masters degree in education for healthcare
professionals: A first step to the development of quality standards. Journal of Interactive
Online Learning, 10(1), 49-63.
Little, B. B. (2009). The use of standards for peer review of online nursing courses: A pilot
study. Journal of Nursing Education, 48(7), 411-415.
Lorenzo, G., & Moore, J. (2002, November). Five pillars of quality online education. Retrieved
from The Sloan Consortium Report to the Nation: http://sloanconsortium.org/
Maki, R. H., Maki, W. S., Patterson, M., & Whittaker, P. D. (2000). Evaluation of a web-based
introductory psychology course: I. learning and satisfaction in on-line versus lecture
courses. Behavior Research Methods, Instruments, & Computers. 32, pp. 230-239.
Lubbock: Texas Tech University.
Maryland Online, Inc. (2006). Quality matters: Inter-institutional quality assurance in online
learning. Retrieved October 10, 2008, from Quality Matters:
http://www.qualitymatters.org/
MarylandOnline. (2010). Higher Ed Program Subscriptions. Retrieved January 30, 2012, from
Quailty Matters: http://www.qmprogram.org/faq/statewide
McFerrin, K., Duchardt, B., Furr, P. F., Horton, S. G., & Gentry, V. (2009). Quality matters:
designing, implementing, and assessing alternate certification programs. Proceedings of
Society for Information Technology & Teacher Education International Conference (pp.
1-9). Chesapeake, VA: Editlib.
McGivney, V. (2004). Understanding persistence in adult learning. Open Learning, 19(1), 33-46.
Means, B., Toyama, Y., Murphy, R., Bakia, M., & Jones, K. (2010, September). US Department
of Education. Retrieved from Evaluation of Evidence-Based Practices in Online
Learning: http://www2.ed.gov/rschstat/eval/tech/evidence-based-practices/finalreport.pdf
89
Menchaca, M. P., & Bekele, T. A. (2008). Learner and instructor identified success factors in
distance education. Distance Education, 29(3), 231-252.
Meyen, E. L., Lian, C. H., & Tangen, P. (1997). Teaching online courses. Focus on Autism and
Other Developmental Disabilities, 12(3), 166-174.
Meyer, K. (2010). A comparison of Web 2.0 tools in a doctoral course. Internet and Higher
Education, 13, 226-232.
Moore, J. C. (2004). The Sloan consortium quality framework and the five pillars. Retrieved
October 14, 2011, from Sloan Consortium: http://citeseerx.ist.psu.edu/viewdoc/
download?doi=10.1.1.115.4238&rep=rep1&type=pdf
Moore, J. C. (2005, August). A synthesis of Sloan-C effective practices. Retrieved January 11,
2012, from The Sloan Consortium: http://sloanconsortium.org/publications/books/
v9n3_moore.pdf
Moore, M. (1976). Investigation of the interaction between cognitive style of field independence
and attitude to independent study among adult learners who use correspondence
independent study and self-directed independent study. Doctoral Dissertation. University
of Wisconsin-Madison.
Moore, M. G. (1989). Three types of interaction. The American Journal of Distance Education,
3(2), 1-6.
Moore, M. G., & Kearsley, G. (2005). Distance education: A systems view (2nd ed.). Belmont,
CA: Wadsworth Publishing.
Morris, L. V., Wu, S., & Finnegan, C. (2005). Predicting retention in online general education
courses. The American Journal of Distance Education, 19(1), 23-36.
Nash, R. D. (2005, Winter). Course competion rates among distance learners: Identifying
possible methods to improve retention. Retrieved October 31, 2011, from University of
West Georgia, Distance Education Center: http://distance.westga.edu/~distance/
ojdla/winter84/nash84.htm
Neuhauser, C. (2002). Learning style and effectiveness of online and face-to-face instruction.
The American Journal of Distance Education, 16(2), 99-113.
Nichols, M. (2010). Student perceptions of support services and the influence of targeted
interventions on retention in distance education. Distance Education, 31(1), 93-113.
90
North American Council for Online Learning. (2009). National standards for quality online
teaching. Retrieved January 12, 2009, from iNACOL: http://www.inacol.org/
resources/nationalstandards/index.php
Oblinger, D., & Hawkins, B. (2006). The myth about online course development. Educause
Review, 41(1), 14-15.
Page, E. H., Canova, B. S., & Tufarolo, J. A. (1997, July). A case study of verification,
validation and accreditation for advanced distributed simulation. ACM Transactions on
Modeling and Computer Simulation, 7(3), 393-424.
Palomba, C. A., & Banta, T. W. (1999). Assessment essentials. San Francisco: Jossey-Bass.
Patterson, B., & McFadden, C. (2009). Attrition in online and campus degree programs. Online
Journal of Distance Learning Administration, 12(2), 1-11.
Peterson, C. (2003). Bringing ADDIE to life: Instructional design at its best. Journal of
Educational Multimedia and Hypermedia, 12(3), 227-241.
Phipps, R., & Merisotis, J. (2000). Quality on the line: Benchmarking for success in internet-
based distance education. Washington, DC: Institute for Higher Education Policy.
Piccoli, G., Ahmad, R., & Ives, B. (2001). Web-based virtual learning environments: A research
framework and a preliminary assessment of effectiveness in basic IT skills training. MIS
Quarterly, 25(4), 401-426.
Pollacia, L., Russell, J., & Russell, B. (2009). Developing an online program in computer
information systems using quality matters standards. Journal of Online Learning and
Teaching, 5(2), 1-9.
Preston, G., Phillips, R., Gosper, M., McNeill, M., Woo, K., & Green, D. (2010). Web-based
lecture technologies: Highlighting the changing nature of teaching and learning.
Australasian Journal of Educational Technology, 26(6), 717-728.
Puzziferro, M., & Shelton, K. (2008). A model for developing high-quality online courses:
Integrating a systems approach with learning theory. Journal of Asynchronous Learning
Networks, 12(3-4), 119-136.
Quality Matters. (2010). Program rubric. Retrieved October 2, 2011, from Quality Matters:
http://www.qmprogram.org/rubric
Ralston-Berg, P., & Nath, L. (2008). What makes a quality online course? The student
perspective. 24th Annual Conference on Distance Teaching & Learning (pp. 1-5).
Madison: WI.
Richardson, J. C., & Swan, K. (2003). Examining social presence in online courses in relation to
students’ perceived learning and satisfaction. Journal for Asynchronous Learning
Networks, 7(1), 68-88.
91
Robels, M., & Braathen, S. (2002). Online assessment techniques. The Delta Pi Epsilon Journal,
44(1), 39-49.
Roblyer, M. D., McDaniel, M., Webb, M., Herman, J., & Witty, J. V. (2010). Findings on
Facebook in higher education: A comparison of college faculty and student uses and
perceptions of social networking sites. Internet and Higher Education, 13, 134-140.
Rodriguez, M. C., Ooms, A., & Montanez, M. (2008). Students’ perceptions of online-learning:
Quality given comfort, motivation, satisfaction, and experience. Journal of Interactive
Online Learning, 7(2), 105-125.
Ross, K. R., Batzer, L., & Bennington, E. (2002). Quality assurance for distance education: A
faculty peer review process. TechTrends, 46(5), 48-52.
Rossin, D., Young, K. R., Klein, B. D., & Guo, Y. M. (2009). The effects of flow on learning
outcomes in an online information management course. Journal of information systems
education, 20(1), 87-98.
Rourke, L., Anderson, T., Garrison, D. R., & Archer, W. (2001). Assessing social presence in
asynchronous text-based computer conferencing. Journal of Distance Education, 14(2),
1-18.
Rovai, A. P., Ponton, M. K., & Baker, J. D. (2008). Distance learning in higher education: A
programmatic approach to planning, design, instruction, evaluation, and accreditation.
New York: Teachers College Press.
Rutherford, C. (2010, December). Using online social media to support preservice student
engagement. Journal of Online Learning and Teaching, 6(4), 1-11.
Sachdeva, A. K., Pellegrini, C. A., & Johnson, K. A. (2008). Support for simulation-based
surgical education through American College of Surgeons--accredited education
institutes. World Journal of Surgery, 32, 196-207.
Sahin, I., & Shelley, M. (2008). Considering students’ perceptions: The distance education
student satisfaction model. Educational Technology & Society, 11(3), 216-223.
Schmidt, W. C. (1997). World-wide web survey research: Benefits, potential problems, and
solutions. Behavior Research Methods, Instruments, & Computers, 29, 274-279.
92
Schroeder, A., Minocha, S., & Schneider, C. (2010). The strengths, weaknesses, opportunities
and threats of using social software in higher and further education teaching and learning.
Journal of Computer Assisted Learning, 26, 159-174.
Shannon, D. M., & Bradshaw, C. C. (2002). A comparison of the response rate, response time,
and costs of mail and electronic surveys. The Journal of Experimental Education, 70(2),
179-192.
Shattuck, K. (2011, September 28). Interview with director of research at quality matters. (J. M.
Simpson, Interviewer)
Shea, P. J., Pickett, A. M., & Pelz, W. E. (2003). A follow-up investigation of “teaching
presence” in the SUNY learning network. Journal of Asynchronous Learning Networks,
7(2), 61-80.
Shea, P., & Armitage, S. (2002). Guidelines for creating student services online. Retrieved from
WCET: http://wcet.wiche.edu/wcet/docs/beyond/overview.pdf
Sloan Consortium. (2004). Entering the mainstream: The quality and extent of online education
in the United States, 2003 and 2004. Retrieved April 17, 2009, from Sloan Consortium:
https://www.sloan-c.org/survey04.asp
Sloan Consortium. (2011). Effective Practices. Retrieved October 9, 2011, from Sloan
Consortium: http://sloanconsortium.org/effective
Sloan-C. (2011a). Effective practices. Retrieved January 11, 2012, from The Sloan Consortium:
http://sloanconsortium.org/effective
Sloan-C. (2011b). The Sloan Consortium. Retrieved January 11, 2012, from
http://sloanconsortium.org/
Sonwalkar, N. (2007, April 26). A new methodology for evaluation: The pedagogical rating of
online courses. Retrieved from Sloan-C: http://sloanconsortium.org/effective_practices/
new-methodology-evaluation-pedagogical-rating-online-courses
Spatariu, A., Quinn, L. F., & Hartley, K. (2007). A review of research on factors that impact
aspects of online discussions quality. TechTrends, 51(3), 44-48.
93
Steinman, D. (2007). Educational experiences and the online student. TechTrends, 51(5), 46-52.
Sun, P. C., Tsai, R. J., Finger, G., Chen, Y. Y., & Yeh, D. (2008). What drives a successful e-
learning? An empirical investigation of the critical factors influencing learner
satisfaction. Computers & Education, 50, 1183-1202.
Swan, K. (2001). Virtual interaction: Design factors affecting student satisfaction and perceived
learning in asynchronous online courses.. Distance Education, 22(2), 306-331.
Swan, K., Matthews, D., Bogle, L., Boles, E., & Day, S. (2011). Linking online course design
and implementation to learning outcomes: A design experiment. The Internet and Higher
Education, doi: 10.1016/j.iheduc.2011.07.002.
Tait, J. (2004). The tutor/facilitator role in student retention. Open Learning, 19(1), 97-109.
Ternus, M. P., Palmer, K. L., & Faulk, D. R. (2007). Benchmarking quality in online teaching
and learning: A rubric for coruse construction and evaluation. The Journal of Effective
Teaching, 7(2), 51-67.
Thach, E. C., & Murphy, K. L. (1995). Competencies for distance education professionals.
Educational Technology Research and Development, 43(1), 57-79.
Thurmond, V. A., Wambach, K., Connors, H. R., & Frey, B. B. (2002). Evaluation of student
satisfaction: determining the impact of a web-based environment by controlling for
student characteristics. The American Journal of Distance Education, 16(3), 169-189.
Tinto, V. (1975). Dropout from higher education: A theoretical synthesis of recent research.
Review of Educational Research, 45(1), 89-125.
Tomei, L. A. (2006). The impact of online teaching on faculty load: Computing the ideal class
size for online courses. Journal of Technology and Teacher Education, 3(14), 531-541.
United States Congress. (1998). Assistive technology act of 1998. Retrieved from Section 508:
http://www.section508.gov/docs/AssistiveTechnologyActOf1998Full.pdf
United States Government. (2011, October). Section 508. Retrieved from An Official Website of
the United States Goverment: http://www.section508.gov/
Vonderwell, S., & Zachariah, S. (2005). Factors that influence participation in online learning.
Journal of Research on Technology in Education, 38(2), 213-231.
94
wcet. (2009, June). Best practice strategies to promote academic integrity in online education:
Version 2.0. Retrieved January 11, 2012, from wcet ADVANCE: http://wcet.wiche.edu/
wcet/docs/cigs/studentauthentication/BestPractices.pdf
Wiesenberg, F., & Stacey, E. (2005). Reflections on teaching and learning online: Quality
program design, delivery and support issues from a cross-global perspective. Distance
Education, 26(3), 385-404.
Workman, J. J., & Stenard, R. A. (1996). Penn State University DEOS NEWS. Retrieved from
DEOS NEWS: http://www.ed.psu.edu/acsde/deos/deosnews/deosnews6_3.asp
Yates, R. W., & Beaudrie, B. (2009). The impact of online assessment on grades in community
college distance education mathematics courses. The American Journal of Distance
Learning, 23, 62-70.
Zhao, F. (2003). Enhancing the quality of online higher education through measurement. Quality
Assurance in Education, 11(4), 214-221.
95
APPENDIX A
96
97
APPENDIX B
98
99
APPENDIX C
100
QUESTION ANSWER
1. How comfortable are you with online learning o Very uncomfortable with online
technology? learning technology
o Uncomfortable with online
learning technology
o Neutral
o Comfortable with online learning
technology
o Very comfortable with online
learning technology
2. A clear introduction (including overall design, o Strongly Disagree
navigation, and faculty information) was o Disagree
available at the beginning of this on-line course. o Neutral
o Agree
o Strongly Agree
3. Technology support was available for using the o Strongly Disagree
online features of this course. o Disagree
o Neutral
o Agree
o Strongly Agree
4. Student support (for example, advising, o Strongly Disagree
financial aid, registration) was available in o Disagree
using the online format of this course. o Neutral
o Agree
o Strongly Agree
5. I find it important to be provided with the o Strongly Disagree
learning objectives of a course. o Disagree
o Neutral
o Agree
o Strongly Agree
6. The objectives for this online course were o Strongly Disagree
provided at the beginning of this course and o Disagree
were clearly described. o Neutral
o Agree
o Strongly Agree
7. The course objectives for this online course o Strongly Disagree
were closely related to what I was expected to o Disagree
learn. o Neutral
o Agree
o Strongly Agree
8. The course objectives for this online course o Strongly Disagree
assisted with guiding my learning activities. o Disagree
o Neutral
o Agree
o Strongly Agree
101
9. I find it important to be provided with the o Strongly Disagree
course assessment methods at the beginning of o Disagree
a course. o Neutral
o Agree
o Strongly Agree
10. The course assessment methods for this online o Strongly Disagree
course were provided at the beginning of the o Disagree
course. o Neutral
o Agree
o Strongly Agree
11. The course assessment methods for this online o Strongly Disagree
course were clearly described. o Disagree
o Neutral
o Agree
o Strongly Agree
12. The course assessment methods for this online o Strongly Disagree
course included a variety of assessment o Disagree
methods. o Neutral
o Agree
o Strongly Agree
13. The course assessment methods for this online o Strongly Disagree
course were closely related to the course o Disagree
objectives. o Neutral
o Agree
o Strongly Agree
14. I find it important to be provided with the o Strongly Disagree
course resources and materials during a course. o Disagree
o Neutral
o Agree
o Strongly Agree
15. The course resources and materials for this o Strongly Disagree
online course were easily accessible during the o Disagree
course. o Neutral
o Agree
o Strongly Agree
16. The purpose of course resources and materials o Strongly Disagree
for this online course were clearly described. o Disagree
o Neutral
o Agree
o Strongly Agree
17. The course resources and materials for this o Strongly Disagree
online course helped me reach the course o Disagree
objectives. o Neutral
o Agree
o Strongly Agree
102
18. The course resources and materials for this o Strongly Disagree
online course included a wide variety of o Disagree
resources and materials. o Neutral
o Agree
o Strongly Agree
19. I find it important to interact with the instructor o Strongly Disagree
during a course. o Disagree
o Neutral
o Agree
o Strongly Agree
20. The course instructor for this online course o Strongly Disagree
interacted with me in a timely fashion. o Disagree
o Neutral
o Agree
o Strongly Agree
21. The course interaction with the instructor for o Strongly Disagree
this online course helped me reach the course o Disagree
objectives. o Neutral
o Agree
o Strongly Agree
22. The amount of course interaction with other o Strongly Disagree
students for this online course was helpful in o Disagree
reaching the course objectives. o Neutral
o Agree
o Strongly Agree
23. I find it important to be provided with course o Strongly Disagree
technology that enhances learning during a o Disagree
course. o Neutral
o Agree
o Strongly Agree
24. The course technology for this online course o Strongly Disagree
was readily available during the course. o Disagree
o Neutral
o Agree
o Strongly Agree
25. The course technology for this online course o Strongly Disagree
functioned very well. o Disagree
o Neutral
o Agree
o Strongly Agree
26. The course technology for this online course o Strongly Disagree
was helpful in reaching the course objectives. o Disagree
o Neutral
o Agree
o Strongly Agree
103
27. What is your gender? o Female
o Male
28. How many online courses have you taken in the
past (enter number)?
29. What is your age (optional)?
30. Overall, I am satisfied with this online course. o Strongly Disagree
o Disagree
o Neutral
o Agree
Strongly Agree
(Aman, 2009, pp. 187-190)
104
APPENDIX D
105
Dear _____ QM Institutional Representative or Course Designer (whichever is applicable for the
given institution),
You are being asked to take part in a research study. The study is called “Student Perceptions of
Quality and Satisfaction in Online Education.” The study is being conducted by principal
investigator Jill Simpson, a doctoral student in the College of Education at the University of
Alabama. The purpose of this study is to identify factors of online course design that are
significant to increased course quality and student satisfaction.
You must be 19 years old to take part in this study. The study involves completing a web survey
that will take about 10 minutes. This survey includes 1-4 questions (depending on the research
category the institution falls within).
This survey is completely anonymous and confidential. At no point will you be asked to give
your name, student number, or any other identification. As well, the link to this survey contains
no identifying information connected with your email address. The investigator is the only
person that will have access to the password-protected research data. Only summarized data
from all participants will be presented in publications or at meetings.
You will not be paid or receive any tangible benefits from this study. However, the results of the
study may benefit the online education community at large. There are no foreseen risks involved
with this study.
If you have questions about this study, you may contact the investigator, Jill Simpson, at (256)
335-1270 or by email at [email protected], or you may contact the investigator’s
advisor, Dr. Angela Benson at (205) 348-7824 or by email at [email protected] . If you
have any questions about your rights as a research participant you may contact Ms. Tanta Myles,
The University of Alabama Research Compliance Officer, at (205)-348-1355 or toll free at (877)
820-3066.
You may also ask questions, make suggestion, or file complaints and concerns through the IRB
Outreach website at http://osp.ua.edu/site/PRCO_Welcome.html or email them at
[email protected].
Your participation is completely voluntary. You are free not to participate or stop participating
at any time.
Sincerely,
Jill M. Simpson
Doctoral Student
University of Alabama
106
* Institutions in the “recognized” category will be directed to this survey
https://www.surveymonkey.com/s/B3Y7N8Q.
Institutions in the “subscribed” category will be directed to this survey
https://www.surveymonkey.com/s/BB7HJGQ
Institutions in the “non-affiliated” category will be directed to this survey
https://www.surveymonkey.com/s/BP8DCNY
107
APPENDIX E
108
Recognized
a. Rubric
b. Faculty development
d. QM peer-review
e. Self-study
Subscribed
1. What QM tools/resources do you have implemented in your online course(s)? Select all
that apply.
Rubric
Faculty development
QM peer-review
Self-study
Both
109
3. Are these tools/resources required of every online course?
Select Yes or No
Non-affiliated
1. Are you subscribed to any type of peer-review quality assurance program for your online
courses/programs?
Select Yes or No
If yes,
110
APPENDIX F
111
Dear _____ Student*,
You are being asked to take part in a research study. The study is called “Student Perceptions of
Quality and Satisfaction in Online Education.” The study is being conducted by principal
investigator Jill Simpson, a doctoral student in the College of Education at the University of
Alabama. The purpose of this study is to identify factors of online course design that are
significant to increased course quality and student satisfaction.
You must be 19 years old to take part in this study. The study involves completing a web survey
that will take about 10 minutes. This survey includes 11 questions (which is broken down into
30 sub-questions).
This survey is completely anonymous and confidential. At no point will you be asked to give
your name, student number, or any other identification. As well, the link to this survey contains
no identifying information connected with your email address. The investigator is the only
person that will have access to the password-protected research data. Only summarized data
from all participants will be presented in publications or at meetings.
You will not be paid or receive any tangible benefits from this study. However, the results of the
study may benefit the online education community at large. There are no foreseen risks involved
with this study.
If you have questions about this study, you may contact the investigator, Jill Simpson, at (256)
335-1270 or by email at [email protected], or you may contact the investigator’s
advisor, Dr. Angela Benson at (205) 348-7824 or by email at [email protected] . If you
have any questions about your rights as a research participant you may contact Ms. Tanta Myles,
The University of Alabama Research Compliance Officer, at (205)-348-1355 or toll free at (877)
820-3066.
You may also ask questions, make suggestion, or file complaints and concerns through the IRB
Outreach website at http://osp.ua.edu/site/PRCO_Welcome.html or email them at
[email protected].
Your participation is completely voluntary. You are free not to participate or stop participating
at any time.
Sincerely,
Jill M. Simpson
Doctoral Student
University of Alabama
112
APPENDIX G
113
Jill M. Simpson, Principal Investigator from the University of Alabama, is conducting a
study called “Student Perceptions of Quality and Satisfaction in Online Education. She
wishes to find out which factors in online course design are significant to higher course
quality and student satisfaction.
Taking part in this study involves completing a web survey of 30 questions that will take
about 10 minutes to complete. This survey contains questions about demographic and your
level of satisfaction in the online course in which you are enrolled.
We will protect your confidentiality by collecting survey results anonymously. This survey
will not collect Internet Protocol (IP) addresses, or any other identifiable information, of its
participants. Only the investigator will have access to the data. Only summarized data will
be presented at meetings or in publications.
There will be no direct benefits to you. The findings will be useful to online educators for
determining which factors in online course design are significant to student satisfaction of
the course.
The chief risk is that some of the questions may make you uncomfortable. You may skip
any questions you do not want to answer.
If you have questions about this study, please contact Jill Simpson at (256) 335-1270 or by
email at [email protected]. If you have questions about your rights as a research
participant, contact Ms. Tanta Myles (the University Compliance Officer) at (205) 348-
8461 or toll-free at 1-877-820-3066. If you have complaints or concerns about this study,
file them through the UA IRB outreach website at
http://osp.ua.edu/site/PRCO_Welcome.html. Also, if you participate, you are encouraged to
complete the short Survey for Research Participants online at this website. This helps UA
improve its protection of human research participants.
If you understand the statements above, are at least 19 years old, and freely consent to be in
this study, click on the CONTINUE button to begin.
114
APPENDIX H
115
Jill M. Simpson, Principal Investigator from the University of Alabama, is conducting a
study called “Student Perceptions of Quality and Satisfaction in Online Education. She
wishes to find out which factors in online course design are significant to higher course
quality and student satisfaction.
Taking part in this study involves completing a web survey of 1-4 questions (depending on
your research category classification) that will take about 10 minutes to complete. This
survey contains questions about the level of peer-review resources implemented in the
design of your online course.
We will protect your confidentiality by collecting survey results anonymously. This survey
will not collect Internet Protocol (IP) addresses, or any other identifiable information, of its
participants. Only the investigator will have access to the data. Only summarized data will
be presented at meetings or in publications.
There will be no direct benefits to you. The findings will be useful to online educators for
determining which factors in online course design are significant to student satisfaction of
the course.
The chief risk is that some of the questions may make you uncomfortable. You may skip
any questions you do not want to answer.
If you have questions about this study, please contact Jill Simpson at (256) 335-1270 or by
email at [email protected]. If you have questions about your rights as a research
participant, contact Ms. Tanta Myles (the University Compliance Officer) at (205) 348-8461
or toll-free at 1-877-820-3066. If you have complaints or concerns about this study, file
them through the UA IRB outreach website at http://osp.ua.edu/site/PRCO_Welcome.html.
Also, if you participate, you are encouraged to complete the short Survey for Research
Participants online at this website. This helps UA improve its protection of human research
participants.
If you understand the statements above, are at least 19 years old, and freely consent to be in
this study, click on the CONTINUE button to begin.
116
APPENDIX I
117
118
CITI Collaborative Institutional Training Initiative (CITI)
Mentoring Case Study: The Graduate Student Laborer 04/25/12 3/3 (100%)
Mentoring Case Study: Too Much Help is Just Too 04/25/12 3/3 (100%)
Much!
119