Statistics Education
Research Journal
Volume 6 Number 1 May 2007
Editors
Iddo Gal
Tom Short
Assistant Editor
Beth Chance
Associate Editors
Andrej Blejec
Carol Joyce Blumberg
Joan B. Garfield
John Harraway
Flavia Jolliffe
M. Gabriella Ottaviani
Lionel Pereira-Mendoza
Peter Petocz
Maxine Pfannkuch
Mokaeane Polaki
Dave Pratt
Chris Reading
Ernesto Sanchez
Richard L. Scheaffer
Gilberte Schuyten
Jane Watson
International Association for Statistical Education
http://www.stat.auckland.ac.nz/~iase
International Statistical Institute
http://isi.cbs.nl/
Statistics Education Research Journal
The Statistics Education Research Journal (SERJ) is a peer-reviewed electronic journal of the
International Association for Statistical Education (IASE) and the International Statistical Institute
(ISI). SERJ is published twice a year and is free.
SERJ aims to advance research-based knowledge that can help to improve the teaching, learning, and
understanding of statistics or probability at all educational levels and in both formal (classroombased) and informal (out-of-classroom) contexts. Such research may examine, for example, cognitive,
motivational, attitudinal, curricular, teaching-related, technology-related, organizational, or societal
factors and processes that are related to the development and understanding of stochastic knowledge.
In addition, research may focus on how people use or apply statistical and probabilistic information
and ideas, broadly viewed.
The Journal encourages the submission of quality papers related to the above goals, such as reports of
original research (both quantitative and qualitative), integrative and critical reviews of research
literature, analyses of research-based theoretical and methodological models, and other types of
papers described in full in the Guidelines for Authors. All papers are reviewed internally by an
Associate Editor or Editor, and are blind-reviewed by at least two external referees. Contributions in
English are recommended. Contributions in French and Spanish will also be considered. A submitted
paper must not have been published before or be under consideration for publication elsewhere.
Further information and guidelines for authors are available at: http://www.stat.auckland.ac.nz/serj
Submissions
Manuscripts must be submitted by email, as an attached Word document, to co-editor Tom Short
<
[email protected]>. Submitted manuscripts should be produced using the Template file and in
accordance with details in the Guidelines for Authors on the Journal’s Web page:
http://www.stat.auckland.ac.nz/serj
© International Association for Statistical Education (IASE/ISI), May 2007
Publication: IASE/ISI, Voorburg, The Netherlands
Technical Production: California Polytechnic State University, San Luis Obispo, California, United
States of America
Web hosting and technical support: Department of Statistics, University of Auckland, New Zealand
ISSN: 1570-1824
International Association for Statistical Education
President: Gilberte Schuyten (Belgium)
President-Elect: Allan Rossman (United States of America)
Past- President: Chris Wild (New Zealand)
Vice-Presidents: Andrej Blejec (Slovenia), John Harraway (New Zealand), Christine Reading
(Australia), Michiko Watanabe (Japan), Larry Weldon (Canada)
SERJ EDITORIAL BOARD
Editors
Iddo Gal, Department of Human Services, University of Haifa, Eshkol Tower, Room 718, Haifa
31905, Israel. Email:
[email protected]
Tom Short, Mathematics Department, Indiana University of Pennsylvania, 210 South 10th St.,
Indiana, Pennsylvania 15705, USA. Email:
[email protected]
Assistant Editor
Beth Chance, Department of Statistics, California Polytechnic State University, San Luis Obispo,
California, 93407, USA. Email:
[email protected]
Associate Editors
Andrej Blejec, National Institute of Biology, Vecna pot 111 POB 141, SI-1000 Ljubljana, Slovenia.
Email:
[email protected]
Carol Joyce Blumberg, Energy Information Administration, US Department of Energy, 1000
Independence Avenue SW, EI-42, Washington DC, USA. Email:
[email protected]
Joan B. Garfield, Educational Psychology, University of Minnesota, 315 Burton Hall, 178 Pillsbury
Drive, S.E., Minneapolis, MN 55455, USA. Email:
[email protected]
John Harraway, Department of Mathematics and Statistics, University of Otago, P.O.Box 56,
Dunedin, New Zealand. Email:
[email protected]
Flavia Jolliffe, Institute of Mathematics, Statistics and Actuarial Science, University of Kent,
Canterbury, Kent, CT2 7NF, United Kingdom. Email:
[email protected]
M. Gabriella Ottaviani, Dipartimento di Statistica Probabilitá e Statistiche Applicate, Universitá degli
Studi di Roma “La Sapienza”, P.le Aldo Moro, 5, 00185, Rome, Italy. Email:
[email protected]
Lionel Pereira-Mendoza, 3366 Drew Henry Drive, Osgoode, Ottawa K0A 2W0 Ontario
Canada K0A 2W0.. Email:
[email protected]
Peter Petocz, Department of Statistics, Macquarie University, North Ryde, Sydney, NSW 2109,
Australia. Email:
[email protected]
Maxine Pfannkuch, Mathematics Education Unit, Department of Mathematics, The University of
Auckland, Private Bag 92019, Auckland, New Zealand. Email:
[email protected]
Mokaeane Polaki, School of Education, National University of Lesotho, P.O. Box Roma 180,
Lesotho. Email:
[email protected]
Dave Pratt, Institute of Education, University of London, 20 Bedford Way, London WC1H 0AL.
[email protected]
Christine Reading, SiMERR National Centre, Faculty of Education, Health and Professional Studies,
University of New England, Armidale, NSW 2351, Australia. Email:
[email protected]
Ernesto Sánchez, Departamento de Matematica Educativa, CINVESTAV-IPN, Av. Instituto
Politecnico Nacional 2508, Col. San Pedro Zacatenco, 07360, Mexico D. F., Mexico. Email:
[email protected]
Richard L. Scheaffer, Department of Statistics, University of Florida, 907 NW 21 Terrace,
Gainesville, FL 32603, USA. Email:
[email protected]
Gilberte Schuyten, Faculty of Psychology and Educational Sciences, Ghent University, H. Dunantlaan
1, B-9000 Gent, Belgium. Email:
[email protected]
Jane Watson, University of Tasmania, Private Bag 66, Hobart, Tasmania 7001, Australia. Email:
[email protected]
1
TABLE OF CONTENTS
Editorial
2
Call for Papers: Special Issue on Reasoning about Informal Statistical Inference
3
Luc Budé, Margaretha W. J. Van De Wiel, Tjaart Imbos, Math J. J. M. Candel, Nick
J. Broers, Martijn P. F. Berger
Students’ Achievements in a Statistics Course in Relation to Motivational
Aspects and Study Behaviour
5
James E. Corter and Doris C. Zahner
Use of External Visual Representations in Probability Problem Solving
22
Mark A. Earley
Students’ Expectations of Introductory Statistics Instructors
51
Past Conference
67
Forthcoming IASE Conferences
68
Other Forthcoming Conferences
74
2
EDITORIAL1
I have recently returned from the second biennial United States Conference on
Teaching Statistics (USCOTS 2), held in Columbus, Ohio. At the conference Joan
Garfield, a member of the SERJ Editorial Board, received the CAUSE/USCOTS Lifetime
Achievement award. In a special presentation at the conference, Joan reflected on the
statistics education research collaborations in her past and present. It is clear that the field
of statistics education research has developed because of Joan’s curiosity, motivation, and
perseverance in locating research spread among many and varied disciplines including
mathematics, psychology, education and statistics. From 1987 through 1999 Joan
published the Newsletter of the International Study Group for Research on Learning
Probability and Statistics, which later evolved into the Statistical Education Research
Newsletter (2000 through 2001) and finally into what SERJ is today. Thank you, Joan,
and congratulations on your well-deserved recognition.
A healthy strand of statistics education research sessions ran through the program at
USCOTS 2. Some presenter names familiar to SERJ readers included Beth Chance, Bob
delMas, Marsha Lovett, and Mike Shaughnessy, in addition to Joan. We were also given
a glimpse into the future of research in our area when Joan acknowledged her current
Statistics Education Doctoral students at the University of Minnesota. I hope that the
success of Joan’s program will inspire others to sprout up around the world.
Iddo Gal and I have been working behind the scenes to disseminate awareness of
SERJ, and I am pleased to report that SERJ is now listed in two indexing services. One is
“Cabell’s Directory of Publishing Opportunities” (www.cabells.com) and the second is
PsycINFO, which is managed by the American Psychological Association (see
www.apa.org/psycinfo/about/covlist.html). We are working out the details of also listing
SERJ abstracts in EBSCOhost (see www.epnet.com), and will continue to seek other
opportunities to abstract SERJ in indices so that researchers will be able to easily find the
work we publish.
Iddo and I have also been monitoring the acceptance rate for manuscripts submitted to
SERJ. We received 30 manuscripts in 2006, and four of them have so far been accepted
for publication in SERJ. A few more of these are still in revision and may eventually be
accepted and published. The acceptance rate indicates the selectivity and high standards
we maintain for SERJ, but it also suggests that there is plenty of room for more high
quality manuscripts from statistics education researchers.
As Iddo’s term as SERJ co-editor comes to an end, we will be announcing a new coeditor later this year. Thanks to Iddo for his leadership, creativity, and attention to detail
during his term.
Please enjoy the articles and announcements in this new issue of SERJ, and note the
Call for Papers for a Special Issue on Reasoning about Informal Statistical Inference.
Thank you for reading the journal, and please consider sending the results of your own
research to us!
TOM SHORT, for IDDO GAL
Statistics Education Research Journal, 6(1), 2, http://www.stat.auckland.ac.nz/serj
© International Association for Statistical Education (IASE/ISI), May, 2007
3
CALL FOR PAPERS: REASONING ABOUT
INFORMAL STATISTICAL INFERENCE2
The Statistics Education Research Journal (SERJ), a journal of the International
Association for Statistical Education (IASE), is planning a special issue for November
2008, focused on research on Reasoning about Informal Statistical Inference.
Submission deadlines: Letters of intent by Sept. 15, 2007; Full papers by Nov. 1,
2007. Guest Editors will be Dave Pratt (University of London, UK, <
[email protected]>),
and Janet Ainley (University of Leicester, UK, <
[email protected]>).
1. ABOUT INFORMAL STATISTICAL INFERENCE
The aim of the special issue is to advance the current state of research-based
knowledge about the development, learning, and teaching of statistical inference, a
foundational area in statistics education. For the special issue we seek articles focused on
a critical subset of issues in this broad area, describing research related to the
understanding, learning, or teaching of informal aspects of inferential statistical
reasoning, and demonstrating a contribution to research-based educational practice in this
area.
It is recognized at the outset that the definition of what counts as “informal statistical
inference” may at times be slippery, that is, what is informal could depend on the nature
of the inferential tasks being studied, on the complexity of the statistical or probabilistic
concepts involved, on the educational stage, and on other factors. The editors will select
papers for the special issue that focus on learners’ informal ideas about statistical
inference or on learners’ intuitive ways of reasoning about statistical inference in diverse
contexts (see possible research topics below), not on mastery of formal procedures or on
the learning/teaching of formal methods of statistical estimation, significance tests, etc.
The papers being sought will be based on empirical research of a quantitative and/or
qualitative nature on individuals or groups involved in all stages of education, including
all levels of schooling, teacher education, professional development, and workplace and
adult education. Papers on informal inferential reasoning invoked when people face
everyday statistical tasks may also be considered, provided that they discuss clear
educational implications.
2. POSSIBLE RESEARCH TOPICS
Key examples of relevant topics for papers that may fit under the general heading of
Reasoning about Informal Statistical Inference include:
a. How does reasoning about statistical inference develop from simple forms
towards more complex ones? What stages exist in the acquisition of informal
knowledge about statistical inference, or in learning to communicate information
or interpret displays about statistical inference, and how do students develop and
understand the concepts and language that can be used in this regard (e.g.,
sampling, significance, confidence)?
b. What technological tools can be used to promote the understanding of statistical
inference? How are such tools utilized by learners to help in understanding the
building blocks or intermediate steps in statistical inference?
Statistics Education Research Journal, 6(1), 3-4, http://www.stat.auckland.ac.nz/serj
© International Association for Statistical Education (IASE/ISI), May, 2007
4
c. What tasks or sequences of instructional activities can help learners develop a
conceptual understanding of some aspect of statistical inference? How does the
design of tools and/or tasks shape students’ informal inferential reasoning?
d. What types of barriers to students’ informal reasoning about statistical inference
are found, and how can they be avoided or overcome?
e. What types of foundational knowledge (statistical, general) or thinking processes
are needed for or used by learners to informally understand and reason about
statistical inference? How does an informal understanding of statistical inference
connect with or depend on understanding of other statistical concepts?
f. What assessment approaches and research methodologies can be used to
effectively assess understanding, reasoning or learning of informal statistical
inference?
3. OTHER TOPICS RELATED TO FORMAL STATISTICAL INFERENCE
SERJ is also inviting research-based papers on learning, reasoning or understanding
of formal aspects of statistical inference, that is, papers that fall outside the scope of the
notion of “informal statistical inference” as described above, but that otherwise fit the
general aims of the Journal. Such papers would be processed by SERJ as regular papers
and if accepted will be published in a regular issue. Should enough such papers be
accepted for publication, they will be grouped together in a special section and prefaced
with an introductory paper by a member of the SERJ Editorial Board.
4. SUBMISSION GUIDELINES
Authors are advised to aim for papers in the range of 4000-6000 words of body text
(not counting abstract, tables and graphs, references, appendices). Manuscripts for the
special issue will be limited to a maximum of 7500 words of body text, but shorter,
concise papers are encouraged. All manuscripts will be refereed following SERJ’s regular
double-blind peer-review process. Manuscripts should be submitted in accordance with
SERJ’s standard Author Guidelines and using the Template file found on the Journal’s
website: www.stat.auckland.ac.nz/serj.
5. DEADLINES AND CONTACT INFORMATION
Interested authors should send a letter of intent by Sept. 15, 2007, but preferably
earlier, with a 150-250 word abstract describing key aspects of the research. This letter
should be sent by e-mail to SERJ co-editor Iddo Gal: <
[email protected]>, and
authors can expect to get a quick response within 10 days. Authors wishing to send
informal queries regarding the suitability of a planned paper can also contact Iddo Gal.
Full manuscripts must be submitted by Nov. 1st, 2007 at the latest to Iddo Gal at the
address above, in accordance with the submission guidelines listed earlier.
Decisions about the suitability of proposed papers and the allocation of accepted
papers to the special issue or to a regular SERJ issue will be made jointly by the SERJ
Editors and Guest Editors.
5
STUDENTS’ ACHIEVEMENTS IN A STATISTICS
COURSE IN RELATION TO MOTIVATIONAL
ASPECTS AND STUDY BEHAVIOUR3
LUC BUDÉ
Maastricht University, The Netherlands
[email protected]
MARGARETHA W. J. VAN DE WIEL
Maastricht University, The Netherlands
[email protected]
TJAART IMBOS
Maastricht University, The Netherlands
[email protected]
MATH J. J. M. CANDEL
Maastricht University, The Netherlands
[email protected]
NICK J. BROERS
Maastricht University, The Netherlands
[email protected]
MARTIJN P. F. BERGER
Maastricht University, The Netherlands
[email protected]
ABSTRACT
The present study focuses on motivational constructs and their effect on students’
academic achievement within an existing statistics course. First-year Health Sciences
students completed a questionnaire that measures several motivational constructs:
dimensions of causal attributions, outcome expectancy, affect, and study behaviour,
all with respect to statistics. The results showed that when the cause of negative
events was perceived as uncontrollable, outcome expectancy was negative. When the
cause of negative events was perceived as stable, affect toward statistics was
negative. Furthermore, negative affect toward statistics and limited study behaviour
led to unsatisfactory achievements. Path analysis (Lisrel) largely confirmed the
causal relations in a model that was based on attributional and learned helplessness
theories. The consequences of these findings for statistics education are discussed.
Keywords: Statistics education research; Motivation; Conceptual understanding;
Study behaviour
Statistics Education Research Journal, 6(1), 5-21, http://www.stat.auckland.ac.nz/serj
© International Association for Statistical Education (IASE/ISI), May, 2007
6
1. INTRODUCTION
Motivation influences the scope and the quality of study behaviour of students (see
e.g., Bruning, Schraw, & Ronning, 1999; Deci & Ryan, 1985; Graham & Weiner, 1987;
Pintrich, 2000). High-quality study behaviour involves active knowledge construction.
Active knowledge construction is known to enhance understanding of the material in
many courses (see e.g., Chi, de Leeuw, Chiu, & LaVancher, 1994; Phye, 1997; Steffe &
Gale, 1995), including statistics courses (see e.g., Garfield, 1993; Giraud, 1997; Keeler &
Steinhorst, 1995; Magel, 1998). Therefore, in attempts to improve statistics education, it
is fundamental to stimulate motivation.
Research on motivation is quite extensive and covers heterogeneous constructs (see
e.g., Ames, 1992; Boekaerts, 1997; Volet, 1997; Weiner, 1992). Some of these constructs
involve phenomena that are difficult to change, because they are to a large extent
determined by traits of the individual that is involved, such as goal orientation, selfdetermination, and competence. Our aim is not to focus on such phenomena, but rather to
focus on constructs that have practical implications for statistics education, that is,
constructs that can be manipulated and acted upon while trying to improve statistics
education.
For that reason we have focused on two motivational theories that offer opportunities
to intervene in motivational processes. Both theories take the starting-point of the
explanations people perceive for events they experience. These so called causal
explanations have cognitive, affective, and behavioural consequences. Examples of
cognitive consequences in a statistics educational context are expected outcomes of
attending lectures or studying a course book; examples of affective consequences are
enjoyment, pleasure, and interest; and examples of behavioural consequences are effort
and persistence. The influence of causal explanations on cognition, affect, and behaviour
might be manipulated and driven toward outcomes that are more positive, in terms of
motivation. As a consequence, these causal explanations have practical implications for
statistics education, because the obtained improvement of motivation might result in
study behaviour that enhances understanding. The goal of the study was to investigate
these phenomena in the context of statistics education.
2. MOTIVATIONAL MODEL
In statistics education one can sometimes encounter students who think that there is a
stable cause for failing an exam (e.g., statistics is a difficult subject). These students may
no longer expect to benefit from studying statistics; they may start to dislike it and will
not spend much study time on this subject. Other students may think that they have no
control over the outcomes of their actions. For example, “no matter how hard I study, I
will not be able to understand it.” These students may in advance expect to fail on the
exam, will also start to dislike statistics, and will not spend much time studying the
material. These examples show the influence of causal attributions (stability of causes,
non-controllability of causes) on cognitions such as outcome expectancies (no benefit
from studying statistics, expectancy to fail on the exam) and consequently on emotions
(affective reactions of starting to dislike statistics) and behaviour (disregarding statistics),
which will finally have an effect on achievement. This chain effect, which is
consequential for statistics education, is reflected in a model that was developed and
tested in this study.
The model as a whole stands for motivation (see Figure 1). Motivation is not a
separate entity in our model for two reasons. Firstly, it is difficult to insert it separately
7
into a model, because it is an abstract, complex (Weiner, 1992), and ill-defined (Murphy
& Alexander, 2000) construct, which is frequently used in colloquial language and
consequently has several connotations. Moreover, motivation is studied in different
domains and from different perspectives, which has led to distinct and changing
conceptualisations and approaches. Various motivational constructs are studied, such as
self-efficacy, goal orientation, metacognitive strategies, value, strategy use, causal
perceptions, autonomy, social relatedness, as so forth. (See e.g., Ames, 1992; Boekaerts,
1997; Dweck, 2000; Pintrich & Schunk, 1996; Volet, 1997; Weiner, 1986.) In these
studies it is often left implicit whether these constructs are part of motivation or are
merely related to motivation (Murphy & Alexander, 2000). Our model as a whole reflects
our perspective on motivation.
Control
Stable Explanation
Outcome Expectancy
Effort
Persistence
Affect
Achievement
Figure 1. Statistics motivational model based on the attributional
and the learned helplessness theory
Secondly, it is in our view not necessary to integrate motivation as a separate
construct in the model. Traditionally, motivation was seen as an isolated latent construct
that drives behaviour, cognition, and affect. We think that motivation merely is the sum
of behaviour, cognition, and affect. Our opinion is in accordance with the remark of
Weiner (1992), referring to Kelly (1958), that motivation as a model construct might be
redundant; it is sufficient to represent only those variables that make up motivation. This
view is also compatible with the fact that most motivational models do not explicitly
contain motivation as a construct (see e.g., Bruning, Schraw & Ronning, 1999; Deci &
Ryan, 1985; Pintrich, 2000; Pintrich & Schunk, 1996; Weiner, 1992). Therefore, the
model that we developed contains only manifest variables that together stand for
motivation, and does not contain motivation as a separate latent entity.
Two specific motivational theories were used for our model; namely the attributional
and the learned helplessness theory, because they both use the starting-point of perceived
causes for aversive events. The attribution-based theory of motivation (Graham &
Weiner, 1987; Pintrich & Schunk, 1996; Weiner, 1986, 1992) commences with perceived
causes for failure, unexpected outcomes, unusual events, and important situations.
Perceived causes are the way people explain to themselves such outcomes, events, and
situations. The connotations of the explanations are determined by underlying properties.
In attribution-based theory these underlying properties of such explanations are divided
into three dimensions: stability, control, and locus. Pintrich and Schunk (1996) propose,
8
however, that the stability dimension is most closely linked to beliefs regarding future
success (outcome expectancy) and subsequently to affect and actual achievement
behaviour. Therefore, we integrated stable explanation in our model in Figure 1. It can be
defined as the invariability over time of such perceived causes, namely causal
explanations.
Peterson, Maier, and Seligman (1993) present a motivational theory, which originally
emanates from the learned helplessness paradigm. In this paradigm, individuals are
thought to become passive and to develop affective deficits if they cannot control and
avoid the causes of aversive stimuli. They claim therefore, in contrast to Pintrich and
Schunk (1996), that controllability is the major factor contributing to a negative outcome
expectancy. Uncontrollable events will, according to Peterson et al., lead to a perceived
non-contingency between people’s actions and the outcomes of their actions. This
negative outcome expectancy will lead to pessimistic thoughts, negative emotions
(affect), and passivity (behaviour). This is what is called learned helplessness. We
integrated control influencing outcome expectation as a separate construct in our model.
Control is defined as the ability to avoid the causes of aversive stimuli.
Although the two presented theories slightly differ in the emphasis of the causal
dimensions control and stability, they both reflect the way these properties of negative
causal explanations contribute to a negative outcome expectancy, and how this will act
upon affect and on behaviour, such as effort and persistence, which will finally result in
an effect on achievement. The causal relations among these constructs are symbolised by
arrows in our model that is presented in Figure 1.
This model was examined within the domain of statistics education. This means that
all the constructs were measured with respect to statistical events and phenomena. It is
known that perceived causal explanations via expectancy, affect, and behaviour
determine future achievements in mathematics (see e.g., Seegers & Boekaerts, 1993;
Vålas & Søvik, 1994). Our question was whether this is also true for statistics education
and if the results would provide useful information for the reformation of statistics
education.
The following research questions were addressed:
1. How do students causally explain statistics related events? Do they think that
they have control over, for example, the mastery of the material, the amount of
time they can spend on studying statistics, and the result on the tests? We also
wanted to know whether or not the causes that the students reported for these
events were stable.
2. We further measured the outcome expectancies, that is, whether students
experience a contingency between studying statistics and their understanding of
the topics and the grades they receive on statistics tests. We also investigated the
influence of outcome expectancy on effort, persistence and affect.
3. Finally, we investigated the relations between these motivational constructs and
achievement. The potential causal relations among these constructs were tested
with structural equation modelling via Lisrel (Jöreskog & Sörbom, 1989).
3. METHOD
3.1. PARTICIPANTS
Two hundred (n = 200) first-year students of the faculty of Health Sciences
participated in a pilot study to establish the reliability of a questionnaire that was
developed to measure the motivational constructs. In the subsequent year n = 94 first-year
9
students of the faculty of Health Sciences participated in the main study; 79 of these
participants were female, 15 were male. The ages ranged from 19 to 26 years.
Approximately 75 percent of the first-year Health Sciences student body is female. The
participants were recruited during educational activities before the start of the
introductory statistics course in which this study was executed. During recruitment they
were told that they had to answer questions about statistics education and that they would
be paid 10 euro. This payment was given to avoid attracting only motivated students who
were particularly interested in statistics. All participants took the introductory statistics
course.
3.2. MEASUREMENT INSTRUMENTS AND PROCEDURE
A questionnaire to measure the motivational constructs that are relevant for our model
was developed. This Motivation toward Statistics Questionnaire (MSQ) consisted of 38
items, divided into six subscales. The items were phrased as statements and participants
responded on a 7-point Likert scale. The questionnaire is partly a Dutch translation of the
Survey of Attitudes Toward Statistics (SATS) (Gal, Ginsburg, & Schau, 1997).
Additional items with regard to causal explanations were formulated using the same
principles as the Attributional Style Questionnaire (ASQ) (Peterson et al., 1993), in
particular for two attributional dimensions: stability and control. Finally, items were
added to measure the two aspects of study behaviour: effort and persistence. All MSQ
items concentrated on statistics related events. Because the MSQ was for the greater part
based on existing surveys that have been proven to be valid (Peterson et al., 1993; Schau,
Stevens, Dauphinee, & Del Vecchio, 1995), it can be considered an adequate
measurement instrument regarding the relevant motivational constructs. Example
questions are presented in Table 1. Based on content the items were divided into six
subscales. To establish the reliability of the MSQ, it was administered to 200 first-year
Health Sciences students and Cronbach’s alpha was computed for each subscale. Six
questions that did not fit in the subscale were identified. Four questions were removed;
two were rephrased. The MSQ was used the subsequent year for collecting data for the
main study. It was administered to the students at the beginning of the introductory
statistics course. Students received written instructions before they completed the MSQ.
The whole procedure took approximately half an hour.
A second instrument was used to assess participants on effort and persistence,
because it is well known that self reports and students’ responses to questionnaires may
not always adequately reveal mental processes and behaviour (Biggs, 1993; Nisbett &
Wilson, 1977; Schwartz, 1999; Watkins, 1996). The goal was to obtain more reliable data
on study behaviour. The instrument consisted of two rating scales ranging from zero to
ten. It was distributed to the tutors of tutorial group meetings. These are weekly two hour
sessions supervised by a tutor, in which the students discuss the subject matter. The
sessions are an essential part of the course. The tutors were given instructions on how to
infer students’ effort and persistence. They were told what was meant by effort and
persistence, examples were given, and they were told how to use the rating scale (grades
ranging from zero to ten are customary in our education). This came down to instructing
them to ask and register whether students attended the lectures, whether students were
prepared for the tutorial group meetings, and whether students were actively involved in
the discussion during the obligatory meetings. The tutors had to convert their impression
concerning these aspects into a grade called effort. Persistence was analogously a grade
based on the tutors’ judgement concerning whether students continued asking questions
during the meetings until they really understood the subject matter, whether students at
10
home persisted in trying to solve their assignments by using lecture notes and/or their
books, or whether they consulted their teacher when they were not able to solve an
assignment. The participants were evaluated by their tutors in the week before the end of
the course. Finally, the scores on the exam at the end of the course were used as an
indicator for participants’ achievements. The exam consisted of 30 multiple choice
questions and grades could range from zero to ten. Example questions of the exam are
presented in the appendix.
3.3. ANALYSIS
Sum scores of the responses to the questionnaire were computed for each subscale by
summing the scores of individual items. Some items were positively phrased, others
negatively. Responses on the negatively phrased items were mirrored so that all answers
were in the same direction. The sum scores were called: Stable Explanation, Control,
Outcome Expectancy, Affect, Effort, and Persistence. To reflect the facts that people seek
causes especially for failure (Graham & Weiner, 1987) and that motivation to study
statistics is usually modest, the coding on the variables Stable Explanation and Control
was done in such a way that high scores corresponded with respectively a stable negative
explanation and lack of control. Cronbach’s α was computed for each subscale. The exam
grades (Achievement) and the tutor ratings Effort(T) and Persistence(T) consisted of
grades ranging from zero to ten. They were included into the analyses as raw data.
Four analyses were done. First, several t tests were done to test for possible selection
biases. A comparison was done between the male and female participants on
Achievement, Stable Explanation, Control, Outcome Expectancy, Affect, Effort(T), and
Persistence(T). Moreover, achievement was compared between the participants in our
study and the rest of the cohort that took the introductory course. Second, bivariate
correlations between all variables were calculated to inspect the correlation patterns. The
covariance structure modelling was, because of the rather small sample size, done in two
separate steps (Scott Long, 1983), resulting in the third and fourth analysis. The third
analysis was a robust maximum likelihood confirmative factor analysis (the simultaneous
analysis of the covariance and the asymptotic covariance matrix; Jöreskog & Sörbom,
1989), which was done to confirm the measurement structure. Fourth, a path analysis (a
robust maximum likelihood structural equation modelling) was done with Lisrel. Due to
the sample size it was necessary to disregard the measurement structure in this analysis.
Hence, the analysis was done without latent variables and the sum scores of the separate
items of the MSQ served as manifest variables. With this path analysis the model
presented in Figure 1 was tested.
4. RESULTS
From the pilot study, Cronbach’s α for each subscale (after the removal of the four
items) and some example questions are presented in Table 1.
A robust maximum likelihood confirmatory factor analysis was executed on those
data of the MSQ that were also used in the path analysis of the main study (n = 94). The
content based classification of the items on the subscales Control, Stable Explanation,
Outcome Expectancy, and Affect was supported by the results of this confirmatory factor
analysis; indices showed a proper fit. The Satorra-Bentler chi-square was used. It is
considered to be more robust against a small sample size and violations of distributional
assumptions (Hu, Bentler, & Kano, 1992; Satorra & Bentler, 1994).
11
Table 1. Subscales of the MSQ (n = 200)
Subscales and example questions
Stable explanation:
Statistics is just a difficult subject.
I have always had difficulties with statistics.
Control:
The result on the statistics exam is determined by my own
endeavour.
Whenever I don’t understand a statistical topic, I know
what to do.
Outcome Expectancy:
It pays off to study statistics.
The time I spend on statistics is wasted.
Affect:
To study statistics is enjoyable.
I think statistics is interesting.
Effort:
I spend a lot of time on statistics.
I never prepare myself for the statistics tutorial group
meeting.
Persistence:
Whenever I don’t understand something from statistics, I
quit.
When I cannot complete a statistics assignment, I go
through the book once again.
Number of items
4 items
Cronbachs’s α
.8427
5 items
.7797
6 items
.6048
8 items
.7813
8 items
.8058
7 items
.7405
The Lisrel program provides several additional indices for how well the model fits the
data (Jöreskog & Sörbom, 1988). A goodness of fit index (GFI) is given for the whole
model. It compares the tested model with a so called null-model, that is, all parameters
are fixed on zero. A second index is the normed fit index (NFI), which compares the
tested model with an independence model (variances are set free, covariances are fixed on
zero). This index, however, continues to improve when paths are added and therefore
does not appraise parsimonious models adequately. The most meaningful index is the
non-normed fit index (NNFI). In this index the degrees of freedom are taken into account
and consequently it appraises not only the best fitting, but also the most parsimonious
model. All three fit indices should be close to one. Finally the root mean square residual
(RMR) is given. This index, as the residuals, is ideally close to zero. The indices
presented in Table 2 show a proper fit for this model; that is, the items adequately fit into
their subscales.
Table 2. Fit indices for the confirmatory factor analysis on Control,
Stable Explanation, Outcome Expectancy, and Affect
Satorra-Bentler chi-square
(df = 224, n = 94)
* p < 0.05
277.18; p = .04*
GFI
.86
NFI
.89
NNFI
.93
CFI
.94
Standardised RMR
.22
12
In Table 3 descriptive statistics of all the variables as measured by the MSQ, as well
as the tutor ratings and the exam grades are given.
Table 3. Descriptives of the motivational variables and achievement
Stable Explanation
Control
Outcome Expect
Affect
Effort
Persistence
Effort(T)
Persistence(T)
Achievement
Mean
SD
Items
16.39
16.22
29.13
26.66
37.88
31.93
7.11
6.64
7.05
5.61
4.84
5.13
7.60
7.62
6.48
1.55
1.73
1.90
4
5
6
8
8
7
4
4
30
Scale
min
4.00
5.00
6.00
8.00
8.00
7.00
0.00
0.00
0.00
Scale Min
max score
28.00 4.00
35.00
5.00
42.00 14.00
56.00 12.00
56.00 16.00
49.00 11.00
10.00
2.00
10.00
1.00
10.00
1.60
Max
score
28.00
31.00
40.00
51.00
54.00
46.00
10.00
10.00
9.40
Skewness
.100
.697
-.562
.196
-.477
-.207
-.803
-1.028
-.780
Kurtosis
-.459
.637
.826
.174
.347
.293
1.212
1.862
-.133
The results of the t tests showed no significant differences between female and male
participants. This might partly be because of the restricted power of the tests, so
additionally the effect sizes (Cohen’s d) were computed. The results are respectively for
Achievement (d = .13; p = .65), Control (d = .17; p = .51), Stable Explanation (d = .53;
p = .08), Outcome Expectancy (d = .22; p = .49), Affect (d = .008; p = .97), Effort(T)
(d = .02; p = .94), and Persistence(T) (d = .20; p = .52). Combined, these results indicate
no substantial differences between male and female participants. An additional t test was
done to test for another possible selection bias. In this t test the achievement of the
students who participated in our study was compared to the rest of the cohort (n = 122).
No significant difference was found, nor a consequential effect size (p = .82; d = .06).
A correlation matrix of all variables was computed and is presented in Table 4. The
significance level was adjusted with a Bonferroni correction. Both dimensions of
attribution (Stable Explanation and Control) were significantly correlated to Outcome
Expectation. The notion of having no control was most strongly correlated to Outcome
Expectation. Outcome Expectation was significantly correlated with Affect toward
statistics.
Affect was significantly correlated to Achievement, but as expected not to the selfreported behavioural constructs (Effort and Persistence), which were also not correlated
to Achievement. The tutor ratings Effort(T) and Persistence(T) on the other hand were
much better predictors for Achievement and were more highly correlated to Affect. This is
consistent with research that established the inaccuracy of self-reports and research that
showed that students’ responses to questionnaires may not always adequately reveal their
own learning (Biggs, 1993; Glenberg, Sanocki, Epstein, & Morris, 1987; Nisbett &
Wilson, 1977; Schwartz, 1999; Watkins, 1996).
A path analysis with Lisrel was conducted, because of this above-mentioned
inaccuracy of self-reports, on a model where the tutor ratings Effort(T) and Persistence(T)
were inserted instead of the self-reported study behaviour (Effort and Persistence). We
started with our model that was presented in Figure 1. The relation between Stable
Explanation and Outcome Expectancy based on attributional theories was not significant
(Standardised Path coefficient β = .06; p = .31). We did find a strong negative relation
between the notion of having no control (Control) and Outcome Expectancy (β = -.68; p <
.001). Apparently, if a student thinks that there is no contingency between, for example,
his study activities and the result on an exam, he will not expect a positive outcome of his
actions.
13
Table 4. Correlations between the motivational variables and achievement.
Stable Explanation
Stable Explanation
1
Control
Outcome Expectancy
Affect
Effort
Persistence
Effort(T)
Persistence(T)
Achievement
* p ≤ 0.001 (Bonferroni corrected)
Control
.584*
p < .001
Outcome Expectancy
-.336*
p = .001
Affect
-.550*
p < .001
Effort
.156
p = .067
Persistence
-.052
p = .310
Effort(T)
-.116
p = .132
Persistence(T)
-.138
p = .093
Achievement
-.392*
p < .001
1
-.647*
p < .001
-.306
p = .003
.152
p = .072
-.099
p = .172
-.020
p = .423
.016
p = .439
-.121
p = .123
1
.312*
p = .001
.020
p = .424
.157
p = .065
.127
p = .112
-.006
p = .479
.226
p = .015
1
.125
p = .115
.239
p = .010
.266
p = .005
.216
p = .018
.429*
p < .001
1
.746*
p < .001
.273
p = .004
.263
p = .005
.261
p = .006
1
.368*
p < .001
.337*
p = .001
.294
p = .002
1
.843*
p < .001
.455*
p < .001
1
.478*
p < .001
1
14
Control
Stable Explanation
-.68**
.06
Outcome Expectancy
-.58**
.24**
-.08
Affect
.05
.25**
.24**
.33**
Persistence(T)
Effort(T)
.33**
.08
.34**
Achievement
theoretical relations that were confirmed
theoretical relations that are not confirmed
meaningful relations that were not in the theoretical model
Figure 2. Statistics motivational model, as confirmed by path analysis (Lisrel)
Notes: Coefficients are standardised; *p < 0.05; **p < 0.01.
The relations among the motivational constructs as well as the coefficients are
displayed in Figure 2. The solid arrows in Figure 2 stand for the theoretical relations that
were confirmed, the dotted arrows stand for the theoretical relations that were not
confirmed, and the dashed arrows indicate meaningful relations that were not in the
hypothesised theoretical model, as shown in Figure 1.
Figure 2 shows a strong direct relation between Stable Explanation and Affect, that is,
if students think that there are stable causes for negative statistics related events, failing
their exams for example, they will develop negative feelings toward statistics. In the
model, as displayed in Figure 1, this relation was mediated by Outcome Expectancy.
A negative Outcome Expectancy also had an adverse effect on Affect. Affect is related
to all other constructs except to the notion of no control (Control) (β = .19; p = .08). To
emphasise the importance of Affect, it has been placed in a more central position in
Figure 2. It is strongly related to Achievement directly, as well as via Persistence(T). Also
important is that Achievement is determined by Persistence(T) (β = .34; p < .001) but not
by Effort(T) (β = .08; p = .36).
To enhance the fit of the model, the residuals of the behavioural constructs Effort(T)
and Persistence(T) had been set free to correlate (error covariance = 2.12; t = 5.34) in
15
Lisrel. With this relaxation of the model (as presented in Figure 2), all fit indices showed
a good fit. The values of these indices for our model are provided in Table 5. Again the
Satorra-Bentler chi-square is presented because of its robustness against a small sample
size and violations of distributional assumptions (Hu et al., 1992; Satorra & Bentler,
1994).
Table 5. Fit indices for the model in Figure 2
Satorra-Bentler chi-square
(df = 7, n = 94)
13.40; p = .063
GFI
.96
NFI
.95
NNFI
.93
CFI
.98
Standardised RMR
.042
5. DISCUSSION
This study was done in an introductory statistics course. It focussed on causal
explanations of statistics related events, perceived outcome expectancy of students’
activities within this statistics course, affect and study behaviour toward statistics, and the
relation of these constructs to the results on the exam at the end of the course. These
constructs were chosen because of their practical implications for the teaching of
statistics.
Our first findings concern causal explanations. In the two presented motivational
theories, perceived causes for events have underlying properties that have affective,
behavioural, and cognitive consequences (Peterson et al., 1993; Pintrich & Schunk,
1996). In our study we focused on the dimensions of control and stability of causal
explanations.
The first result concerns control. The model in Figure 2 indicates that the perception
of having no control over causes of statistics related events may lead to decreased
outcome expectancy. For example, a student who thinks that there is nothing he can do
about the causes for failing the statistics exams, or thinks that he is not able to understand
statistics anyway, may not expect a positive outcome from attending the lectures or
studying the material. This mechanism is intuitively appealing.
The second result indicates that the stability of causal explanations may be more
directly related to affect. As is seen in Figure 2 we found a significant path from Stable
Explanation of such causes to Affect. The path that we found may be interpreted as
follows. The perception of stable causes for aversive events related to statistics may lead
to displeasure and frustration. If students perceive that failing statistics exams is not
easily changeable, students may start to dislike statistics. This was reflected in responses
like: I dislike statistics; I do not have a positive perception of statistics; and so forth.
In sum, these two findings indicate that students who think that they lack control may
not expect to profit from studying statistics, and students who do invest time but think
that there are stable causes for failing in spite of that, may start to dislike statistics.
The last path from Stable Explanation to Affect, though intuitively appealing, was not
anticipated. The model in Figure 1 contained a relation between Stable Explanations and
Outcome Expectancy. This relation was based on the general attributional position that
the stability of a cause has the most influence on shifts in expectancy (Pintrich, 2000;
Pintrich & Schunk, 1996; Weiner, 1986, 1992). Our findings are more consistent with the
basic assumption from Peterson et al. (1993) that controllability is the major factor
16
influencing outcome expectancy. Yet, the direct influence of Stable Explanation on Affect
may also have important practical implications for statistics education.
The implication for education from our findings may be that when students discover
the material is comprehensible to them and they experience success, they will be
stimulated to study the material. This means that in constructing a learning environment,
there should be tasks built in that are feasible for students. In that way the sequence of
events that may lead to diminished motivation (Weiner, 1986) may be interrupted.
Students will gradually sense that they can master the topics, they will discover they can
control their learning outcomes, they will experience success, and they will abandon the
idea that there are stable causes for failure. Control over learning outcomes may foster the
positive expectation of future study activities. This positive expectation, together with the
reduction of the perception of stable negative causes for failure, may even promote
students to enjoy studying statistics. Only then should more difficult tasks be
administered.
A second finding of interest in our study seems to be the central position of Affect in
our model in Figure 2. Students who appreciate the value and relevance of statistics, who
think it is interesting, challenging, and who like statistics, appear to study statistics more
and qualitatively better, and perform better on the exams. In attributional theories
(Pintrich, 2000; Pintrich & Schunk, 1996; Weiner, 1986, 1992) as well as in the learned
helplessness theory of Peterson et al. (1993), affect is on the same level as behaviour and
cognition. In the model in Figure 1 Affect was therefore put on par with behavioural
consequences of Outcome Expectancy. However, affect seems to have a more prominent
role in motivational processes in the present statistics education context. In our study we
found that Affect directly and positively influenced Achievement. It also influenced study
behaviour, namely Effort(T) and Persistence(T). Persistence(T) in turn also influenced
Achievement. Thus, Affect seems to determine achievement directly, as well as indirectly.
Moreover, we found that Affect functioned as a mediator between Control, Stable
Explanations, and Outcome Expectancy on the one hand, and the rest of the motivational
constructs on the other. For this reason Affect holds a more central position in our model
in Figure 2 than in the model presented in Figure 1.
The central role of Affect suggests that the students’ feelings toward statistics appear
to be an important theme for innovating and improving statistics education. Our results
with respect to Affect are in line with Malone and Lepper (1987), who state that
implementing features that make learning more appealing, enjoyable, and challenging
makes learning more intrinsically motivating. Our finding that the feelings toward
statistics are crucial in reaching satisfactory achievements corroborates the results of Isen,
Daubman, and Gorgoglione (1987). In their study they found that positive affect may
foster student’s tendencies to see relations among stimuli, because positive affect leads to
different ways of information processing, for example using different strategies. More
relations between concepts are characteristic for richer knowledge networks, which
indicate better integrated knowledge and deeper understanding (Kintsch, 1988, 1998).
It seems to be of relevance in the improvement of statistics education to make
statistics courses more attractive, interesting, and enjoyable. One of the ways this might
be achieved is by making the courses less theoretical. We think that a small experiment
may engage students in a more active way, it may be fun to analyse data that are collected
by the students themselves, and it may foster the notion of relevance of statistics.
A final result in our study was that Effort(T) had no significant relation with
Achievement. Both Effort(T) and Persistence(T) were determined by the tutors. Effort(T)
reflected the amount of time students studied, and whether students prepared themselves,
attended lectures, or were actively involved in the discussion during group meetings.
17
Effort per se seemed to have a minor effect on achievement. What counts seems to be the
way students study. In our study, Persistence(T) contributes significantly to exam
performance. Students who did not quit that easily, who persisted, who turned to their
lecture notes or their books, or consulted a teacher when they were not able to solve a
statistical problem, those students did better on the exam. This result suggests that
persisting is the best way to study statistics. It is in line with research in other subjects
that established the importance of learning strategies and mastery goals for achievement
in educational settings. (see e.g., Ames, 1992; Boekaerts; 1997; Pintrich, 2000; Dweck,
2000).
This finding may also be important for educational purposes. In the teaching of
statistics, students should be stimulated to try to solve their problems. They should try to
persist instead of quitting all too easily. This can be done by guiding them through the
topics and by pointing them in the correct direction, instead of giving the solution to a
problem promptly. Persisting and learning from mastering their own difficulties may be
the most valuable way of learning.
The student population from which we recruited our participants consists largely of
female students. Consequently, most of our participants were female (79 female versus
15 male). This could have affected our results. However, t tests on all the core variables
(Control, Stable Explanation, Outcome Expectancy, Affect, Effort(T), Persistence(T), and
Achievement) in our models showed no significant differences between the female and
male students. Therefore, the fact that the majority of our participants was female seems
not to affect the motivational processes that were studied.
The tutor ratings that we used to measure effort and persistence are another limitation
of our study. We instructed the tutors in great detail and asked them to record students’
activities that we hold indicative for effort and persistence. We are confident that the
ratings of the tutors are a quite valid and reliable measurement of the relevant behaviour.
Still these ratings only reflect observable, external behaviour. Consequently we cannot
discuss internal processes of reflection and mental activity. Our results only pertain to
self-reported cognitions, affect, and observed behaviour.
In the present study only first-year students were studied. In future research secondand third-year students could be studied. Secondly, our results could be corroborated in
studies with a larger sample. In our study a rather small sample was used (n = 94). It
could also be investigated how in a practical educational context we can determine
whether students persist during studying statistics. How can students optimally be guided
to the correct solution of the problems? Will this reduce the perception of stable negative
causes for failure and enhance the notion of control? Will such a reduction lead to a
positive expectation of future study activities and to more enjoyment? Will all this
eventually lead to more persistence and better results on the exam? Finally, further
research is needed to investigate additional ways statistics education can be made more
enjoyable. In the past, our department spent most attention on how to make lectures more
informative, to select the best instruction books, and to develop assignments that are
mainly educational. Now our attention has somewhat shifted toward making the courses
more attractive, interesting, and enjoyable. We have tried to make the courses less
theoretical by introducing a small experiment. Even so, future research may include
investigating the most effective ways of making statistics education more enjoyable.
REFERENCES
Ames, C. (1992). Classrooms: Goals, structures, and student motivation. Journal of
Educational Psychology, 84(3), 261-271.
18
Biggs, J. B. (1993). What do inventories of students’ learning processes really measure?
A theoretical review and clarification. British Journal of Educational Psychology,
63(1), 3-19.
Boekaerts, M. (1997). Self-regulated learning: A new concept embraced by researchers,
policy makers, educators, teachers, and students. Learning and Instruction, 7(2), 161186.
Bruning, R. H., Schraw, G. J., & Ronning, R. R. (1999). Cognitive psychology and
instruction. Upper Saddle River, NY: Merrill, Prentice Hall.
Chi, M. T. H., de Leeuw, N., Chiu, M-H., & LaVancher, C. (1994). Eliciting selfexplanations improves learning. Cognitive Science, 18(3), 439-477.
Deci, E. L., & Ryan, R. M. (1985). Intrinsic motivation and self-determination in human
behavior. New York: Plenum Press.
Dweck, C. S. (2000). Self-theories: Their role in motivation, personality, and
development. Philadelphia: Psychology Press.
Gal, I., Ginsburg, L., & Schau, C. (1997). Monitoring attitudes and beliefs in statistics
education. In I. Gal, & J. B. Garfield (Eds.), The assessment challenge in statistics
education (pp 37-51). Amsterdam: IOS Press and the International Statistical
Institute.
Garfield, J. (1993). Teaching statistics using small group cooperative learning. Journal of
Statistics Education, 1(1).
[Online: www.amstat.org/publications/jse/v1n1/garfield.html]
Giraud, G. (1997). Cooperative learning and statistics education. Journal of Statistics
Education, 5(3).
[Online: www.amstat.org/publications/jse/v5n3/giraud.html]
Glenberg, A. M., Sanocki, T., Epstein, W., & Morris, C. (1987). Enhancing calibration of
comprehension. Journal of Experimental Psychology, 116(2), 119-136.
Graham, S., & Weiner, B. (1987). Some educational implications of sympathy and anger
from an attributional perspective. In R. E. Snow & M. J. Farr (Eds.), Aptitude,
learning, and instruction. Volume 3: Conative and affective process analyses (pp.
199-221). Hillsdale: Lawrence Erlbaum Associates.
Hu, L., Bentler, P. M., & Kano, Y. (1992). Can test statistics in covariance structure
analysis be trusted? Psychological Bulletin, 112(2), 351-362.
Isen, A. M., Daubman, K. A., & Gorgoglione, J. M. (1987). The influence of positive
affect on cognitive organisation: Implications for education. In R. E. Snow & M. J.
Farr (Eds.), Aptitude, learning, and instruction. Volume 3: Conative and affective
process analyses (pp. 143-164). Hillsdale: Lawrence Erlbaum Associates.
Jöreskog, K. G., & Sörbom, D. (1989). LISREL 7. A guide to the program and
applications. Chicago: Scientific Software, Inc.
Keeler C. M., & Steinhorst, R. K. (1995). Using small groups to promote active learning
in the introductory statistics course: A report from the field. Journal of Statistics
Education, 3(2).
[Online: www.amstat.org/publications/jse/v3n2/keeler.html]
Kelly, G. A. (1958). Man’s construction of his alternatives. In G. Lindzey (Ed.),
Assessment of human motives (pp. 33-64). New York: Grove.
Kintsch, W. (1988). The role of knowledge in discourse comprehension: A constructionintegration model. Psychological Review, 95(2), 163-182.
Kintsch, W. (1998). Comprehension. A paradigm for cognition. Cambridge: Cambridge
University Press.
Magel, R. C. (1998). Cooperative learning and statistics instruction. Journal of Statistics
Education, 6(3).
19
[Online: www.amstat.org/publications/jse/v6n3/magel.html]
Malone, T. W., & Lepper, M. R. (1987). Making learning fun: A taxonomy of intrinsic
motivations for learning. In R. E. Snow & M. J. Farr (Eds.), Aptitude, learning, and
instruction. Volume 3: Conative and affective process analyses (pp. 223-253).
Hillsdale: Lawrence Erlbaum Associates.
Murphy, P. K., & Alexander, P. A. (2000). A motivated exploration of motivation
terminology. Contemporary Educational Psychology, 25(1), 3-53.
Nisbett, R. E., & Wilson, T. D. (1977). Telling more than we know: Verbal reports on
mental processes. Psychological Review, 84(3), 231-259.
Peterson, C., Maier, S. F., & Seligman, M. E. P. (1993). Learned helplessness. A theory
for the age of personal control. New York: Oxford University Press.
Phye, G. D. (1997). Handbook of academic learning: Construction of knowledge. San
Diego: Academic Press.
Pintrich, P. R. (2000). The role of goal orientation in self-regulated learning. In M.
Boekaerts, P .R. Pintrich, & M. Zeidner (Eds.), Handbook of self-regulation (pp. 451502). San Diego: Academic Press.
Pintrich, P. R., & Schunk, D. H. (1996). Motivation in education: Theory, research and
applications. Upper Saddle River, NY: Merrill, Prentice Hall.
Satorra, A., & Bentler, P. M. (1994). Corrections to test statistics and standard errors in
covariance structure analysis. In A. von Eye & C.C. Clogg (Eds.), Latent variables
analysis: Application for developmental research (pp. 399-419). Thousand Oaks:
Sage.
Schau, C., Stevens, J., Dauphinee, T. L., & Del Vecchio, A. (1995). The development and
validation of the Survey of Attitudes Toward Statistics. Educational and
Psychological Measurement,55(5), 868-875.
Schwartz, N. (1999). Self-reports. How the questions shape the answers. American
Psychologist, 54(2), 93-105.
Scott Long, J. (1983). Covariance structure models. Newbury Park: Sage Publications.
Seegers, G., & Boekaerts, M. (1993). Task motivation and mathematics achievement in
actual task situations. Learning and Instruction, 3(2), 133-150.
Steffe, L. P., & Gale, J. (1995). Constructivism in education. Hillsdale: Erlbaum.
Valås, H., & Søvik, N. (1994). Variables affecting students’ intrinsic motivation for
school mathematics: two empirical studies based on Deci and Ryan’s theory on
motivation. Learning and Instruction, 3(4), 281-298.
Volet, S. E. (1997). Cognitive and affective variables in academic learning: The
significance of direction and effort in students’ goals. Learning and Instruction, 7(3),
235-254.
Watkins, D. (1996). The influence of social desirability on learning process
questionnaires: A neglected possibility? Contemporary Educational Psychology,
21(1), 80-82.
Weiner, B. (1986). An attributional theory of motivation and emotion. New York:
Springer Verlag.
Weiner, B. (1992). Human motivation. Metaphors, theories, and research. Newbury
Park: Sage Publications Inc.
LUC BUDÉ
Department of Methodology and Statistics
Maastricht University
PO Box 616
6200 MD Maastricht
20
APPENDIX
Example questions from the exam at the end of course (used for the measurement of
achievement).
1. In a sample of 101 newborn babies, the mean birth weight is 3.8 kg and the standard
deviation is 0.85. The null hypothesis is H0: μ = 4 kg.
If this null hypothesis holds, then:
a. The probability that we will find a sample mean smaller than or equal to 3.8 kg is 50%
b. The probability that we will find a sample mean smaller than or equal to 3.8 kg is 80%
c. The probability that we will find a sample mean smaller than or equal to 3.8 kg is less than
50%
d. The probability that we will find a sample mean smaller than or equal to 3.8 kg is greater
than 50%
2. Given the same sample as in question 1, we are testing H0: μ = 4 kg against H1: μ ≠ 4
kg. The p-value of the sample mean of 3.8 kg is:
a. p ≤ .01
b. .01 < p ≤ .02
c. .02 < p ≤ .05
d. p > .05
3. Given the same sample as in question 1, we are again testing H0: μ = 4 kg against H1:
μ ≠ 4 kg. Suppose the null hypothesis is rejected at α = .10. What is the implication of
this α = .10?
a. In 10 % we will wrongfully conclude that H0: μ = 4 kg holds.
b. In 10 % we will wrongfully conclude that H1: μ ≠ 4 kg holds.
c. In 5 % we will wrongfully conclude that H0: μ = 4 kg holds.
d. In 5 % we will wrongfully conclude that H1: μ ≠ 4 kg holds.
4. The effects of 3 instructional methods on comprehensibility of the information
(SCORE) were investigated. The 3 methods were: a standard method and 2
experimental methods (experimental method 1 and experimental method 2). The
coding of the dummy variables was as follows:
Standard method
Experimental method 1
Experimental method 2
D_EXP1
0
1
0
D_EXP2
0
0
1
It is tested whether the comprehensibility of the information (SCORE) for all methods is
equal (H0), or if at least one of the three methods is different (H1).
Part of the output of the SPSS analysis is presented below:
ANOVA
Model
1
Regression
Residual
Total
Sum of Squares
3341.722
8727.917
12069.639
df Mean Square
2
1670.861
33
264.482
35
F
6.317
Sig.
.005
21
Predictors: (Constant), D_EXP2, D_EXP1
Dependent Variable: SCORE
Coefficients
Unstandardized
Coefficients
Model
B Std. Error
1 (Constant)
37.750
4.695
D_EXP1
14.250
6.639
D_EXP2
23.417
6.639
Dependent Variable: SCORE
Standardized
Coefficients
Beta
.367
.603
t
Sig.
8.041
2.146
3.527
.000
.039
.001
Question: What conclusion can be drawn? Assume α = 0.05.
a. There is a difference between the instructional methods because the p-value of F is
smaller than 0.05.
b. There is a difference between the instructional methods because the p-value of F is
smaller than 0.05/2 =0.025.
c. There is no difference between the instructional methods because the p-value of F is
smaller than 0.05/2 =0.025.
d. There is no difference between the instructional methods because the p-value of F is
smaller than 0.05.
Given the same research and the same results as in question 4, suppose that the F-test
indicates a difference between the three methods. Which groups differ significantly?
Assume α = 0.01.
a. Each method differs significantly from the others.
b. The standard method differs significantly from experimental method 1.
c. The standard method differs significantly from experimental method 2.
d. The experimental method 1 differs significantly from experimental method 2.
5. Given the same research and the same results as in question 4, what is the proportion
of explained variance in the SCORE variable?
a. 0.28
b. 0.72
c. 0.38
d. 0.86
22
USE OF EXTERNAL VISUAL REPRESENTATIONS IN
PROBABILITY PROBLEM SOLVING4
JAMES E. CORTER
Teachers College, Columbia University
[email protected]
DORIS C. ZAHNER
Teachers College, Columbia University
[email protected]
ABSTRACT
We investigate the use of external visual representations in probability problem
solving. Twenty-six students enrolled in an introductory statistics course for social
sciences graduate students (post-baccalaureate) solved eight probability problems in
a structured interview format. Results show that students spontaneously use selfgenerated external visual representations while solving probability problems. The
types of visual representations used include: reorganization of the given information,
pictures, novel schematic representations, trees, outcome listings, contingency tables,
and Venn diagrams. The frequency of use of each of these different external visual
representations depended on the type of probability problem being solved. We
interpret these findings as showing that problem solvers attempt to select
representations appropriate to the problem structure, and that the appropriateness of
the representation is determined by the problem’s underlying schema.
Keywords: Statistics education research; Probability problem solving; Visual
representations; Trees; Outcome listings; Venn diagrams
1. INTRODUCTION
Consider the following probability problem:
An apartment building has four parking spaces in front (call them A, B, C, and D).
There are four apartments in the building (#1, #2, #3, and #4), and each apartment has
a single occupant with a single car. Every evening, all four occupants come home and
park in a randomly chosen space. What is the probability that this evening they park
so that the occupant of Apt #1 is in space A, the occupant of #2 is in space B, the
occupant of #3 in space C, and the occupant of #4 in space D?
How would you go about solving this problem? Many people report visualizing the
cars and parking spaces. After that, strategies for solving the problem tend to diverge (as
do success rates). One of our points in presenting this problem (used in the present study)
is that probability word problems are often simple to pose, yet difficult for many students
to solve. Another point is that visualization and visual solution methods, such as selfgenerated external pictures and diagrams, can be very helpful in solving some probability
problems.
Statistics Education Research Journal, 6(1), 22-50, http://www.stat.auckland.ac.nz/serj
© International Association for Statistical Education (IASE/ISI), May, 2007
23
Probability problem solving (PPS) can be quite difficult for students (Garfield &
Ahlgren, 1988; Konold, 1989; O’Connell, 1993; Pollatsek, Well, Konold, Hardiman &
Cobb, 1987), even when the mathematics involved is simple. Of course, other types of
mathematics word problems are also difficult for many students, perhaps because solving
them requires the problem solver to think abstractly about situations, and then model
these situations using mathematical concepts. However, some researchers (Garfield &
Ahlgren, 1988; Konold, 1989) have suggested that probability problem solving may be
especially difficult because people have natural misconceptions about probabilistic
concepts (e.g., Kahneman, Slovic, & Tversky, 1982).
Recommendations have been made for how to teach concepts in probability (e.g.,
Bantanero, Godino, & Roa, 2004; Gelman & Nolan, 2002; Gigerenzer, 1994; Keeler &
Steinhorst, 2001; Konold, 1995, 1996; Sedlmeier & Gigerenzer, 2001). However, as
pointed out by Garfield and Ahlgren (1988), only a few articles have tried to gather
empirical evidence on the processes by which students solve probability problems. In one
such study, O’Connell (1999; O’Connell & Corter, 1993) described a pedagogical model
of recommended process steps by which students should solve probability problems.
O’Connell (1993, 1999) classified student errors in probability problem solving, showing
that they could be grouped into several categories: text comprehension errors, conceptual
errors, procedural errors, and computational errors. Konold, Pollatsek, Well, and
Lohmeier, and Lipson (1993) documented inconsistencies in probabilistic reasoning and
discussed implications for probability problem solving. Due to this paucity of research on
PPS, Chance and Garfield (2002) call for more research on the cognitive processes of
probability problem solvers using innovative methods such as videotaped clinical
interviews. The present study is intended as a step in that direction.
1.1. IS THERE A SPECIAL ROLE FOR VISUALIZATION IN PROBABILITY
PROBLEM SOLVING?
In studying the cognitive processes of probability problem solvers, one issue that
deserves special attention is the role of visualization. After all, anecdotal evidence
suggests that visualization plays an important role in how experts solve probability
problems (and mathematics problems generally). Also, informal observations of how
students in statistics courses solve probability problems provide ample evidence that they
sometimes spontaneously use visual devices (e.g., outcome trees) in their written work.
Finally, Sedlmeier (2000) has suggested that common cognitive “fallacies” in reasoning
about conditional probabilities may be ameliorated by graphical representations.
Visualization may be especially important for probabilistic reasoning and probability
problem solving because of the inherently abstract nature of the concepts introduced in
probability.
To better understand the literature on visualization in mathematics problem solving, it
is important to distinguish between internal visual representations (i.e., “mental
imagery”) and external visual representations (e.g., graphs, charts, pictures, etc.). Another
distinction about the way external representations may be used in problem solving
concerns whether the external representations are provided to the student by an instructor
or experimenter, or are spontaneously generated by the student in the course of solving
the problem. Although there is an extensive literature on how instructor-provided
graphics can aid in scientific problem solving (summarized below), there has been little
or no research on students’ spontaneous creation and use of pictures, graphics and other
visual devices in the course of mathematics problem-solving activities. In the present
study, we use written and think-aloud protocols to study when and why probability
24
problem solvers spontaneously produce external visual representations in their written
work (when not required to do so), and what types of visual representations they employ.
1.2. PREVIOUS RESEARCH ON EXTERNAL VISUAL REPRESENTATIONS
AND PROBLEM SOLVING
Results from previous research on scientific problem solving by schoolchildren (e.g.,
Lehrer & Schauble, 1998; Penner, Giles, Lehrer, & Schauble, 1996) and by high school
and college students (e.g., Hall, Bailey, & Tillman, 1997; Hegarty & Just, 1993;
Kaufmann, 1990; Mayer, 1989; Mayer & Anderson, 1991, 1992; Mayer & Gallini, 1990;
Mayer, Mautone, & Prothero, 2002; Molitor, Ballstaedt, & Mandl, 1989; Santos-Trigo,
1996; B. Tversky, 2001) suggest that experimenter-provided external visual
representations can aid scientific problem solving. The visual representations investigated
in these studies ranged from diagrams that accompanied text (Hall, Bailey, & Tillman,
1997; Mayer, 1989; Mayer & Anderson, 1991, 1992; Mayer & Gallini, 1990; Mayer,
Mautone, & Prothero, 2002) to actual physical models of scientific systems (Lehrer &
Schauble, 1998; Penner, Giles, Lehrer, & Schauble, 1996). In spite of the wide range of
external visual representations used in these studies, a common finding was that
experimenter-provided external visuals often facilitate problem-solving success. Many of
the studies also conclude that such external visual representations can aid in the
development of student understanding of physical systems and mechanisms.
Incidentally, it is likely that individuals vary in the extent to which they use and
benefit from visual representations. Some researchers in this area (e.g., Hegarty &
Kozhevnikov, 1999; Kozhevnikov, Hegarty, & Mayer, 2002) have taken an individual
differences perspective, grouping problem solvers into one of several types: those who
tend to use verbal representations, and those who primarily use visual/spatial
representations. Kozhevnikov et al. (2002) suggest that the visualizer group can be
further split into object visualizers and spatial visualizers, with spatial visualizers
showing some advantages in scientific and mathematical tasks.
Research conducted specifically in the domain of mathematics has also shown that
experimenter-provided external visual representations can be useful in mathematics
problem solving (e.g., Sedlmeier & Gigerenzer, 2001; Koedinger & Anderson, 1997;
Nemirovsky, 1994). In particular, a number of studies (e.g., Hollebrands, 2003; Hannafin,
Burruss, & Little, 2001; Hannafin & Scott, 1998) have found that the use of Geometer’s
Sketchpad®, a geometry graphing computer program, can be helpful in developing
students’ concepts and problem solving in geometry. Schwartz and Martin (2004)
investigated the use of graphical tools in statistics instruction and found that
experimenter-prompted graphical “invention activities” by students led to significant
gains in understanding of statistical concepts.
Previous work (e.g., Russell, 2000; Zahner & Corter, 2002) in our own lab has shown
that most probability problem solvers choose to use external visual representations while
solving problems (after being taught the use of such visuals in an introductory statistics
course), and that a wide variety of such external visual devices are used. External visual
representations used by probability problem solvers include at least these types: graphs,
tree diagrams, contingency tables, Venn diagrams, and pictures. Arguably, formulas and
mathematical symbols could be included in this list, because they incorporate visuospatial
relationships (cf. Presmeg, 1986). However, their usefulness in solving probability
problems is not in question.
25
1.3. WHY ARE EXTERNAL VISUAL REPRESENTATIONS USEFUL IN
PROBLEM SOLVING?
In order to use research results on visualization and problem solving to improve
mathematics teaching and learning, it is important to ask why external visual
representations are useful in mathematics problem solving. One possible answer to this
question is that external visual representations help to augment cognitive capabilities in
certain ways (e.g., Lowrie & Kay, 2001; Novick, 2001; Qin & Simon, 1995), for example
by aiding memory. Tversky (2001) lists a number of possible functions of external
diagrams and visual devices, including attracting attention, recording information and
supporting memory, communication, providing models, and facilitating inference and
discovery. Another possibility is that using multiple representations of a problem
(including visual ones) leads to a fuller understanding of the problem and an increased
“depth of processing” (Logie & Baddeley, 1990; Mayer, 1989, 2001; Mayer & Gallini,
1990). Other potential explanations for the use of external visual representations include
the possibility that such representations can help problem solvers build a mental model of
the described problem situation (Schwartz & Black, 1996). Finally, for certain problems
the graphical devices may be used as a solution tool in a more specific way: for example,
reading a value from a graph, or counting outcomes in an outcome tree. Alternatively, it
might be that there is no benefit in using external visual representations, rather their use is
just an epiphenomenon, a reaction to training from classroom instruction.
Of course, these accounts of why visualizations might be useful are not all mutually
exclusive or contradictory. But only fragmentary data exist that might support or discredit
any of these explanations. Some hints might come from studies examining when problem
solvers choose to use external visual representations. For example, there is some evidence
that both internal (Hampson & Morris, 1990) and external (Lowrie & Kay, 2001; Zahner
& Corter, 2002) visual representations tend to be used more for unfamiliar or more
difficult problems. This observation seems to support certain explanations (e.g., visuals as
supporting memory, or facilitating inference and discovery) more than others.
1.4. THE PRESENT STUDY
This study focuses on the use of external visual representations in probability problem
solving (PPS). We are interested in what types of problems tend to elicit use of visual
representations, how and when external visual representations are used in PPS, and
finally, if external visual representations facilitate correct solution of the problems. We
used a variety of problem types, in order to investigate if the usefulness of visuals and the
type of visual device chosen by the problem solver depends on specific aspects of the
problem being solved. Specifically, we ask: Are particular types of representations used
with particular problem topics (for example, problems dealing with permutations)? Also,
we investigate if external visual representations are used more often with unfamiliar types
of problems, because the student may have a higher cognitive load in these cases, or
because the elicitation of a familiar problem-solving schema may be less likely.
As background to the present work, we assume that the process of solving a
probability word problem can be broken down into roughly sequential stages (cf. Kintsch
& Greeno, 1985; Mayer, 1992; O’Connell, 1993, 1999; O’Connell & Corter, 1993;
Reusser, 1996). These stages are assumed to be:
i) initial problem understanding (text comprehension),
ii) formulating the mathematical problem,
iii) finding a solution method or schema,
26
iv) computing the answer.
Novick and Hmelo (1994) make a more gross distinction between problem
representation and solution procedure phases of problem solving. Consistent with this
simpler classification, our coding scheme for written protocols of students did not attempt
to code use of visuals separately for stages i-iii, because we do not believe this can be
done reliably with the present data. Rather, our scheme coded two types of uses of graphs
or other external visual representations: a) for problem understanding, mathematical
formulation, or for selection of a solution schema (i.e., any such use in the first three
stages above), and b) for any use in the final stage, that of actually computing a numerical
answer. We refer to the latter type of use of visual devices under the term “computational
method.” As an example of the first type of use, consider the use of a picture of a spinner
or a Venn diagram to depict aspects of the probability word problem. An example of the
second type of use of external visualizations, using them to compute an answer, would be
counting the number of outcomes (leaves) in an outcome tree to find the denominator for
a probability calculation.
Finally, we are interested in knowing if the use of external visual representations
is associated with solution success for these probability problems. If external visual
devices are used because they are helpful, then we ought to be able to find evidence of
that. However, there are several factors that complicate this relationship, including the
student’s prior knowledge of the visual devices used, student spatial and mathematical
ability, student cognitive style and the difficulty of the problem. Alternatively, it might be
that the use of external visual representations is associated with solution failure, because
participants might be more likely to use visual representations when they find a problem
confusing or difficult (cf. Hegarty & Kozhevnikov, 1999; Lowrie & Kay, 2001).
2. METHOD
2.1. PARTICIPANTS
Twenty-six students were recruited from introductory probability and statistical
inference classes during the Fall semester of 2002 from an urban college of education and
psychology in the U.S.A. All participants were graduate students (post-baccalaureate) in
education and social sciences, with widely varying math backgrounds. Participants were
each paid ten dollars. They were informed that they were going to participate in a study
of probability problem solving, and that the primary focus of the study was on the
methods by which students solve problems. Because all the participants were enrolled in
the same introductory statistics class at the college, their recent curricular background in
probability problem solving was well-controlled and known, though the degree to which
each participant mastered the material in that course was not measured. This course
included approximately six lectures in probability. Topics included: events and outcome
spaces, definition of probability for equally-likely and unequally-likely events,
combinatorics, compound events, conditional probability, independence of events, and
Bayes’ Rule.
2.2. MATERIALS
Each respondent was asked to solve eight probability problems. This set of eight
problems (see Appendix) was designed to include four different probability topics
(“problem types”) each represented at two different levels of typicality for that topic. The
four different problem types were labeled: “Combinations,” “Sequential,”
27
“Permutations,” and “Conditional probability.” The problems representing each of these
four topics were thought to have distinct “deep structures” corresponding to four distinct
problem schemas, tapping somewhat different sets of knowledge and solution skills. This
factor will be referred to as “problem type.” As an example, consider Problem P1, an
example of a Combinations problem:
P1. There are 10 books on Mary’s bookcase. She randomly grabs 2 books to read on
the bus. What is the probability that the 2 books are “Little Women” and “War &
Peace”? (Both these books are on her bookshelf.)
In the curriculum to which these participants had recently been exposed, this problem
typically would have been solved using the formula for the number of combinations of n
objects selected k at a time. That formula gives the number of possible outcomes in the
sample space:
n
Ck =
n!
10!
=
= 45 .
k!(n − k )! 2!(8!)
Then the probability of Mary selecting one particular combination (two specific
books) can easily be calculated to be 1/45. Any problem requiring use of this formula (or
a variant of it) is therefore assumed to share the same problem schema, and is said to be
of the same basic Problem Type (Combinations in this case).
For each problem type, there was one typical variant and one atypical variant. The
typical version was a problem that could be solved using a straightforward application of
a standard probability formula known to have been taught in the participants’
introductory statistics class. The atypical version was a problem that was very unlikely to
be isomorphic to any problem encountered in their course, and that could not be solved
using a single application of a standard probability formula. This manipulation of
typicality may be clarified by Table 1, which presents summaries of what we judge to be
appropriate formula-based solutions for the typical and atypical variants of each problem
type. These solutions are presented to show the basic underlying structure of each
problem type, without reference to surface content, and to illustrate how each problem
might be solved by application of one or more standard probability formulas.
Table 1 also makes clear the types of specific problem manipulations that were used
to create the atypical variant of each problem type. For the Combinations problem, the
predicted solution for the typical version requires the problem solver to use the standard
formula for the number of combinations of n things selected k at a time to calculate the
number of possible outcomes. Our predicted solution for the atypical Combinations
problem requires using this formula twice, once in the numerator and once in the
denominator. For the Sequential problems, problem solvers must use the formula for
calculating the probability of three independent events. In the typical variant the three
events are identical, whereas in the atypical variant they are different events with
differing probabilities. The typical variant of the Permutations problem asks how many
different ways four items can be matched up with four “slots.” The atypical version of
this problem asks the same question, but orders the objects only with respect to the first
two slots. We consider this an atypical problem because the computational method does
not correspond to straightforward application of the formula for number of permutations
of n objects, which is known to have been taught in the participants’ introductory
statistics course. Note that this atypical variant is actually simpler computationally than
its typical version. For the Conditional Probability problems, the typical version closely
resembles examples used in the students’ introductory statistics course, and requires the
problem solver to use the formula for conditional probability (twice). The atypical variant
adds a final step, in which the formula must be used a third time.
28
Table 1. Example formula-based solutions for typical and
atypical variants of each problem type
Problem Type
Typical
n
Ck =
Combinations
Atypical
n!
k!(n − k )!
where n = total number of books
k = number selected
m!
C
k!(m − k )!
P ( A) = m k =
n!
n Ck
k!(n − k )!
where n = total number of books
m = number of novels
k = number selected
Sequential
P( A1 ∩ A2 ∩ A3 ) = P( A) P ( A) P( A)
P( A1 ∩ A2 ∩ A3 ) = P ( A1 ) P ( A2 ) P( A3 ),
A1, A2, A3 not necessarily equal
Permutations
# outcomes = n! = (n)(n-1)…(2)(1)
where n = number of objects
# outcomes = (n)(n-1)
where n = number of objects
Conditional
Probability
P( B) = P( A ∩ B) + P( A c ∩ B)
P( A ∩ B ) = P ( B | A) × P( A)
P( A c ∩ B) = P( B | A c ) × P( A c )
P( B) = P( A ∩ B) + P( A c ∩ B)
P( A ∩ B) = P( B | A) × P ( A)
P( A c ∩ B) = P( B | A c ) × P( A c )
P( A ∩ B)
P( A | B) =
P( B)
Across two different forms (A and B) of the test booklet, each of the eight problems
was formulated with two different cover stories, or surface content. Surface content was
counterbalanced with problem typicality across test forms. For example, the typical
Combinations problem P1 given above involved books on a bookshelf, so a participant
who saw that problem would see an atypical Combinations problem involving cookies in
a cookie jar. For another participant who saw the second test form, the atypical
Combinations problem would involve books on a bookshelf, and the typical version
would involve cookies in a cookie jar. The Appendix shows only test form A.
2.3. PROCEDURE
A structured interviewing protocol (cf. Ginsburg, 1997) was developed for use in the
interviews, and was designed mainly to elicit a reasonable level of detail in the participant
protocols. The same interviewer worked with all of the participants, interviewing only
one participant at a time. Participants were asked to think aloud while solving the
problems, and also to show their written work with provided pen and paper. The task was
not timed. However, most participants finished in less than an hour. A probability
formula sheet was available to them at all times (but left face down), though no
participant was observed to use it. A videotape recorder was used to capture the
participants’ work and student/interviewer comments. The present analyses mainly focus
on the participants’ written work, though the verbal transcripts were analyzed as well.
The interviewer stepped in with verbal prompts in any of four circumstances, to elicit
continuation of the work or more detail about the participant’s solution process. The first
29
circumstance was if the participant could not see any way to begin solving the problem.
In this situation the script called for the interviewer to ask, “In general, what would be a
good first step in solving this problem?” with other follow-up questions (“How would
you apply it in this case?”) if the first prompt did not elicit useful work. The second type
of circumstance in which the interviewer stepped in was when the participant paused for
a long time (more than 30 seconds) without thinking aloud or writing. This could indicate
either that the participant was thinking silently or was at an impasse, and was responded
to with “What are you thinking?” and other follow-up prompts (“Let’s back up and look
at this again. How else could you solve this?”). The third situation in which prompting
occurred was when the participant’s verbal or written process explanations lacked
sufficient detail, for example, consisting of only a few calculations with no explanation
(“Can you explain how you arrived at this?”). The fourth situation eliciting interviewer
intervention was when the participant indicated that he or she was finished with the
problem. In most cases, this occurred when the participant had arrived at what he or she
believed was the correct answer. In other cases, this was because the participant gave up
on solving the problem. In either case, the interviewer then asked the participant to
explain in detail all of the steps used in the solution attempt.
Coding of the written protocols The focus of the present study is on use of external
visual representations in problem solving and on the methods used to solve problems.
Thus, the analyses reported in the present study focus mainly on coding of the
participants’ written work. Three particular aspects of the written problem solutions were
coded, based upon a scheme developed in previous research (Russell, 2000). The first
coded aspect was whether or not the participant gave the correct answer to the problem.
The second aspect coded, described in more detail below, was the type of external visual
representation used (if any) by the participant. The third aspect coded, also described in
more detail in the next section, was the type of general computational method used to
solve the problem. Here the identified types were: formula, graphical, or procedural.
External visual representations Written protocols for each problem solution were
coded for use of different types of external visual representations. The coded types of
external visual representation included pictures, outcome listings, trees, contingency
tables, Venn diagrams, novel schematic representations, and spatial reorganization of the
given information.
An external visual representation was coded as a picture if it attempted to represent
the real-world situation conveyed in the problem in a non-symbolic, pictorial way. For
example, in a problem about use of a spinner with separate areas marked “red,” “blue,”
and so on, any picture of a spinner type device would count as a picture (see Figure 1 for
an example). A visual device was coded as an outcome listing if it gave a list of outcomes
in some relevant outcome space, for example: {HH, HT, TH, TT} as the outcomes space
for the experiment of flipping a coin twice. A visual representation was coded as a tree
diagram if the participant attempted to organize the information from the problem in
either a complete or a partial outcome tree. An example of the use of a tree diagram is
shown in Figure 2. A visual representation was coded as a contingency table if the
participant presented the information from the problem as probabilities or frequencies in
a two-way table. A visual representation was coded as a Venn diagram if the participant
used a Venn diagram to represent set relationships.
30
Figure 1. A participant’s written work for the typical version of the Sequential problem,
illustrating use of a picture
Figure 2. A participant’s written work for the typical version of the Sequential problem,
illustrating use of an outcome tree (and a picture)
Besides these standard pictorial and schematic representations used in previous
studies (Russell, 2000; Zahner & Corter, 2002), we created two additional coding
categories to cover cases not handled by the above classes. The first is a code indicating
any attempt to invent and use what we termed a “novel schematic representation.” Use of
the term “novel” is meant to denote a schematic visual device that was not taught in the
introductory class the participants were taking or had taken, nor used in standard
probability texts. It is not meant to imply that the student invented and used a previously
unknown type of visual device. An example of this category is a graphic used by several
subjects for the Permutations problems: a list of four names whose elements are
connected by lines or arrows to elements in a list of four numbers (see Figure 3). This
type of representation (that we would classify as a directed graph) is apparently an
attempt to develop or discover the correct outcome space for the problem. This type of
representation uses spatial information and schematic elements (lines or arcs) to represent
relational aspects of the problem, and is thus different from a simple outcome listing. The
second additional coding category was defined to include any spatial reorganization of
the given information. Use of a spatial organization scheme for information is not a
31
formal graphical representation nor a purely pictorial representation. However, we have
included this coding category because we have observed frequent use of spatiallyorganized rewriting of information to aid in problem solving. In the present study, many
participants were observed to line up corresponding given probabilities or conditional
probabilities (see Figure 4). This practice may make it easier for novice problem solvers
to check for needed or missing information, to break down problem solution into
subparts, or to make visual associations to relevant formulas.
Figure 3. A participant’s written work for typical version of the Permutation problem,
illustrating use of a novel schematic representation
Figure 4. A participant’s written work for the atypical version of the Conditional
problem, illustrating use of spatial reorganization of given information
Computational method We also coded the computational method used by the
problem solver, that is, the means by which the problem solver actually computes the
answer required by the problem. We did this because in the course of coding the student
protocols, we noticed that sometimes visual representations were used very early in the
problem solving process, for example while the problem solver seemed to be still trying
to understand the given problem information or to classify the problem, and sometimes
later in the solution process, for example when the subject was trying to compute the
actual numerical answer. In an effort to begin to understand what specific role or function
external visual representations are serving in probability problem solving, we decided to
separately code the method by which the problem solver actually computed the numerical
answer required in each of these problems. We classified this later stage of each problemsolving protocol into three broad classes of computational method: formula, procedural,
and graphical.
32
The computational method was characterized as formula-based problem solving if the
participant wrote down an explicit (standard) formula, then substituted in quantities and
solved the problem. An example of a formula-based computational method would be the
use of the combinations formula followed by the necessary calculations: 5C2 = 5!/(2!)(3!)
= (5)(4)/(2)(1) = 10. This complete and rather formal method was distinguished from a
procedural approach, which was used to code solutions carrying out a calculation
involving only numbers without reference to any general formula or underlying principle.
An example of a procedural approach would be if the participant calculated the
probability of getting three heads in three flips of a coin by simply multiplying
(1/2)(1/2)(1/2) = 1/8 without indicating any rationale for that procedure. A computational
method was considered graphical if the subject used an external visual device to solve
the problem, but only if the graphical device was judged to be instrumental to the method
by which the student arrived at the actual numerical solution. An example of a graphical
computational method would be if the subject multiplied two conditional probabilities
that were taken from branches of a tree diagram.
No computational method was coded if the participant did not attempt to solve the
problem. Occasionally, multiple computational methods were coded for a single problem.
This occurred only when a participant attempted the problem, then abandoned that
attempt, and attempted another computational method.
In order to assess reliability of the coding of the written protocols for external
representations and computation method, a second rater coded all student solutions.
Initial percent agreement between the two raters was over 90% for both external
representation and computation method. Discrepancies were discussed by the two raters
and the resulting consensus was used in all analyses reported.
Coding of audio protocols In order to better understand how the external visual
representations are being used by problem solvers, we also transcribed and examined the
audio portion of the videotapes capturing the participants’ think-aloud protocols. Each
utterance in a participant’s audio transcript was coded to indicate if the participant was
engaged in either of two broad phases or stages of problem solution: 1) a problemrepresentation phase that involves understanding the problem text and reformulating the
problem in mathematical form; or 2) a solution phase, that involves selecting a solution
strategy and implementing it. The video track of the tapes focused on participants’
written inscriptions, including use of external representations. The video tapes were used
to match uses of external visual devices with verbal statements by the participants about
their thoughts and actions and the general phase of problem solving that they were
engaged in: either problem representation or strategy selection and solution.
2.4. RESULTS
Preliminary analyses showed that individual problems and problem types varied
considerably in difficulty. The rightmost column in the Appendix shows the proportion of
subjects who correctly solved each problem. These proportions vary from a low of .08 for
Problem P8 (Conditional probability, atypical) to .73 (for Problems P3 and P4, the
typical and atypical Sequential problems). Regarding problem type, it was found that
participants were most successful at solving the Sequential problems (.73 correct overall),
followed by the Permutations problems (.48), then the Combinations problems (.29), and
finally the Conditional probability problems (.25). These differences in solution rate
among problem types were significant: in a log-linear analysis with factors Problem
33
Type, Typicality, and Correctness (the dependent variable), the Problem Type ×
Correctness association was significant (χ2(3) = 31.56, p < .05).
We were also interested in whether the atypical variant of each problem type was
more difficult for problem solvers than the typical variant. Results indicated that typical
and atypical variants did not differ in mean difficulty across all four problem types: mean
proportion correct for the four typical problems encountered by each subject was .43,
whereas for the four atypical problems it was .44. However, it is clear from the solution
rate for individual problems (see Appendix) that the typical-atypical difference in
solution rates varies across problem types. This interaction was tested by the three-way
association of Problem Type × Typicality × Correctness in the loglinear analysis
described above. This association was significant, (χ2(3) = 17.00, p <.05). Consequently,
it is necessary to examine the effects of typicality separately for each problem type.
In particular, a difference in the expected direction was found for the Conditional
Probability problems, with 42% of participants correctly solving the typical version of
the problem versus only 8% for the atypical version. Unexpectedly, for the Combinations
problem the solution rate for the atypical variant of the problem was much higher (at
73%) than for the typical variant (at 12%). In order to understand this unexpected result,
we analyzed participants’ written protocols to identify the specific solution methods used
by participants for these problems. We found that most participants did not use the
combinations formula at all to solve the atypical variant of the Combinations problem;
rather they tended to solve this problem by treating it as a “sequential” problem involving
sampling without replacement. For example, the problem can be solved using the
formula: P ( A1 ∩ A2 ) = P ( A1 ) P ( A2 | A1 ) .
Inspection of the individual student protocols revealed that 92% (24 out of 26) of the
participants selected this alternate method to solve the atypical Combinations problems.
This probably occurred because the atypical problem is extremely difficult using the
Combinations approach: the only two participants who tried this approach both failed to
solve it. In contrast, exactly half of the 24 participants who adopted the sequential-events
approach for the atypical variant succeeded in solving it. Note that many participants
(73%) also tried to solve the typical Combinations problem using a sequential-events
approach. However, all of these subjects failed to solve the problem, contributing to the
overall low solution rate for the typical version. The difference in apparent difficulty of
the sequential approach to these two problems probably involves that fact that in the
typical Combinations problem, order is not important (but the use of the sequential
solution method tends to elicit a solution attempt involving ordered pairs). Thus, many
participants gave the answer 1/90 for this problem using the sequential approach, whereas
the correct answer is 1/45. In the atypical variant, in contrast, there is a symmetry to the
outcomes in the outcome space such that order is irrelevant.
What kinds of external visual representations are used? For each specific form of
external visual representation, we calculated the percentage of participants who used that
representation at least once. As shown in Table 2, we found that participants most often
used reorganization of the given information (used at least once by 96.2% of the
participants), followed by use of pictures (by 84.6% of the participants), novel schematic
representations (65.4% of the participants), trees (53.8%), outcome listings (38.5%),
contingency tables (7.7%) and finally Venn diagrams (3.8%).
34
Table 2. Frequency and percentage of participants using each type of external visual
representation at least once, with frequency and percentage use of each representation
across all problem solutions
Representation
Reorganize
Outcome Listings
Contingency Tables
Venn Diagrams
Trees
Novel Schematic
Pictures
By participant (N=26)
Frequency
%
25
96.2
10
38.5
2
7.7
1
3.8
14
53.8
17
65.4
22
84.6
By problem solution (N=208)
Frequency
%
72
34.6
20
9.6
6
2.9
1
0.5
27
13.0
24
11.5
64
30.8
Are different types of external visual representations used with different types of
problems? We investigated the relationship between the type or topic of the probability
problem (Combinations, Sequential, Permutations, and Conditional) and the type of
representation that participants chose to use for it. In this analysis no distinction was
made between the typical and atypical versions of each problem type. Table 3
summarizes how often each type of external representation was used for each type of
problem. Because there were two problems of each type, each entry in this table is
calculated across a total of 52 problem solutions.
Table 3. Frequency and percentage of problems of each type for which a given type of
external representation was used (out of N=52 problem solutions), with χ 2 goodness-of
fit tests evaluating differences in the frequencies of use of each representation across the
four problem types
Representation
Reorganize
Outcome Listings
Contingency Tables
Venn Diagrams
Trees
Novel Schematic
Pictures
*p < .05
Combinations
Freq
%
26
50.0
9
17.3
0
0.0
0
0.0
5
9.6
0
0.0
16
30.8
Sequential
Freq
%
4
7.7
7
13.5
0
0.0
0
0.0
6
11.5
0
0.0
31
59.6
Permutations
Freq
%
4
7.7
4
7.7
4
7.7
0
0.0
5
9.6
24
46.2
15
28.8
Conditional
Freq
%
38
73.1
0
0.0
2
3.8
1
1.9
11
21.2
0
0.0
2
3.8
χ2(3)
47.6*
9.2*
----3.7
72.0*
26.4*
For each row of the table, we used a chi-square goodness-of-fit test to test if each
visual representation was used with unequal frequencies across problem types (i.e.
columns). The chi-square goodness-of-fit test revealed that the frequency of use of
reorganization of given information differed significantly across problem types (χ2(3)=
47.6, p < .05). This representation was used most often for the Conditional Probability
problems (73.1% of the time) and the Combinations problems (50%). Usage was also
distributed unequally across problem type for outcome listings (χ2(3) = 9.2, p < .05), with
the most frequent use being for Combinations (17.3%) and Sequential (13.5%) problems.
Use of Novel schematic representations was also distributed unequally across problem
type (χ2(3) = 72.0, p < .05), because these representations were used only for the
Permutations problems (46.2% of the time). Novel schematic devices may be tried
especially often for the specific permutations problems used here because these problems
35
are difficult for novices to recognize as permutation problems. That is because the cover
stories for these particular permutation problems involve matching two sets of entities
(e.g., tutors with students) rather than simply ordering one set of objects. This situation
does not plug neatly into any formula or solution schema that students had been taught.
This situation apparently spurred participants to try to understand these relatively unusual
problems by inventing or adapting “novel” graphical representations.
Also, the use of pictures was distributed unequally across problem type (χ2(3) = 26.4,
p < .05), due to very frequent use of pictures for the Sequential problems (in 59.6% of
problem solutions), Combinations (30.8%), and Permutations (28.8%) and infrequent use
(3.8%) for the Conditional Probability problems. The use of trees did not vary
significantly across problem type (χ2(3) = 3.7, p < .05). Inspection of Table 3 reveals that
trees were used about 10% of the time or more for all four problem types. This result
seems to show that at least for these types of probability problems, trees were perceived
by study participants as widely applicable. The use of Contingency tables and Venn
diagrams was too infrequent to be tested in the manner.
Sometimes problem solvers used more than one form of external visual representation
in a single problem solution. Figure 5 shows the percentage of use of single and multiple
representations across all problem solutions, separately by problem and problem type.
Across the eight problems, multiple external visual representations were used in 23.6% of
the problem solutions. The most common combinations of multiple representations were
pictures with reorganization (used in 13% of the problem solutions), and pictures with
trees (used in 6%; see Figure 2 for an example). All other instances of multiple external
representations occurred less than 2% of the time. Multiple representations were used
most often for Combinations and Sequential problems. This may simply reflect the fact
that pictures were used quite often for these problem types, as shown in Table 3.
100
% of Use
80
60
Single
Multiple
40
20
0
P1
P2
Combinations
P3
P4
Sequential
P5
P6
Permutations
P7
P8
Conditional
Problem/Problem Type
Figure 5. Percentage of use of single and multiple external representations in the
problem solutions (N=208), by problem and problem type
The above results showing differences in frequency of use of specific representations
across the four problem types demonstrate that participants are selecting representations
based on the type of problem they are trying to solve, presumably reacting to differences
in the problem schema for the four problem types. This suggests that participants’
solution methods (at least, their use of external visual representations) vary depending on
36
the problem’s underlying schema or “deep structure.” We return to this issue in the
Discussion section.
Is solution success associated with use of external representations? If external
visual devices do indeed serve some purpose for problem solvers, then we might expect
an association between solution success and the specific external representation used (if
any). We explored this idea by estimating the conditional probability of solution success
given use of each type of external visual representation. The results, shown in Table 4,
show that use of particular external visual representations was associated with higher
rates of solution correctness for some problem types (compared to baseline performance
for that problem type), and with lower rates of success for others. For example, the use of
reorganization is associated with a higher rate of solution success for Combinations
problems (.32 versus .29), but with a lower rate of success for Sequential (.50 versus .73),
Permutations (.00 versus .48), and Conditional (.15 versus .25) problems.
Table 4. Proportion of correct solutions, given the use of a particular representation,
separately by problem type. Dashed lines indicate a cell with fewer than four uses of that
representation (i.e., n ≤ 3).
Representation
Reorganize
Outcome Listings
Contingency Tables
Venn Diagrams
Trees
Novel Schematic
Pictures
Mean P(correct)
Combinations
.32
.43
--.60
-.23
.29
Sequential
.50
.42
--.67
-.63
.73
Permutations
.00
-.00
-.25
.44
.59
.48
Conditional
.15
---.25
--.25
Table 4 shows that for the Combinations problems, use of reorganization, outcome
listings, or trees were all associated with higher rates of solution success, whereas use of
pictures was associated with a lower rate of success. Presumably, the first three types of
representations are useful here because the essence of such combinatorics problems is to
identify the number of outcomes in the outcome space. However, trees are not usually
useful for problems involving simultaneous sampling of multiple objects (where order is
not important). We therefore reexamined participants’ solutions to try to understand this
association. We found that trees were used in only five solutions for the Combinations
problems, and all of these were cases where the problem solver was treating the problem
as a sequential problem rather than using the combinations formula.
For the Sequential problems, use of any external visual representation was associated
with a lower rate of solution success. Sequential problems were the easiest type of
problem overall, with P(correct) = .73, so it may be that participants did not feel any need
to call upon visual representations unless they were among the few who experienced
difficulty with these problems.
For the Permutations problems, use of reorganization, contingency tables, and trees
was associated with lower rates of solution success. Contingency tables in particular do
not seem appropriate for permutation problems, which involve ordering a single set of
objects. Trees are rarely used to represent sequential sampling without replacement,
though in principle they could be applied. However, use of pictures was associated with a
higher rate of solution success for these Permutations problems. Pictures may be
37
especially useful for these particular (unusual) permutations problems, which are unusual
in that they describe matching two sets of objects rather than ordering a single set. It may
be that pictures facilitated the realization that the ordering of one of these sets is arbitrary.
Finally, for the Conditional probability problem, external visual representations were
not often used. Use of reorganization was associated with slightly lower rates of solution
success. It is surprising that trees were not often used for these problems because their
use for such problems was explicitly described in the course.
Thus, this table seems to offer mixed evidence concerning the usefulness of external
visual representations in probability problem solving. The positive associations found
seem easily explainable. We argue that the few observed negative associations between
external visual representations and solution success do not prove that use of external
representations is harmful in probability problem solving. Rather, the negative
associations may arise because external visual representations are more often called upon
when the problem is especially difficult for the problem solver (cf. Hegarty &
Kozhevnikov, 1999; Lowrie & Kay, 2001).
In addition to the analysis of solution correctness given the use of a particular external
visual representation, we correlated the number of times a participant used any external
visual representation (which ranged from 2 to 18 for participants) and the participant’s
overall solution success (defined as number of problems correct out of 8 for a
participant). Results from this analysis show that there is a significant negative overall
correlation between the use of an external visual representation and solution success (r =
-.40, p < .05). We also found marginally significant negative correlations between use of
certain specific representations and solution success. Specifically, the use of
reorganization of the given information was negatively correlated with solution success (r
= -.37, p = .06), as was the use of outcome listings (r = -.37, p = .06).
What computation methods are used in PPS? Are external visual devices used in
computing problem solutions? We calculated how often each of the three computational
methods (formula, procedural, graphical) was used for each problem. Results showed that
students used the procedural computational method most often (on average in 5.5 out of 8
problems) followed by formula-based computational methods (1.19 out of 8 problems)
and finally, graphical solutions (0.42 problems out of 8.) Thus, external visual
representations were rarely used to compute solution. For 13.5% of problems overall,
subjects did not complete the problem to the point of computing a solution. Multiple
computation methods (coded only when the participant made multiple solution attempts)
were observed only 2.4% of the time.
Are different computation methods used with different problem types? We
calculated frequency and percentage of use of each of the three types of computation
method across problem types. Note that more than one type of solution method could be
coded for a given solution, and that if the problem solver did not attempt to compute a
numeric solution no computation method was coded. Results showed that formulas were
used most often for the Conditional problems (21.2% of the time) and Combinations
(21.2%). A procedural computation method was used most often for Sequential problems
(84.6%), Combinations (75.0%), and Permutations (75.0%). Finally, a graphical
computational method was used most often with the Conditional Probability problems
(9.6%). Note that for the Conditional Probability problems the observed student solutions
were not purely graphical; rather the tree graphs were typically used in conjunction with
procedural calculations. We also performed a chi-square goodness-of-fit test to determine
if each computational method was used equally often across all four problem types. The
38
results (Table 5) show differential use of the procedural computation method across
problem types (χ2(3) = 8.57, p < .05). Frequency of use did not differ significantly across
problem types for the procedural computation method (χ2(3) = 6.05, p > .05), nor for the
graphical computation method (χ2(3) = 5.21, p > .05). The pattern of results in Table 5
suggests that a procedural solution method is used relatively less often for the
Conditional Probability problems.
Table 5. Frequency and percentage of problem solutions (n = 52) of each type for which
a given computational method was used, with χ 2 goodness-of-fit tests of differences in
frequency of use of each method across the four problem types
Computational method
Formulas
Procedural
Graphical
*p < .05
Combinations
Freq
%
11
21.2
39
75.0
3
5.8
Sequential
Freq
%
3
5.8
44
84.6
3
3.8
Permutations
Freq
%
6
11.5
39
75.0
0
0.0
Conditional
Freq
%
11
21.2
21
40.4
5
9.6
χ2(3)
6.05
8.57*
5.21
Is solution success associated with computational method? We also checked for
associations between solution success and computational method, separately by problem
type (Table 6). Use of a formula-based computation method was associated with a higher
rate of solution success only for Combinations problems. This makes sense, because the
combinations problems are arguably best solved via formulas. Use of a procedural
method was associated with the highest rates of solution success for the other three
problem types. This is probably because if it is intuitively clear to a student how to solve
a problem, only the computations need be written down (and the solution would be coded
as a procedural one). Using a graphical method to aid in computing the solution was
observed infrequently, except for the Conditional probability problems. For these
problems, the tree can be used to organize the procedural calculations.
Table 6. Proportions of correct solutions given the use of a particular computation
method, separately by problem type. Dashed lines indicate a cell with fewer than four
uses of that representation (i.e., n ≤ 3).
Computational method
Formula
Procedural
Graphical
Mean
Combinations
.38
.30
-.29
Sequential
-.77
-.73
Permutations
-.57
-.48
Conditional
.11
.46
.30
.25
Are there differences between typical and atypical problems in the use of external
visual representations or computational method? We investigated whether there is a
difference in the rates of use of visual representations for the typical and atypical
problems. To test this, for each type of representation we compared the summed
frequency of its use for the four problems presented in their typical versions to the
summed frequency of its use for the four problems presented in their atypical versions.
The results show that the only significant difference in use of an external representation
between typical and atypical problems was for pictures (paired-sample t(25) = -3.86, p <
.05). Specifically, pictures were used more often for atypical problems (for 38.5% of
problems) than for typical problems (23.0%), as shown in Table 7. This result is not at all
39
surprising – pictures may be used especially often to try to better understand or structure
a difficult problem, especially one that does not plug in easily to a familiar solution
schema. On the other hand, if the solution method is obvious, nothing is gained (and time
and effort are expended) in drawing a picture. The only other representation used more
often for atypical problems (though the difference is not significant) is reorganization. As
for pictures, it can be argued that this is a general-purpose method that is often useful
when the problem text itself is difficult to understand.
Table 7. Frequency and percentage use of each type of external representation,
separately for the four typical and four atypical problems experienced by each
participant. (* = significant difference between the total number of uses for typical and
atypical using a dependent samples t test, df = 25)
Representation
Reorganize
Outcome Listings
Contingency Tables
Venn Diagrams
Trees
Novel Schematic
Pictures*
Typical
Freq
%
34
32.8
12
11.5
4
3.8
1
1.0
15
14.5
13
12.5
24
23.0
Atypical
Freq
%
38
36.5
8
7.8
2
2.0
0
0.0
12
11.5
11
10.5
40
38.5
*p < .05
Table 8 reports the rates of use of different computation methods for typical versus
atypical problems. There were no significant differences in the use of different
computation methods for typical and atypical problems.
Table 8. Percentage use of each type of computational method, summed across problems,
separately for the four typical and four atypical problems experienced by each
participant
Computational Method
Formula
Procedural
Graphical
Typical
Freq
%
15
14.5
73
70.3
6
5.8
Atypical
Freq
%
16
15.5
70
67.3
10
10.0
2.5. ANALYSIS OF AUDIO PROTOCOLS
As described in the Methods section, each utterance in the audio track of the session
videotapes was coded as relevant to either the problem solvers’ problem-representation
phase or as part of the solution execution phase (cf. Novick & Hmelo, 1994). We also
matched any use of an external visual representation in a solution (as captured in the
video track) to any utterances made simultaneously. This enabled us to classify uses of
external visual representations as being associated with either or both of these broad
temporal stages of problem solving.
Of the 2,756 utterances in the audio transcripts, approximately 63% of them were
coded as part of the Problem Representation phase (1,734 utterances) and approximately
32% of the utterances were coded as Solution Execution (881 utterances). The remaining
40
utterances were not coded as part of either phase. This occurred if the utterance was a
meta-comment or irrelevant, for example concerning how difficult the problem was or the
temperature of the room in which the study was being conducted.
Of the 1,734 utterances that were coded as part of the Problem Representation phase,
417 (24.08%) of them were matched with the use of an external visual representation. Of
the 881 utterances that were coded as involving the Solution Execution phase, 31
utterances (3.68%) were matched with the use of an external visual representation.
Although participants did use external visual representations during both phases of
problem solving, they tended to use them more often during the Problem Representation
phase than during the Solution Execution phase.
These associations were broken down by type of representation used (Table 9).
Results indicated that participants more often tended to use the external visual
representations to help understand and organize the problem text (i.e., in problem
understanding) than to select or execute solutions. This trend was especially strong for
pictures, reorganization of the given information, and novel schematic representations.
For example, pictures were used significantly more often during the problem
representation phase because participants claimed that they helped them visualize the
problem more clearly. One subject explained, “I drew the ten cookies because I needed
literally to visualize it, and then based on what the information is in this problem, there’s
obviously….there’s ten different types of cookies.” (Subject #16)
Table 9. Total frequency of use of particular representations, by problem solving phase
Representation
Reorganize
Outcome Listings
Contingency Tables
Venn Diagrams
Trees
Novel Schematic
Pictures*
Total
*p < .05
Problem Representation
169
17
11
4
50
45
120
417
Strategy & Execution
9
3
3
0
13
1
4
32
3. DISCUSSION
Our results show that students sometimes choose to use self-generated external visual
representations while solving probability word problems. Presumably, this is because
problem solvers believe that these representations are useful in solving the problems,
because in this study they were requested merely to “show their work,” and not explicitly
requested to produce any diagrams or other visual devices. A skeptical observer might
worry that the verbal prompts occasionally issued by the experimenter here could have
served as a general prompt to try alternative representations. However, Russell (2000)
found similar levels and patterns of use of these types of representations in students’
actual answers to course assignments, lending confidence to the conclusions that problem
solvers choose to use such representations because they are thought to be useful.
The results also document what types of spatial and graphical devices are used in
probability problem solving. Using a very broad definition of external visual
representations, the types we identified included (in decreasing order of frequency of
use): reorganization of the given information, pictures, novel schematic representations,
trees, outcome listings, contingency tables, and Venn diagrams (cf. Russell, 2000). Of
41
course, our reported relative frequencies of use for these representations may not
generalize to other curricula, other specific sets of problems, and other problem solvers.
We found evidence that the frequency of use of each of these different external visual
representations depends on the type of probability problem being solved. Type of
problem refers to the basic problem schema, and not to surface characteristics of the
problem. The types of probability problems studied here were Conditional Probability,
Combinations, Permutations, and Sequential problems. Our results showed that pictures
were used most often for Sequential, Combinations, and Permutations problems; outcome
listings were used more often for Combinations and Sequential problems; trees were used
most often for Conditional problems; novel schematic representations were used mainly
with Permutation problems; and reorganization of given information was used more
often with Conditional and Combinations problems. One way to interpret these findings
is that problem solvers attempt to select representations appropriate to the problem’s
structure, and that the appropriateness of the representation is determined by the
problem’s solution schema, not by surface characteristics.
However, any conclusions as to specific associations between the type of visual
device and the problem type must be tempered by consideration of the particular set of
problems used here to represent these general types. First, the Permutation problems
studied here were unusual in that they described situations in which two sets of entities
(e.g., tutors and students) were to be matched in a one-to-one fashion, but the ordering of
one set was arbitrary, making the problem isomorphic to an ordering problem. This
“schema mismatch” may have made these problems particularly difficult for our problem
solvers, spurring more attempts to use novel schematic representations and pictures.
Second, problems were experienced with our manipulation of typical and atypical
problems for the Combinations problems. The atypical Combinations problems were
most often solved by an alternative method, using a sequential-events approach, that
resulted in a higher rate of success than for the typical variant of this problem type. Thus,
although it was our intention to manipulate problem typicality in such a way that the
atypical problems were at least as difficult as the typical ones, this did not happen for this
problem type. In future studies, we hope to more fully refine and explore the notions of
problem typicality and difficulty, and to try to disentangle their effects experimentally by
careful development and piloting of materials.
In future research, we also hope to more fully investigate aspects of the schematic
devices that play a part in determining the appropriateness of a representation for a given
problem. We believe that the seven types of external visual representations studied here
differ in some important ways. Three of the visual representations (reorganization,
outcome listings, and contingency tables) can be considered forms of tabulation. Another
three (Venn diagrams, trees, and novel schematic representations) could be classified as
schematic devices, and the final type (pictures) refers to iconic representations of concrete
aspects of the problems. We term the second group of representations (Venn diagrams,
trees, and novel schematic representations) schematic because structural aspects of the
graphs symbolically represent meaningful aspects of the problem.
Novick and Hurley (2001) propose that different types of schematic devices (or
“diagrams”) have structural aspects or properties that determine their range of
applicability. The associations we have found between use of the different types of
representations and specific problem types suggest that properties of the diagrams and
properties of the problem schema are being matched (though not always successfully) by
participants. For example, trees seem naturally appropriate for sequential problems such
as the results of multiple coin flips or successive spins of a spinner, whereas contingency
42
tables and Venn diagrams are particularly appropriate for representing joint or compound
events.
These seven identified types of external representations also differ in terms of their
degree of structure. This has implications for how broadly or how narrowly the visual
representation may apply. Specifically, we argue that reorganization of the given
information, pictures, and outcome listings are relatively general representations that can
be applied to a wide variety of problems, whereas trees, contingency tables, and Venn
diagrams have more inherent structure, thus may be applicable to a more limited set of
problems. Finally, the category of novel schematic representations is by definition not
limited to any specific type of structure, thus this category of representation is also widely
applicable (although any specific novel graph may have limited applicability). However,
novel schematic representations seem to be used only when the problem solver
encounters a very atypical or unusual problem that does not seem to plug into any
familiar schema.
If we are right that reorganization of the given information, pictures, and outcome
listings are very general tools, whereas the schematic devices (trees and Venn diagrams)
and contingency tables are more limited in scope of application because they have more
constrained structures, then those variations in scope of application ought to show up in
our data. Calculating the average percentage of problems for which each type of
representation was used (Table 2), provides some supporting results for this idea. The
three types of representation argued here to be general ones (reorganization, pictures, and
outcome listings) were used in 25% of problem solutions on average, whereas the three
specific types were used in only 6% of problem solutions on average. However, the
picture given by Table 3 is a bit less clear. Here it can be seen that uses of reorganization
of the given information and pictures are spread across all four problem types, and
outcome listings are used for three out of four types, whereas the more constrained types
of representation Venn diagrams and contingency tables are used for only one or two
types of problem. However, trees are used across all four problem types. Thus, except for
trees, the predicted pattern does hold.
3.1. DO SPONTANEOUSLY SELF-GENERATED VISUAL REPRESENTATIONS
HELP PROBABILITY PROBLEM SOLVERS?
The present study provides mixed evidence for the idea that external visual devices
are used by probability problem solvers because they are helpful (i.e., they aid in solving
the problem). For example, we found higher rates of solution success given use of
reorganization, outcome listings, and trees for the Combinations problems. The finding
regarding outcome listings makes sense intuitively because the essence of the
combinations problems involves determining the number of outcomes in the outcomes
space. Furthermore, use of trees is associated with a higher success rate (60%) for those
solving the Combinations problems via a sequential approach, and use of trees is a
relatively successful strategy (67% success rate) for the “true” Sequential problems. This
finding seems easily interpretable, because these sequential events problems have
structures that map directly onto tree diagrams. Specifically, the Sequential problems
used here described a sequence of trials or events, each of which had several possible
outcomes. Thus the process determining the outcome space can be described by a
branching set of possibilities. In the corresponding tree, each node of the tree graph
corresponds to one of the sequential events (e.g., one spin of the spinner), and the
branches that ensue from that node represent the several possible outcomes of that
uncertain event.
43
For the Permutations problems, the use of pictures led to a higher success rate. Our
explanation for this finding is based on the point we have already made, that these
particular permutation problems were atypical in that their semantic content (i.e., the realworld situation they described) describes a matching process between two sets of objects
(e.g., tutors and students). Typical permutation problems the students had seen in their
course consisted of problems in which a single set of objects is randomly ordered. Thus,
it is only through a relatively sophisticated symmetry argument (requiring what is
perhaps a rare or difficult insight) that the student was likely to see that one set of objects
could be arbitrarily ordered, hence ignored, reducing the problem to one about ordering a
single set of objects. For this reason we suspect that the increased solution success
associated with use of pictures for this problem type may indicate a facilitative effect of
pictures for problem restructuring. Such restructuring seems a necessary insight to deal
with this relatively novel type of problem.
Although problem solvers showed relatively frequent use of novel schematic
representations for the Permutations problems, these novel or invented types of external
graphical representations were apparently not always useful, because their use did not
lead to increased solution success.
Why are positive correlations between specific types of visual representations and
specific problem types relatively rare in our data? As Novick and Hmelo (1994) observe,
having an appropriate problem representation does not guarantee that the problem can be
solved, because computational or other issues may intrude, lowering correlations between
initial problem representations and solution success. Furthermore, even if graphics could
be helpful, prior research shows that students are not always successful in finding correct
representations for problems (Novick, 1990). Our data provide additional evidence that
this is true. Additionally, some evidence from our study suggests that choosing an
inappropriate representation might be harmful to a student’s chance of successfully
solving a problem. For example, for the Permutations problems solution success was
negatively associated with use of contingency tables and with use of reorganization of the
given information. The former finding can be explained because contingency tables are
not appropriate for representing problems involving the ordered selection of a single set
of objects. The latter finding can be explained by viewing the reorganization strategy as a
response commonly chosen when the problem solver is confused. Thus, the negative
association may indicate that when a student is stymied by a problem, rewriting the given
information might be seen as a general-purpose strategy, to be tried if the student is
merely casting about for any approach that might help.
We also found lower rates of success associated with use of outcome listings for the
Sequential problems. Here, we suspect that the choice of representation could be based on
a wrong understanding of the problem situation, or might just be an unfortunate (being
potentially misleading) choice. For the Sequential problems used here, the listings seem
to be appropriate, but they may cue (incorrect) approaches based on treating the outcome
space as consisting of equally-likely outcomes.
In addition to lower rates of solution success associated with particular external visual
representations, we also found a significant negative correlation (r = -.4, p < .05) between
solution success and the overall use of external visual representation, suggesting that our
participants were often using the external visual representations in futile solution
attempts. Looking more closely at the correlations, we found that reorganization of the
given information and outcome listings were marginally significantly correlated with
solution failure. These two types of representations are very general tools for problem
solving and participants may use these types of representations mainly when they are
having trouble solving the problems. This tendency could produce such correlation with
44
solution failure. A related possibility is that participants who are adept at solving
probability problems may not need to use or report the use of a visual representation;
whereas weaker problem solvers may choose very general external visual representations
in the absence of insights that might allow them to select a more specific representation,
leading to a lower rate of success overall for these lower-ability participants who needed
to resort to external representations.
Prior research by Hegarty and Kozhevnikov (1999) suggests that there are two main
types of external visual representations, schematic and pictorial. They found that the use
of schematic representations was positively correlated with solution success and the use
of pictorial representations was negatively correlated with success. In our data, solution
success was negatively (but non-significantly) correlated with use of all types of external
visuals, but the correlations were more negative with the use of pictures, reorganization,
and outcome listings than with contingency tables and with the schematic representations
(Venn diagrams, trees, and novel schematic representations), lending tentative support to
the importance of this distinction.
In summary, appropriate use of a correct external visual representation may generally
be helpful in problem solving, but this effect is difficult to measure in the present type of
study, in which the student only generates an external representation if he or she so
chooses. First, there is evidence (Lowrie & Kay, 2001; Table 7) that self-generated
external visuals may be tried more often for difficult or novel problems, which have a
lower solution rate in general. It should be easier to demonstrate facilitative effects of
external visuals in less naturalistic studies in which the visual representations are
provided to the student, or the student is explicitly asked to generate an appropriate
representation before attempting to compute the answer. That type of study has been
common in the literature on uses of visual representations in (non-mathematical) problem
solving. However, the present data showing which types of graphical representations are
spontaneously used for which types of problems may aid in designing such experimental
studies and educational interventions.
Finally, facilitative effects of using visual representations may not be easy to detect in
the present type of experiment because choosing the correct representation is a non-trivial
task, and may require a certain level of problem understanding to accomplish (Novick,
1990; 2001; Novick & Hmelo, 1994; Novick & Hurley, 2001). With novice problem
solvers, knowledge of why one representation is more appropriate than another may still
be incomplete, because they have not yet mastered the appropriate schemas. Riley,
Greeno and Heller (1983) found that failure to solve word problems might be caused
more often by a lack of appropriate schemas than by poor arithmetic skills. They
observed that problem solvers often carried out correct arithmetic procedures on incorrect
representations of the problems. The negative associations we found between certain
types of (presumably inappropriate) representations and solution success seem consistent
with their conclusions. Interestingly, De Bock, Verschaffel, Janssens, Van Dooren, and
Claes (2003) also found negative effects on solution success of asking students to
generate specified diagrams for geometry problems, showing that not all experimenterselected representations are useful as well (cf. Tversky, 2001; Mayer & Gallini, 1990;
Scaife & Rogers, 1996), or perhaps merely indicating that not all student-generated visual
representations are produced correctly, even when appropriately cued.
3.2. SOME FINAL ISSUES
One potential limitation of the present study is the question of how well the results
will generalize to other populations of students. Participants in the present study were
45
graduate students in social sciences and education, who were finishing or had recently
finished an introductory course in probability and statistics. However, the participants
were actually quite diverse in terms of mathematics background, including people who
had not taken any mathematics in college and those who had taken a number of
undergraduate or graduate mathematics and statistics courses. Thus, we believe that our
results would generalize to other populations, such as high school students who had
completed a similar probability course. But this question, and the question of how well
the results would generalize to a wider set of probability problems, should be addressed
by future research.
Another factor possibly affecting the generalizability of these results is that
participants were taught probability problem solving using external visual
representations, and associations between specific problem types and specific types of
external graphical devices may have been implicitly or explicitly taught. Thus, the results
of our study may just be a reflection of the instruction. However, the use of “novel”
graphical representations by some participants indicates that although the participants
were taught to use visual devices when solving probability problems, they do not always
use the specific representations they were taught to use in class. This finding suggests that
students believe that visual representations are useful, and try to use even representations
that they have not been explicitly taught.
Studying how students solve probability problems (or any type of mathematics word
problem) is a complex endeavor. One reason is that students can use any of several
solution methods or strategies for many problems. Even worse, an individual student may
switch approaches across similar problems, or even during the solution of a single
problem. As an example of how multiple solution strategies can complicate the research
process, we designed each of our probability problems with a particular formula or
problem-solving schema in mind. However, in producing atypical problems for a given
method, in at least one case (the atypical Combinations problem) we produced a problem
that could easily be solved by another method entirely (treating the problem as involving
sequential events), with a different appropriate external representation.
Thus, another limitation to the present study is that our manipulation of problem
“typicality” was not fully successful, due to the use of alternative solution strategies by
many participants. In a well-controlled experimental study with novice probability
problem solvers, this problem could be avoided by introducing only one solution method
or probability principle at a time. However, in a naturalistic study like the present one,
where participants have been taught an array of probability problem-solving techniques,
the problem of alternate strategy choice is difficult to avoid. Certainly such effects could
be minimized by more careful piloting of materials in future studies.
Another issue deserving of future study is to more closely investigate the temporal
process of probability problem solving. In the present study we have used a coding
scheme that separates uses of external visuals for problem understanding and
representation from the type of method used to compute the problem solution, but we still
do not have a clear picture of the temporal stages of probability problem solving. We plan
future studies that will use think-aloud protocols and structured interviews to try to
distinguish sequential stages of probability problem solving, and that will examine
specifically when and how external visual representations are used in the temporal
process of PPS. We also plan to investigate the coordination of external visual
representations with internal visualizations (cf. Scaife & Rogers, 1996). Results of these
studies may bring us to a more complete understanding of the role played by visual
representations and visualization skills in probability problem solving.
46
Results from the present study might be useful in improving instruction in the domain
of probability problem solving. The very general types of external representations
considered here (pictures, reorganization, and outcome listing) might be taught to
students as general methods that can help them restructure particularly difficult problems.
In contrast, the schematic representations studied here (Venn diagrams and outcome
trees) and contingency tables could be taught as applicable to particular problem types. In
line with the work of Novick and colleagues (e.g., Novick, 1990; Novick & Hmelo,
1994), abstract aspects or “features” of problems and of specific graphical representations
could be taught to students, and it could be emphasized that a given representation will
most likely be useful when these structural aspects of the problem and the visual device
match. To some extent such principles may already be employed by instructors of
probability courses, but future research should explore and better document the success of
such practices.
REFERENCES
Batanero, C., Godino, J., & Roa, R. (2004). Training teachers to teach probability.
Journal of Statistics Education, 12(1).
[Online: www.amstat.org/publications/jse/v12n1/batanero.html]
Chance, B., & Garfield, J. (2002). New approaches to gathering data on student learning
for research in statistics education. Statistics Education Research Journal, 1(1), 3841.
[Online: www.stat.auckland.ac.nz/~iase/serj/SERJ1(2).pdf]
De Bock, D., Verschaffel, L., Janssens, D., Van Dooren, W., & Claes, K. (2003). Do
realistic contexts and graphical representations always have a beneficial impact on
students’ performance? Negative evidence from a study on modeling non-linear
geometry problems. Learning and Instruction, 13(4), 441-463.
Garfield, J., & Ahlgren, A. (1988). Difficulties in learning basic concepts in probability
and statistics: Implications for research. Journal for Research in Mathematics
Education, 19(1), 44-63.
Gelman, A., & Nolan, D. (2002). Teaching statistics: A bag of tricks. Oxford, England:
Oxford University Press.
Gigerenzer, G. (1994). Why the distinction between single-event probabilities and
frequencies is important for psychology (and vice versa). In G. Wright & P. Ayton
(Eds.), Subjective Probability (pp. 129-161). Somerset, NJ: John Wiley & Sons.
Ginsburg, H. (1997). Entering the child's mind: The clinical interview in psychological
research and practice. New York: Cambridge University Press.
Hall, V., Bailey, J., & Tillman, C. (1997). Can student-generated illustrations be worth
ten thousand words? Journal of Educational Psychology, 89(4), 677-681.
Hampson, P., & Morris, P. (1990). Imagery and working memory. In P. J. Hampson, D.
E. Marks, & J. T. E. Richardson (Eds.), Imagery: Current developments (pp. 78-102).
New York: Routledge.
Hannafin, R. D., Burruss, J. D., & Little, C. (2001). Learning with dynamic geometry
programs: Perspectives of teachers and learners. The Journal of Educational
Research, 94(3), 132-144.
Hannafin, R. D., & Scott, B. (1998). Identifying critical learner traits in a dynamic
computer-based geometry program. The Journal of Educational Research, 92(1), 312.
Hegarty, M., & Just, M. (1993). Constructing mental models of machines from text and
diagrams. Journal of Memory and Language, 32(6), 717-742.
47
Hegarty, M., & Kozhevnikov, M. (1999). Types of visual-spatial representations and
mathematical problem solving. Journal of Educational Psychology, 91(4), 684-689.
Hollebrands, K. (2003). High school students’ understandings of geometric
transformations in the context of a technological environment. Journal of
Mathematical Behavior, 22(1), 55-72.
Kahneman, D., Slovic, P., & Tversky, A. (1982). Judgment under uncertainty: Heuristics
and biases. Cambridge; New York: Cambridge University Press.
Kaufmann, G. (1990). Imagery effects on problem solving. In P. J. Hampson, D. E.
Marks, & J. T. E. Richardson (Eds.), Imagery: Current developments (pp. 169-197).
New York: Routledge.
Keeler, C. M., & Steinhorst, R. K. (2001). A new approach to learning probability in the
first statistics course. Journal of Statistics Education, 9(3).
[Online: www.amstat.org/publications/jse/v9n3/keeler.html]
Kintsch, W., & Greeno, J.G. (1985). Understanding and solving word arithmetic
problems. Psychological Review, 92(1), 109-29.
Koedinger, K. R., & Anderson, R. (1997). Intelligent tutoring goes to school in the big
city. International Journal of Artificial Intelligence in Education, 8, 30-43.
Konold, C. (1989). Informal conceptions of probability. Cognition and Instruction, 6(1),
59-98.
Konold, C. (1995). Confessions of a coin flipper and would-be instructor. The American
Statistician, 49(2), 203-209.
Konold, C. (1996). Representing probabilities with pipe diagrams. The Mathematics
Teacher, 89(5), 378-384.
Konold, C., Pollatsek, A., Well, A. D., Lohmeier, J., & Lipson, A. (1993). Inconsistencies
in students’ reasoning about probability. Journal for Research in Mathematics
Education, 24, 392-414.
Kozhevnikov, M., Hegarty, M., & Mayer, R. (2002). Revising the visualizer-verbalizer
dimension: Evidence for two types of visualizers. Cognition and Instruction, 20(1),
47-77.
Lehrer, R., & Schauble, L. (1998). Reasoning about structure and function: Children’s
conceptions of gears. Journal of Research in Science Teaching, 35(1), 3-25.
Logie, R., & Baddeley, A. (1990) Imagery and working memory. In P. J. Hampson, D. E.
Marks, & J. T. E. Richardson (Eds.), Imagery: Current Developments (pp. 103-128).
New York: Routledge.
Lowrie, T., & Kay, R. (2001). Relationship between visual and nonvisual solution
methods and difficulty in elementary mathematics. The Journal of Educational
Research, 94(4), 248-255.
Mayer, R. (1989). Systemic thinking fostered by illustrations in scientific text. Journal of
Educational Psychology, 81(2), 240-246.
Mayer, R. (1992). Mathematical problem solving: Thinking as based on domain specific
knowledge. In R. E. Mayer (Ed.), Thinking, problem solving, cognition (pp. 455-489).
New York: W. H. Freeman & Co.
Mayer, R. (2001). Multimedia learning. Cambridge: The Cambridge University Press.
Mayer, R., & Anderson, R. (1991). Animations need narrations: An experimental test of a
dual-coding hypothesis. Journal of Educational Psychology, 83(4), 484-490.
Mayer, R., & Anderson, R. (1992). The instructive animation: Helping students build
connections between words and pictures in multimedia learning. Journal of
Educational Psychology, 84(4), 444-452.
Mayer, R., & Gallini, J. (1990). When is an illustration worth ten thousand words?
Journal of Educational Psychology, 82(4), 715-726.
48
Mayer, R., Mautone, P., & Prothero, W. (2002). Pictorial aids for learning by doing in a
multimedia geology simulation game. Journal of Educational Psychology, 91(4),
171-185.
Molitor, S., Ballstaedt, S. P., & Mandl, H. (1989). Problems in knowledge acquisition
from text and pictures. In H. Mandl & J. Levin (Eds.), Knowledge acquisition from
text and pictures (pp. 3-35). North-Holland: Elsevier Science Publishers.
Nemirovsky, R. (1994). On ways of symbolizing: The case of Laura and the velocity
sign. Journal of Mathematical Behavior, 13(4), 389-422.
Novick, L. (1990). Representational transfer in problem solving. Psychological Science,
1(2), 128-132.
Novick, L. (2001). Spatial diagrams: Key instruments in the toolbox for thought. In D.
Medin (Ed.), The psychology of learning and motivation: Advances in research and
theory, Volume 40 (pp. 279-325). San Diego, CA: Academic Press.
Novick, L., & Hmelo, C. (1994). Transferring symbolic representations across
nonisomorphic problems. Journal of Experimental Psychology: Learning, Memory,
and Cognition, 20(6), 1296–1321.
Novick, L.R., & Hurley, S. M. (2001). To matrix, network or hierarchy: That is the
question. Cognitive Psychology, 42(2), 158-216.
O’Connell, A. A. (1993). A classification of student errors in probability problemsolving. Unpublished doctoral dissertation, Teachers College, Columbia University,
New York.
O’Connell, A. A. (1999). Understanding the nature of errors in probability problem
solving. Educational Research and Evaluation, 5(1), 1-21.
O’Connell, A., & Corter, J. E. (1993, April). Student misconceptions in probability problemsolving. Paper presented at annual meeting of the American Educational Research
Association, Atlanta, GA.
Penner, D. E., Giles, N. D., Lehrer, R., & Schauble, L. (1996). Building functional
models: Designing an elbow. Journal of Research in Science Teaching, 34(2), 125143.
Pollatsek, A., Well, A., Konold, C., Hardiman, P., and Cobb, G. (1987). Understanding
conditional probabilities. Organizational Behavior and Human Decision Processes,
40(2), 255-269.
Presmeg, N. C. (1986). Visualization in high school mathematics. For the Learning of
Mathematics – An International Journal of Mathematics Education, 6(3), 42-46.
Qin, Y., & Simon, H. (1995). Imagery and mental models in problem solving. In B.
Chandrasekaran, J. Glasgow, & N. H. Narayanan (Eds.), Diagrammatic reasoning
(pp. 403-434). Menlo Park, CA: The AAAI Press.
Reusser, K. (1996). From cognitive modeling to the design of pedagogical tools. In S.
Vosniadou, E. De Corte, R. Glaser & H. Mandl (Eds.), International perspectives on
the design of technology–supported learning environments (pp. 81-103). Mahwah,
NJ: Erlbaum.
Riley, M., Greeno, J., & Heller, J. (1983). Development of children’s problem-solving
ability. In H. Ginsburg (Ed.), The development of mathematical thinking. New York:
Academic Press.
Russell, W. E. (2000). The use of visual devices in probability problem solving.
(Doctoral Dissertation, Columbia University, 2000). Dissertation Abstracts
International, 61, 1333.
Santos-Trigo, M. (1996). An exploration of strategies used by students to solve problems
with multiple ways of solution. Journal of Mathematical Behavior, 15(3), 263-284.
49
Scaife, M., & Rogers, Y. (1996). External cognition: How do graphical representations
work? International Journal of Human-Computer Studies, 45(2), 185-213.
Schwartz, D. L., & Black, J. B. (1996). Shuttling between depictive models and abstract
rules: Induction and fallback. Cognitive Science, 20(4), 457-497.
Schwartz, D. L., & Martin, T. (2004). Inventing to prepare for future learning: The
hidden efficiency of encouraging original student production in statistics instruction.
Cognition and Instruction, 22(2), 129-184.
Sedlmeier, P. (2000). How to improve statistical thinking: Choose the task representation
wisely and learn by doing. Instructional Science, 28(3), 227-262.
Sedlmeier, P., & Gigerenzer, G. (2001). Teaching Bayesian reasoning in less than two
hours. Journal of Experimental Psychology: General, 130(3), 380-400.
Tversky, B. (2001). Spatial schemas in depictions. In M. Gattis (Ed.), Spatial schemas
and abstract thought (pp. 79-112). Cambridge, MA: MIT Press.
Zahner, D. C., & Corter, J. E. (2002, April). Clinical interviewing to uncover the
cognitive processes of probability problem solvers. Paper presented at the 2002
Annual Meeting of the American Educational Research Association, New Orleans,
LA.
James E. Corter
Doris C. Zahner
Department of Human Development
Teachers College, Columbia University
525 West 120th Street
New York, NY 10027
50
APPENDIX: PROBABILITY PROBLEMS
The eight probability problems (test form A) classified by topic (problem type) and level
of typicality, with observed proportion correct for each item. Test form B
counterbalanced surface content of the text and level of typicality for each problem.
Topic
Typicality
Typical
Combinations
Atypical
Typical
Sequential
Atypical
Typical
Permutations
Atypical
Typical
Conditional
Probability
Atypical
Problem Text
There are 10 books on Mary’s bookcase. She randomly grabs 2
books to read on the bus. What is the probability that the 2 books
are Little Women and War & Peace? (Both these books are on her
bookshelf.)
There are 10 cookies in a cookie jar. Three of the cookies are
chocolate chip, seven are sugar. A child blindly picks 2 cookies
from the cookie jar. What is the probability that both cookies are
chocolate chip?
There are three balls in an urn. One is red, one is white, and one is
blue. Jane randomly draws a ball from the urn, then replaces it,
three times in all. What is the probability that she draws a red ball
on all three turns?
Three spinners are constructed. The first spinner has 2 equal areas
(colored red and blue), the second has three equal areas (red, blue,
and white), and the third again has two equal areas (red and white).
All three spinners are spun and the result of each spin is recorded.
What is the probability of getting ‘red’ on all three spins?
An apartment building has four parking spaces in front (call them
A, B, C, and D). There are four apartments in the building (#1, #2,
#3, and #4), and each apartment has a single occupant with a single
car. Every evening, all four occupants come home and park in a
randomly chosen space. What is the probability that this evening
they park so that the occupant of Apt #1 is in space A, the occupant
of #2 is in space B, the occupant of #3 in space C, and the occupant
of #4 in space D?
There are four math students (Ed, Fred, Mary, Pia) waiting to be
randomly matched with four math tutors (#1, #2, #3, and #4). Each
tutor works one-on-one with a student. What is the probability that
Ed will be matched with tutor #1, and Fred will be matched with
tutor #2?
Joe applies for a state-subsidized mixed-income housing project
being built in his neighborhood. If he is classified as a low-income
applicant, he has a 70% chance of getting an apartment. Applicants
not classified as low-income have only a 10% chance of getting an
apartment. Joe believes that on the basis of the records he is
submitting that he has a 40% chance of being classified as low
income. What is the probability that he gets an apartment?
Assume that in the city of Metropolis, if a criminal defendant in
fact committed the crime, he has a 70% chance of being found
guilty by the jury. A defendant who is in fact innocent has a 10%
chance of being found guilty by the jury. Assume that 40% of
defendants who are tried in Metropolis in fact committed the crime.
We meet a Metropolis defendant in prison. What is the probability
that he is fact committed the crime, given that we know he was
found guilty by the jury?
Proportion
of correct
responses
0.115
0.731
0.462
0.423
0.462
0.731
0.500
0.077
51
STUDENTS’ EXPECTATIONS OF INTRODUCTORY
STATISTICS INSTRUCTORS5
MARK A. EARLEY
Bowling Green State University
[email protected]
ABSTRACT
The purpose of this phenomenological study was to talk to students about their
experiences taking introductory statistics. The author met with eleven students
individually for four interviews throughout the semester, followed by a memberchecking focus group during the last week of classes. One of the most salient themes
to emerge was the students’ reliance on their instructor for feedback about
performance, directions on taking notes, and the creation of a classroom environment
that motivated them to study. As part of the phenomenological tradition, the author
presents his own reflections based on these students’ comments. Conclusions include
the encouragement of instructors to be more mindful of students’ reactions to course
content, and suggestions for developing a more learner-centered learning
environment.
Keywords: Statistics education research; Teaching statistics; Statistics classroom;
Learning environment; Phenomenology
1. INTRODUCTION
The current statistics reform paradigm stresses instructors teaching and students
learning statistical concepts over mechanics (Chance & Garfield, 2002). One goal of this
model is to help students develop relational or structural knowledge in addition to
declarative and mechanical knowledge (Earley, 2001; Schau & Mattern, 1997). With
relational knowledge, students go beyond just “knowing that” standard deviation is the
square root of variance as well as “knowing how” to compute a standard deviation. How
students connect standard deviation with concepts they already know demonstrates their
relational or structural knowledge. The shift away from mechanics and toward
understanding is one attempt to decrease students’ anxiety levels, under the assumption
that reducing the mathematical content and rote memorization of definitions and formulae
reduces students’ worries about course performance (Onwuegbuzie, DaRos, & Ryan,
1997). The most frequently cited implication is the need to develop class activities and
assessment tools that are more concept-based and less calculation-based (Gal & Garfield,
1997). Carpenter and Lehrer (1999) discuss understanding in mathematics courses as a
“mental activity” as well as something “emerging or developing rather than presuming
that someone either does or does not understand a given topic, idea, or process”
(emphasis added, p. 20). Garfield (1995) warns instructors, “no [teaching] method is
perfect and will work with all students” (p. 32) and “teachers often overestimate how well
their students understand basic concepts” (p. 31). These statements imply researchers and
classroom teachers would be wise to invest time listening to individual students to get a
Statistics Education Research Journal, 6(1), 51-66, http://www.stat.auckland.ac.nz/serj
© International Association for Statistical Education (IASE/ISI), May, 2007
52
sense of how well their own teaching methods help students develop conceptual
understandings of introductory statistics concepts.
What we are not seeing in the literature are discussions of how students’
understandings are impacted by the classroom environment. What is also not heavily
discussed in the literature is how students respond to more conceptually-based classrooms
(Batanero, Garfield, Ottaviani, & Truran, 2000). Of course, this does not include
references to the well known statistics anxiety phenomenon. Few researchers have
explored any of our students’ experiences in statistics courses – there is a more consistent
focus on achievement outcomes (Becker, 1996). Gordon stated that “as in other fields, a
major challenge currently facing researchers in statistics education is to improve our
understanding of learning” (1995, para. 6) and “in order to teach statistics effectively, we
must first understand the learners” (2000, p. 16). Gordon’s work focuses on activity
theory and the social, historical, and cultural aspects of classroom environments to help
understand statistics learners (1995, 2000, 2004). Through this perspective, she
concluded in a 1995 study that statistics classrooms must be supportive, instructors must
provide guidance to students, and teaching must “build on the personal experience of the
learner” (para. 47). Nine years later, Gordon (2004) indicated there is still a need for
statistics education researchers and statistics educators to understand our learners. To this
end, her interview work done for a 2004 report led to five categories of meanings
students attach to statistics: statistics as having no meaning, statistics as processes,
statistics as mastery, statistics as a tool, and statistics as critical thinking.
These meaning categories parallel those described by Reid and Petocz (2002), in
which students described statistics as having (a) a focus on techniques, (b) a focus on
using data, and (c) a focus on meaning. In a later report they redefined these conceptions
as (a) doing, (b) collecting, (c) applying, (d) linking, (e) expanding, and (f) changing
(Petocz & Reid, 2003). In both cases, the authors describe these conceptions as a
continuum from limited to expansive. More limited conceptions, according to Gordon
(2004), have the potential to limit students’ study strategies and motivation for learning
statistics. All of these studies suggest understanding students’ descriptions of what
“statistics” means is an important precursor for improving students’ learning and
ultimately their experiences in our courses. One question to address is how students
create these meanings: Do they come in to class with them, do specific aspects of the
classroom environment change initial meanings or create new ones (as Gordon suggests),
and do these meanings change over the course of the instruction? In other words, what
kind of impact does the learning environment created by instructors have on students’
meaning-making?
Petocz and Reid (2002, 2003) have addressed some of these questions by connecting
students’ conceptions of statistics with expectations students have of statistics instructors.
Again, the authors describe these conceptions as most limited to most expansive:
Conception 1: Providing materials, motivation, and structure
Conception 2: Explaining material and helping with student work
Conception 3: Linking statistical concepts and guiding learning
Conception 4: Anticipating student learning needs
Conception 5: Being a catalyst for ‘open-mindedness’
The important addition in this work is the blending of students’ conceptions of statistics
and their expectations of instructors. As an example, Petocz and Reid (2004) describe one
mix as students’ conceptions of “doing” and “collecting” along with students’
expectations of their instructors to provide materials, motivation, and structure
(Conception 1). This would be the most limited blend of both sets. Although her
interviews did not directly ask about students’ expectations of their instructors, Gordon
53
(2004) did interview two instructors and also discusses their impression of the
interactions between how students attach meaning to statistics and what goes on in the
classroom.
One important aspect of these studies is their reliance on interviews and open-ended
survey responses at one point in time. Any discussion of how classroom environments
and instructors impact students’ conceptions should incorporate a more longitudinal data
collection strategy. For example, Murtonen and Lehtinen (2003) conducted a study with
Finnish students in which they asked participants to write in “learning diaries” throughout
the entire term. Students were to record difficult concepts they encountered as well as
what they thought led to the difficulties. Students’ reasons for difficulties fell into five
categories: (a) superficial teaching, (b) difficulty linking the theory of quantitative
methods with the doing, (c) unfamiliarity and difficulty with content (e.g., too many
concepts to absorb), (d) lack of connections across bits of information, and (e) negative
attitudes toward quantitative methods. Similar to the typologies of Gordon (2004) and
Petocz and Reid (2003), Murtonen and Lehtinen’s categories begin to blend students’
understandings of quantitative material with how it is taught. Interestingly, they conclude
“the experience of difficulty did not occur because of the major subject or any specific
teacher, but because of some more general reason” (p. 182), indicating again that
understanding students’ backgrounds could be an important precursor to developing more
effective classroom environments. What is also important from this work is its
longitudinal design. Ultimately, the authors chose to combine responses from throughout
the term, however, so we still have no presentation of how individual students’
experiences change (if at all) as the course progresses. Petocz and Reid (2003) posit that
students “would develop their approach to learning, and maybe change their conceptions
of teaching from time to time” (p. 41), suggesting that rather than exploring class marks
on assessments, an additional approach to data collection that also explores individual
students’ experiences would help statistics educators understand further what students
take away from the course.
This expectation, along with Carpenter and Lehrer’s (1999) assertion that
mathematical learning is developmental, suggests that expanding this body of work
through more longitudinal explorations of students’ experiences throughout their course
would be useful. Talking to students more frequently allows researchers to gather more
information about students’ experiences, such as: What do students do to prepare for
class? What do students do while in class? What goes through students’ minds during
class? What aspects of class time help or hinder students’ understandings of the material?
What do students do when they leave class? For these reasons, the general purpose of the
current phenomenological study was to begin this exploration with one very broad
guiding question: “How do students describe their statistics course at different points
during the term in which they are taking the course?”
2. METHODS
2.1. THE PHENOMENOLOGICAL TRADITION
The use of phenomenology allows us to understand student experiences differently
than through surveys or other qualitatively-oriented traditions. Rather than “averaging” or
grouping students’ experiences together, phenomenology seeks to present “the
qualitatively different ways in which a phenomenon is experienced” (Pietersen, 2002,
para. 14). Van Manen (1990) stresses phenomenology as “human scientific study” with
an emphasis on “explicat[ing] the meanings as we live them in our everyday experience,
54
our lifeworld” (p. 11). Phenomenological research incorporates the researcher as
someone learning about a phenomenon by “borrow[ing] other people’s experiences and
their reflections on their experiences” (Van Manen, 1990, p. 62). In phenomenology, it is
the actual doing and living that constitutes knowledge – “a kind of knowing that can only
be obtained through active engagement” (Ladkin, 2005, p. 116). Ultimately,
phenomenological research attempts to present the unique lived experience of others so
that the reader also develops a new understanding of the experience.
An essential first step in conducting phenomenological research is for the researcher
to first understand his or her own experience with the phenomenon under investigation
(commonly referred to as bracketing or époche) (Ladkin, 2005; Laverty, 2003). As one
reads different descriptions of phenomenology, this process is part of what distinguishes
the various approaches (Kerry & Armour, 2000; Laverty, 2003). Rather than completely
“suspend” our own understandings, as Edmund Husserl’s description of époche would
have us do (Creswell, 1998), hermeneutic phenomenology, most commonly attributed to
Hans-Georg Gadamer, (Kerry & Armour, 2000; Ladkin, 2005), incorporates these
understandings as essential to the interpretive process to follow. In other words, because
the researcher will interpret participants’ experiences while developing a description of
the phenomenon, an understanding of the researcher’s own experiences is necessary so
the reader has an idea of the researcher’s interpretive lens. Van Manen (1990) describes
this as the “intersubjective” nature of phenomenological research: “the human science
researcher needs the other (for example, the reader) in order to develop a dialogic relation
with the phenomenon” (p. 11).
This also means the researcher’s own understanding develops during the research
process – in his description of the phenomenological data collection process, van Manen
refers to this as the “reflective data” (2002; see also Laverty, 2003) researchers collect
alongside the “empirical data” (i.e., interview data) from our participants. van Manen
(1990) indicates social science researchers “borrow” participants’ experiences in order to
develop a complete description of the experience. In this sense, the researcher is
developing the description as his or her own developing understanding of the
phenomenon under study. Kerry and Armour (2000) indicate “these personal histories
lead to a unique perception of different experiences and that this personal history cannot
be bracketed out; it is fundamental for interpretation” (p. 6). For example, Finney (2000)
presents her own experience with success in statistics courses prior to detailing six other
students’ success experiences. The reader then understands how her interpretation and
summary of the experience of success is shaped by her own initial perceptions. Mayhew
(2004) also presents his experiences with spirituality prior to describing the meanings
eight undergraduate students attribute to spirituality “to give the reader a sense of any
prejudice or orientation that may have shaped [his] interpretation or approach to the
study” (p. 656).
2.2. PARTICIPANT SELECTION
In phenomenological research, the researcher bases participant selection on finding
and recruiting individuals who have experienced the phenomenon of interest (rather than
selecting key informants in ethnography, or selecting individuals based on their life
experiences for a life history) (Creswell, 1998; Laverty, 2003; Moustakas, 1994; van
Manen, 1990). The quantity and depth of the data to be collected determines how many
participants the researcher can recruit. In phenomenological inquiry, the primary data
source is the interview, and depending on the phenomenon of interest, there may be more
than one interview per participant. The researcher must have each interview transcribed
55
for analysis, creating a large set of data to read through and analyze, so the number of
participants is generally kept to a minimum (Creswell, 1998). There are no specific
guidelines on what constitutes this “minimum” – “Boyd (2001) regards two to 10
participants … as sufficient to reach saturation and Creswell (1998, pp. 65 & 113)
recommends … 10 people” (as cited in Groenewald, 2004, p. 11). Because I was
interested in interviewing participants four times during the semester and interviewing all
participants within a week of each other, a number around ten seemed manageable and
was my goal.
Six departments across four colleges at our institution offered 40 sections of
introductory statistics courses during the spring 2003 term (this includes multiple sections
of the ten different courses). Students in this study came from multiple sections of four of
these courses. All of these courses are traditional in scope, covering introductory
descriptive and inferential statistical analyses in one term. The “Psychology 270” course
is the required introductory course for psychology majors. The “Statistics 211” is the first
in a required two-course sequence for all majors in the College of Business
Administration. The “Math 115” course is offered to students not majoring in
mathematics; many undergraduates take this course to fulfill the mathematics
requirement for the university if a different course is not specified by their College.
Mathematics majors take a different introductory statistics course, so the “Math 115”
course is not calculus-based. The “EDFI 641” course is a required course for master’s
level students in the College of Education and Human Development. All of these
departments offer more advanced courses, but only students in the College of Business
Administration are required to take a second course.
My goal for recruiting participants was to get a mix of participants from across all of
these courses. Eligible students included those over the age of 18 and enrolled in any
introductory statistics course during the spring 2003 term. In January 2003 I posted flyers
next to the doors of each classroom in which one of the 40 introductory statistics sections
met. This method of recruitment does introduce typical volunteer bias, but the advantage
is that if students are interested in talking about their experiences (whether good or bad),
they will volunteer (a necessary participant characteristic in phenomenology). The flyer
asked students to contact me via phone or e-mail if interested in taking part in my study,
and I set up times to meet with the twelve who did so by the end of the third week of
classes in order to have a first interview for each participant completed during the third or
fourth week of the term. At our first interview, I informed participants of a $120 incentive
to those who completed the study by returning for interviews throughout the semester.
Although the issue of incentives is still controversial (Adler & Adler, 2002), I do believe
in giving participants something in appreciation of their time (Weiss, 1994). I do not feel
this had an impact on the quality of my participants’ contributions because each
participant was engaged during our conversations and each had much to say about their
experiences in their statistics class.
To give the reader a general description of the participants, Table 1 lists demographic
information for the final set of eleven students I interviewed (one student dropped out of
the study after the first interview). I did achieve the variety I initially wanted: My group
includes participants from freshman to senior with one graduate student, a grade point
average (GPA) range of 2.00 – 4.00 (i.e., students fell across the general academic
performance spectrum from “average” to “excellent” students), and students from courses
in four of the six departments offering statistics courses in the spring (no students from
the Sociology or Criminal Justice courses contacted me).
56
Table 1. Demographic summary
Participant a
Gender
Course Taken b
GPA c Class
Aileen
Female
Psychology 270
3.60
Senior
Alastair
Male
Statistics 211
2.60
Sophomore
Brice
Male
Psychology 270
4.00
Senior
Brigit
Female
Math 115
3.77
Sophomore
Cecily
Female
EDFI 641
4.00
1st Year Graduate
Cedric
Male
Statistics 211
2.92
Freshman
Dillan
Male
Math 115
2.90
Sophomore
Eleyn
Female
Math 115
4.00
Freshman
Evan
Male
Statistics 211
2.50
Sophomore
Griffin
Male
Math 115
2.00
Freshman
Ian
Male
Statistics 211
3.66
Sophomore
a
All names used in this report are pseudonyms.
b
All courses are introductory-level statistics courses designed for non-statistics majors.
c
GPA = Grade Point Average, calculated as an average of the students’ grades for all of
their coursework to date after converting the letter grades to “grade points”: A = 4, B = 3,
C = 2, D = 1, F = 0.
2.3. DATA COLLECTION
Phenomenological data collection proceeds in two parallel processes: (a) empirical
data collection, typically in the form of in-depth interviews with participants, and (b)
reflective data collection, in which the researcher reflects on his or her interactions with
the empirical data (van Manen, 2002). The power of phenomenological inquiry is this
interaction, where the researcher first presents his or her own experiences with the
phenomenon of interest, and then reflects on how his or her experience and
understanding of the phenomenon changes and develops while listening to the
participants’ stories (Moustakas, 1994).
I met with each student four times for one-on-one interviews in my office (the only
exception was Ian, who started the study late and only met with me three times). We
spaced interviews approximately 3 to 4 weeks apart depending on the student’s schedule.
I had an initial list of questions I wanted to discuss, based in part on my interests and in
part on themes emerging from the previous interviews. Each interview began, however,
with the general question, “So how is class going for you?” From there, we discussed
whatever the student brought up as important at that time. Interview times ranged from 15
to 45 minutes, entirely dependent on how much the student had to say. I audio taped the
interviews and took fieldnotes during our discussion. I then transcribed audiotapes, taking
additional fieldnotes as I listened to our discussion again.
To increase the accuracy and representativeness of the themes I saw emerge
(Maxwell, 2005), I met with nine of my participants in two member-checking focus
groups during the last week of classes prior to final exams. “Member-checking” is
described as a process of going back to the participants to have them review the
researcher’s interpretations of and conclusions drawn from the data. Through this
process, participants comment on whether the researcher “got it right” (so to speak)
(Creswell, 1998; Glesne, 1999; Maxwell, 2005). Due to the scheduling so close to the end
of the term and final exams, I gave students two options: an afternoon time with lunch,
selected by two participants, or an evening time with dinner, selected by seven. Two of
my participants could not make these focus groups due to scheduling conflicts. I
57
presented students with a summary of the major themes on slides, one at a time, and
solicited group feedback on whether these were actually important to the students. To get
this feedback, I asked students to think about what they would tell a group of statistics
instructors about their experiences taking statistics. Students changed the wording on
some (e.g., the course is “easier when instructors …” not “easy when instructors …,” as I
had first worded one of the themes), but otherwise agreed these were important ideas to
them and ideas they would share with statistics instructors as important.
2.4. DATA ANALYSIS
To come to an understanding of how my participants experienced their statistics
courses, I analyzed and reflected on the transcribed data throughout the study. After each
interview, I took time to make notes on ideas the student seemed to stress or return to
often. I also noted student comments that, for some reason, stood out for me. After each
round of interviews, I explored major patterns across all 11 participants and developed
questions for the next round of interviews. After the study ended, I also developed timeordered displays (Miles & Huberman, 1994) of students’ main comments from the
beginning to the end of the term to explore changes in their descriptions of class. Because
I developed interview protocols based in part on the ongoing analysis of each round of
interviews (i.e., what did I find important or interesting that I wanted to explore again in
the next interview?), I include here a list of the major “topics” for each round of
interviews. I introduced these additional topics into the interview after the opening
question about how class is going.
Interview 1: Discuss a general description of the class, what happens during class,
what student does in between class meetings, other issues student brings up.
Interview 2: How does student take notes? What does student do with notes during
and outside of class time? Why does student take notes? Does student only take notes
in class, or does student read the textbook and take notes as well?
Interview 3: What would happen if there were no identifiable instructor? What would
the student miss?
Interview 4: Summary of student’s thoughts about course, additional comments, and
comparison of their statistics course with another course they are taking (students
chose one of their other courses and we discussed similarities and differences
between that course and their statistics course).
All participants, after the first interview, came prepared to answer the “How is class
going?” question – they clearly had topics they wanted to share with me. After this
discussion, I introduced questions addressing each of the ideas listed above.
3. RESULTS
3.1. BRIEF REFLECTIONS ON MY OWN EXPERIENCE
My experiences taking statistics courses began with a required course for my (then)
economics degree. I clearly remember a lot of frustration and confusion as the instructor
filled blackboard after blackboard with Greek symbols and mathematical formulae. We
used a computer software package for homework problems, without the benefit of a
user’s manual or any training on how to use the software (back then, the software
required writing programs to run analyses as opposed to today’s “point-and-click”
platforms). My solution to this confusion and lack of understanding was to give in, earn a
D-, and retake the course the following term.
58
This new instructor was so radically different from my first – primarily because he
actually made class enjoyable. I wanted to be in class, and worked hard between class
meetings in order to participate fully each week. We used computer software, but now he
guided us through the programming during class time before sending us out on our own. I
recall measuring 100 leaves from trees around campus for one of our early homework
exercises – data we would analyze and discuss throughout the course. After being in class
with this instructor (and concurrently taking an advanced economic theory course in
which I was struggling), I switched my major to mathematical statistics at the beginning
of my sophomore year and finally felt “at home” with my coursework.
Although I have had many (many) statistics courses since my start in 1989, I turn
now to some brief reflections on my observations as an instructor of introductory
statistics in education courses (note that none of the students in my classes participated in
this study). Many of these informal observations are what led me to develop an interest in
finding out what exactly my students experience when they take my course. Have I
avoided creating a classroom environment similar to that very first environment in which
I failed? Have I created a motivating and interesting classroom environment similar to my
second experience (along with many other inspiring instructors and classes along the
way)?
No matter what I do or say, students consistently come in to my course with the
typical fear and anxiety. It seems at times that I am an actor on a stage, begging my
audience to laugh and enjoy themselves. After careful attention to the textbook I use, the
way I structure class meetings, the assessment system, and my own concern for students’
success, some students still carry that fear and anxiety throughout the term. I hear from
students that the textbook is confusing and impossible to read. I hear from students how
they can follow along in class, but seemingly all is lost when they leave the classroom;
they have no idea what they are looking at when they study at home. Student evaluations
of the course consistently rate “relevance to my profession/degree program” the lowest of
all evaluation items. Students come in the first day and announce this is their last course –
all they want is their grade of “C” so they can graduate and move on.
The power of phenomenological inquiry is the “connections” researchers and
participants can make through their shared experiences. Moving into this particular study,
all of my own experiences, attitudes, beliefs, and values related to introductory statistics
form the context in which I talk to participants, listen to their words, and try to make
sense of what we discuss. Rather than considering this a major bias to the study, however,
this context allows the reader to understand this narrative as my understanding of these
students’ experiences (as is the case with any naturalistic inquiry process). I have paid
close attention to verbatim transcripts, and my reflections at the end of this narrative point
to the “surprises” and discussion topics that made me reflect deeper on my role as
instructor. The member-checking focus groups conducted with my participants helped
validate the themes I saw emerge, to which I now turn.
3.2. THE VOICES OF MY PARTICIPANTS
Even with the variety of students I talked to, as well as the diversity of their
classroom environments, three common themes emerged across all of our discussions: (a)
student behaviors and characteristics, (b) instructor behaviors and characteristics, and
(c) resources students use for studying. This report focuses on the second of these: the
students’ perceptions of the instructor and his or her behaviors. During the first two sets
of discussions, this idea came across in a variety of my participants’ responses to
59
questions, regardless of whether I had directly asked about the instructor. Based on this
preliminary analysis, I asked students the following question during our third discussion:
Imagine you had to take this course entirely online. There is no identified instructor
now – you read texts and examples online, complete assignments and tests online,
and receive scores via e-mail. What would that be like?
My intent here was to focus on what students would miss about their instructors, having
discussed them so much during our first two interviews.
Responses to this question, along with questions from earlier discussions, fell into
three major sub-themes:
(1) The instructor’s use of class time (including pace and teaching style),
(2) The instructor’s role in the note-taking process, and
(3) The instructor’s assessment strategies (including homework and exams).
I present next my interpretations of each of these themes along with representative quotes
from my participants’ responses.
The instructor’s use of class time One of my early interests in conducting this study was
to find out what it is like just to be in a statistics class: What are students thinking, what
kinds of behaviors do they feel are important during and between class meetings, and
most importantly, why? For most students, the response to what they do in class was “sit
there and take notes.” What thoughts did they have during class? For some, they were
random because, as Brigit said,
Brigit:
It’s hard to stay focused on it because all she does is write definitions from the book
onto the board and then just reads them and explains them a little bit. Very rarely
we go through examples of things but mostly it’s just definitions.
For Brigit and others, instructors organized class time around the traditional “lecture”
approach with very little interaction among students. There were exceptions, in particular
Griffin’s instructor, who did no formal instruction but rather assigned group work from
the text.
Griffin:
Int:
Griffin:
Int:
Griffin:
He gives us problems to do from the book and then sits up front and watches us.
He is there to answer questions. If we don’t have any questions, we finish our
exercises and leave.
Do you like that approach?
Well he seems pretty unmotivated to teach the class and not interested at all.
What do you think of that?
There is no reason for me to do any work outside of class. Why should I care if he
doesn’t?
When asked about the online course, only Griffin was excited at the prospect of
“taking the class while [he stays] in bed.” The remaining ten participants each expressed
concern that this type of setting, most notably the absence of a live instructor, would
never work for statistics. Cecily was most vocal (her eyes nearly popped out of her head
when I asked her that question) in wanting to avoid this type of course.
Cecily:
Int:
Cecily:
Wow … no way of asking questions? There’s gotta be a way of asking questions …
yeah I may be able to pass the class but I wouldn’t know it. I wouldn’t be able to use it.
So you need the instructor there to ask questions to.
Absolutely. I have a math background, I’m pretty good at math. But this is not just a
60
math class … there’s a lot of concepts here that just reading the book [won’t work] …
Cedric also expressed concern over the “completely online” approach.
Cedric:
Int:
Cedric:
Int:
Cedric:
Int:
Cedric:
I don’t think that would work at all.
Why not?
Because, for example, I’ll hit a stumbling block where I don’t understand it, but then I
talk to a peer and that sheds a different light on it. It makes it so much easier. Without
that interaction between students it makes it more difficult, especially because the
instructor can shed another different light on it.
So having the different viewpoints are helpful.
Yes. The textbook is good at telling you what to do, but not as good at telling you why
this happens.
So is that something the instructor has been helpful for? Presenting the whys?
Yes.
Overall, students who had instructors giving and explaining examples in class were
more comfortable with their experience – students like Cedric who were not getting these
examples wished their instructors would do more of “the whys” instead of just “the
whats.”
I clearly recall my own desire to skip class meetings the first time I took statistics
contrasted with my excitement to be in class the second time – student comments about
how the instructor guided class time resonated strongly with me. As an instructor, I am
keenly aware of students who do not appear engaged or “with me” – most likely a result
of my own early experiences. My participants integrated the first four of Petocz and
Reid’s (2003) conceptions into their expectations of how class time is used: They expect
instructors to be organized, present more than just an outline of the text, explain concepts
beyond what the text offers, and be there to address student questions and provide
alternative explanations when needed. This is different from Petocz and Reid’s analysis,
in which students could be described by one or two of the conceptions – all 11 of my
participants discussed all four of these conceptions throughout the term.
The instructor’s role in the note-taking process A natural progression from the first subtheme on instructor behaviors and characteristics is how these behaviors influence
students’ note-taking processes. Students consistently reported relying predominately on
their notes for studying course material. This is also how they each spent their time in
class (again with the exception of Griffin noted above) – listening to the instructor and
writing down whatever he or she wrote on the blackboard or overhead. I asked students
how they decided what to write down, and most indicated it was “whatever [or
everything] the instructor writes down.” Only Brice and Cedric mentioned they also jot
down additional notes to themselves about key formulae or concepts to remember.
Evan, Ian, and Alastair each had instructors who provided handouts with full or
partial notes already on them. For these students, this meant they just followed along and
maybe filled in an example or two – Evan called this “being spoon fed” the information.
Evan:
Int:
Evan:
He passes out sheets and he goes over the stuff in class … I remember a lot of it
from last semester, so I’ll go through and do the stuff that I know. And I’ll just sit
around and wait for a question that I didn’t know and fill in the answer as he
explains it.
So then when you’re in class you’re going through and doing on your own the stuff
that you remember and then waiting?
If I hear something that I didn’t write down or don’t remember from last semester,
61
Int:
Evan:
I’ll write that down.
Do you take any other notes besides those on the sheet he hands out?
No he usually has a lot of space on the side to put extra stuff there. He spoon feeds
us most of it.
With or without these handouts, students rely very heavily on their notes for studying
outside of the classroom. Cedric further indicated that he has experienced the “false
safety net” of feeling as though he is following along in class, not taking any notes as a
result, and then being lost when he got home to study for the test (having no notes to refer
to). Here again, students all mentioned either being happy with an instructor who walked
through examples in class (so these could go into their notes) or being unhappy with an
instructor who did not do so.
Three of my participants had sold their books back by the first interview (within a
month of class starting), another two sold them back later in the term, and five of the
remaining six students used their books for assigned homework problems only. Only
Cecily read the textbook in between classes, though she indicated this was difficult to do.
Brigit and Dillan struggled with this the most because their instructor “lectured” directly
from the book.
Brigit:
Int
Brigit
This class is different [from my other classes] because instead of having the professor’s
point of view on the notes, it’s just point blank – I look at my book and I looked at my
notes and it’s word for word.
So what do you do then?
I have random thoughts during class. It’s hard to focus when it’s just like being dictated
to. Plus I know if I miss anything it’ll be in the book.
Dillan indicated it was almost “stupid” to go to class because he can sit at home and read
the book – what he and other students want from instructors is explanations and
“different viewpoints,” not dictation. The textbook was not enough for them to learn the
concepts, and the instructor in this case was not serving as an extra resource, which
bothered them. For these students, and students who cannot or do not read the text even
though they keep it, there is an even greater importance placed on course notes.
The instructor’s assessment strategies There was an overwhelming consensus from all of
my participants that regular homework and frequent assessment are necessary in order for
students to know whether or not they are learning anything. As Brice put it:
Brice:
Each person in the class was left on their own to make sense of what was being
presented … maybe it felt like a struggle because if we’d had some more problems …
we’d have some confirmation of whether or not we understood the material. Maybe it
was a struggle because we just didn’t know for sure if we understood it or not – we just
really didn’t have any confirmation of that.
Brice and others were concerned that with no practice, no regular homework, and no
assignments to work on, preparing for exams was more difficult and led to an increase in
anxiety prior to exams. Cecily mentioned that “a computer can give you the answer, but it
doesn’t give you how it got the answer,” and so for her the additional need was for the
instructor to go over the assessments when returning them to students.
Students also indicated that “what the instructor told us would be on the test” was
their study guide, and any deviation from this on tests and quizzes was frustrating for
students. Although I did not look at individual assessments unless the student asked me a
question about them, what they perceived to be poor test development clearly impacted
62
their class experiences. A major challenge was tests based on “applications” when
instructors did not give students application models in class. Even though many students
called for and recognized the need to understand what it is they were doing, they wanted
instructors to guide them in this process. Cedric’s first exam experience was particularly
enlightening:
Cedric:
Int:
Cedric:
I wasn’t happy with my first test.
What happened?
Right answer, wrong work, so a lot of points were taken off. The one thing is, she
doesn’t accept just the answer. Even if you do some work, unless you do the right work
it will be marked wrong. I can understand that if we’re majors, but we’re general
business majors. Just getting the right answer is enough. Recognizing wrong answers is
good, but not exactly knowing step-by-step.
This conflict between Cedric’s realization that understanding is important (from earlier
conversations) but not being comfortable when tested on this understanding is important
to me. These data highlight the need for instructors to match more closely their
instruction and assessment procedures so students do not get these conflicting messages.
These data also suggest that instructors need to make explicit connections between
instruction and assessment, as instructors’ intentions are not always clear.
4. DISCUSSION
All of these sub-themes are intimately related to each other, as well as to the two
other larger themes of student characteristics and behaviors and use of resources.
Students in this study spoke frequently about the role the instructor plays in their class
experiences, but for me the surprising aspect of all of our conversations was the extent to
which students relied on their instructors. What we do in the classroom has far-reaching
impacts beyond how students feel during class. When on their own, students need
resources that will help them make sense of class material (Petocz and Reid’s (2003)
“Conception 3,” p. 46). Those participants most comfortable with their experiences
indicated they could take their notes home, complete homework assignments by referring
to their notes as necessary, and received consistent feedback from their instructor as to
whether they were indeed “getting it.” Oathout’s (1995) participants echoed “strongly
favor[ing] frequent tests” (p. 50) and a lack of “mapping from lectures and assignments to
test content and format [as] equally problematic” (p. 48).
Many students, however, did not do much work outside of class because graded
assignments were either non-existent or infrequent. This meant their statistics course was
the lowest priority class when studying during the week, and they rarely did anything
with their statistics material outside of class time. This changed around exam time, when
there was a need for notes to be clear and complete, and for these notes (typically
verbatim from instructor notes in class) to be reflective of what would be assessed on the
exams (similar to Petocz and Reid’s (2003) “Conception 1” on teaching statistics, p. 44).
The exams became the focal point for these students – they took notes so they would
have something to study so they could do well on the exams. As Garfield (1995) states,
“students learn to value what they know will be assessed” (p. 32) – Petocz and Reid
(2003) discuss this as “doing” statistics (p. 42). Students in the current study usually
waited until an exam was coming up to actually go back over their notes, so any
questions they had could not be addressed and they were sometimes still confused and
unsure going into the exam. How the instructor conducts class, how the students record
classroom events and discussions in their notes, how the instructor assesses students in
63
between exams, and how the instructor develops exams all become a single system to
which instructors need to pay close attention.
Most importantly, students never talked about “getting it” in terms of understanding
what they were doing. Although Cedric indicated that knowing why they are doing an
analysis makes doing the analysis easier, students did not ever indicate they understood
course material. Rather, they were happy if they could get problems correct and earn all
of the possible points. Most of the instructors these students spoke of appear to encourage
the “doing” aspect of statistics – solving problems is a large part of how these students
spent their time both in and out of class. Students knew their tests would involve solving
problems (based on what their instructors told them about the tests), so this was how
students studied. If we want students to hold the more expansive conceptions of statistics
that include “statistics as critical thinking” (Gordon, 2004) and a “focus on meaning”
(Petocz & Reid, 2003), we need instructors to hold and espouse these conceptions as
well. These data suggest that future researchers explore these same conceptions with
instructors to discern their own conceptions as well as how they create learning
environments that support students developing the more expansive conceptions.
5. REFLECTING ON MY OWN CLASSROOM
As I reflected on the discussions I engaged in with these students, my thoughts turned
to my own teaching philosophy and classroom environment. Listening to them talk about
their instructors and class experiences, I often asked myself, “Do I do that in my
classroom?” Reflective questions about my own practices that developed because of these
discussions, questions I believe worthy of further exploration both in my own practice
and in the larger statistics education community, include:
(1) How much do I actually stress “doing” statistics over “understanding”?
(2) What are my expectations of students during and outside of class meetings? Am I
clear in expressing these expectations to my students?
(3) What role does the textbook play in my course?
(4) How can I best assist students in taking complete and accurate notes?
(5) What is the connection between each class activity and/or discussion and my
assessment of student understandings?
I firmly believe as instructors we need to spend more time reflecting on our own
teaching philosophies and practices, and for me, this reflection requires student feedback.
We can focus our reflections on each aspect of the classroom, understanding which
piece(s) are working for our students and which are not. Statistics education researchers
can also consider these questions, and begin to focus research more on these “emerging
understandings.” Through these explorations, we can continue to learn how students are
responding to our desire for them to walk away from the course with some knowledge of
what and why various concepts were covered, rather than walking away with the feeling
that all they did was “plug and chug.”
6. CONCLUSION
Although it appears intuitive that students rely on their instructors in any course,
there are some implications that may not be as intuitive. Based on my own observations
as an instructor, as well as comments provided by the students in this study, a primary
motivation for students in introductory statistics (those who are not statistics majors) to
do anything (take notes, work problems, study outside of class, etc.) is how it will impact
their final grade (Garfield, 1995). Students further mentioned that they appreciate
64
feedback on graded assignments as a way to make sure they know what the instructor
wants them to know. Here we see a direct connection between how we assess students,
the opportunities we provide for assessment (e.g. homework, in-class activities, quizzes,
tests, projects), and why we even assess them (Chance & Garfield, 2002).
Assessment systems in any course gauge whether or not the students are meeting
course objectives. As we work to develop introductory statistics courses oriented more
toward understanding and less toward mechanics, our assessment systems must change as
well. As we change our assessments, it is important to remember that our classroom
environment must also change to encourage students’ development of understanding over
their mastery of mechanics (Treagust, Jacobowitz, Gallagher, & Parker, 2001).
Demetrulias (1988) argues that an “opportunity to understand statistics from an integrated
and flexible viewpoint must go along with a classroom environment that rewards such
exploration” (emphasis added, p. 169). As students in this study report, what happens in
class becomes their main resource for any work they do outside of class (Petocz & Reid,
2003). Spending time in class (and with assessments) on “how to understand” as much as
“what to understand” becomes critical, and instructors must continue to develop and
make use of delivery techniques, classroom examples, and homework activities that focus
on this “how to understand” piece. I also encourage sharing these reflections and
experiences in the literature, so that others may benefit from what we learn in our own
classrooms. Results from this study suggest the need for researchers to encourage more
personal accounts of actual experiences of instructors teaching, and students taking,
statistics, as opposed to relying only on the more quantitative outcomes currently
presented (Becker, 1996). Instructors and their students are “in the trenches,”
experiencing, reflecting on, and ultimately determining whether any particular classroom
environment is or is not successful in increasing students’ understandings (Batanero,
Garfield, Ottaviani, & Truran, 2000).
ACKNOWLEDGEMENTS
I presented a preliminary version of this paper at the 2003 Annual Meeting of the
Mid-Western Educational Research Association. My gratitude goes to Beverly Dretzke
for her thoughtful feedback as discussant of that paper session and to Tony Onwuegbuzie
for his insightful comments on a draft of this manuscript. I am also indebted to the editors
and reviewers who carefully read and provided insightful feedback on earlier drafts of
this manuscript. Of course, I owe much to my eleven participants for meeting with me
throughout the semester and sharing their thoughts and experiences.
REFERENCES
Adler, P. A., & Adler, P. (2002). The reluctant respondent. In J. F. Gubrium & J. A.
Holstein (Eds.), Handbook of interview research: Context & method (pp. 515-536).
Thousand Oaks, CA: Sage.
Batanero, C., Garfield, J. B., Ottaviani, M. G., & Truran, J. (2000, May). Research in
statistical education: Some priority questions. Statistical Education Research
Newsletter, 1(2), 2-6.
[Online: www.stat.auckland.ac.nz/~iase/serj/newsmay00.pdf]
Becker, B. J. (1996). A look at the literature (and other resources) on teaching statistics.
Journal of Educational and Behavioral Statistics, 21(1), 71-90.
65
Boyd, C. O. (2001). Phenomenology the method. In P. L. Munhall (Ed.), Nursing
research: A qualitative perspective (3rd ed., pp. 93-122). Sudbury, MA: Jones and
Bartlett.
Carpenter, T. P., & Lehrer, R. (1999). Teaching and learning mathematics with
understanding. In E. Fennema & T. A. Romberg (Eds.), Mathematics classrooms that
promote understanding (pp. 19-32). Mahwah, NJ: Lawrence Erlbaum Associates.
Chance, B. L., & Garfield, J. B. (2002). New approaches to gathering data on student
learning for research in statistics education. Statistics Education Research Journal,
1(2), 38-44.
[Online: www.stat.auckland.ac.nz/~iase/serj/SERJ1(2).pdf]
Creswell, J. W. (1998). Qualitative inquiry and research design: Choosing among five
traditions. Thousand Oaks, CA: Sage.
Demetrulias, D. M. (1988). (Creatively) teaching the meanings of statistics. The Clearing
House, 62, 168-170.
Earley, M. A. (2001). The development of knowledge structures in introductory statistics.
Unpublished doctoral dissertation, University of Toledo, Toledo, OH.
Finney, S. J. (2000, April). The meaning of success for students in statistical methods
courses: A phenomenological study. Paper presented at the Annual Meeting of the
American Educational Research Association, New Orleans, LA. (Eric Document
Reproduction Service No. ED441026)
Gal, I., & Garfield, J. B. (Eds.). (1997). The assessment challenge in statistics education.
Amsterdam: IOS Press and the International Statistical Institute..
Garfield, J. (1995). How students learn statistics. International Statistical Review, 63(1),
25-34.
Glesne, C. (1999). Becoming qualitative researchers: An introduction (2nd ed.). New
York: Longman.
Gordon, S. (1995). A theoretical approach to understanding learners of statistics. Journal
of Statistics Education,3(3).
[Online: www.amstat.org/publications /jse/v3n3/gordon.html]
Gordon, S. (2000, May). Bibliography on activity theory and related research in
education. Statistical Education Research Newsletter, 1(2), 16-22.
[Online: http://www.stat.auckland.ac.nz/~iase/serj/newsmay00.pdf]
Gordon, S. (2004). Understanding students’ experiences of statistics in a service course.
Statistics Education Research Journal, 3(1), 40 - 59.
[Online: www.stat.auckland.ac.nz/~/iase/serj/SERJ3(1)_gordon.pdf]
Groenewald, T. (2004). A phenomenological research design illustrated. International
Journal of Qualitative Methods, 3(1). Article 4.
[Online: www.ualberta.ca/~iiqm/backissues/3_1/pdf/groenewald.pdf]
Kerry, D. S., & Armour, K. M. (2000). Sport sciences and the promise of
phenomenology: Philosophy, method, and insight. QUEST, 52, 1-17.
Ladkin, D. (2005). ‘The enigma of subjectivity’: How might phenomenology help action
researchers negotiate the relationship between ‘self’, ‘other’ and ‘truth’? Action
Research, 3(1), 108-126.
Laverty, S. M. (2003). Hermeneutic phenomenology and phenomenology: A comparison
of historical and methodological considerations. International Journal of Qualitative
Methods, 2(3). Article 3.
[Online: www.ualberta.ca/~iiqm/backissues/2_3final/pdf/laverty.pdf]
Maxwell, J. A. (2005). Qualitative Research Design: An Interactive Approach (2nd ed.).
Thousand Oaks, CA: Sage.
66
Mayhew, M. J. (2004). Exploring the essence of spirituality: A phenomenological study
of eight students with eight different worldviews. NASPA Journal, 41(3), 647-674.
Miles, M. B., & Huberman, M. A. (1994). Qualitative data analysis: An expanded
sourcebook (2nd ed.). Thousand Oaks, CA: Sage.
Moustakas, C. (1994). Phenomenological research methods. Thousand Oaks, CA: Sage.
Murtonen, M., & Lehtinen, E. (2003). Difficulties experienced by education and
sociology students in quantitative methodology courses. Studies in Higher Education,
28(2), 171-185.
Oathout, M. J. (1995, April). College students’ theory of learning introductory statistics:
Phase one. Paper presented at the Annual Meeting of the American Educational
Research Association, San Francisco, CA. (Eric Document Reproduction Service No.
ED391841)
Onwuegbuzie, A. J., DaRos, D., & Ryan, J. (1997). The components of statistics anxiety:
A phenomenological study. Focus on Learning Problems in Mathematics, 19, 11-35.
Petocz, P., & Reid, A. (2002). How students experience learning statistics and teaching.
In B. Phillips (Ed.), Proceedings of the Sixth International Conference on Teaching
Statistics. Cape Town, South Africa [CD-ROM]. Voorburg, The Netherlands:
International Statistical Institute.
[Online: www.stat.auckland.ac.nz/~iase/publications/1/6b4_peto.pdf]
Petocz, P., & Reid, A. (2003). Relationships between students’ experience of learning
statistics and teaching statistics. Statistics Education Research Journal, 2(1), 39-55.
[Online: www.stat.auckland.ac.nz/~iase/serj/SERJ2(1).pdf]
Pietersen, C. (2002). Research as a learning experience: A phenomenological explication.
The Qualitative Report, 7(2), 1-13.
[Online: www.nova.edu/ssss/QR/QR7-2/pietersen.html]
Reid, A., & Petocz, P. (2002). Students’ conceptions of statistics: A phenomenographic
study. Journal of Statistics Education, 10(2).
[Online: www.amstat.org/publications /jse/v10n2/reid.html]
Schau, C., & Mattern, N. (1997). Assessing students’ connected understanding of
statistical relationships. In I. Gal & J. B. Garfield (Eds.), The assessment challenge in
statistics education (pp. 91-104). Amsterdam: IOS Press and the International
Statistical Institute.
Treagust, D. F., Jacobowitz, R., Gallagher, J. L., & Parker, J. (2001). Using assessment as
a guide in teaching for understanding: A case study of a middle school science class
learning about sound. Science Education, 85(2), 137-157.
van Manen, M. (1990). Researching lived experience: Human science for an action
sensitive pedagogy. New York: SUNY Press.
van Manen, M. (2002). Phenomenology Online.
[Online: www.phenomenologyonline.com]
Weiss, R. S. (1994). Learning from strangers: The art and method of qualitative
interview studies. New York: The Free Press.
MARK A. EARLEY
Division of Educational Foundations & Inquiry
Bowling Green State University
554 Education Building
Bowling Green, Ohio, USA
43403-0246
67
PAST CONFERENCE6
USCOTS 2007
UNITED STATES CONFERENCE ON TEACHING STATISTICS
Columbus OH, USA, May 17-19, 2007
The second biennial United States Conference
on Teaching Statistics (USCOTS 07) was held on
May 17-19, 2007 at the Ohio State University in
Columbus, Ohio, hosted by CAUSE, the Consortium
for the Advancement of Undergraduate Statistics
Education. The target audience for USCOTS is teachers of undergraduate and AP
statistics, from any discipline or type of institution. Teachers from two-year colleges are
particularly encouraged to attend.
The theme for USCOTS 2007 was Taking Statistics Teaching to the Next Level.
“Next level” has many interpretations, such as developing a second course, gaining more
confidence in teaching statistics, moving students beyond statistical literacy to statistical
thinking, and using the latest technology to improve teaching and learning. USCOTS is a
“working conference” with many opportunities for hands-on activities, demonstrations,
networking, idea sharing, and receiving the latest information on research and best
practices in teaching statistics. Leaders in statistics education gave plenary talks,
including Jessica Utts, Paul Velleman, Dick DeVeaux, Allan Rossman, and Mike
Shaughnessy.
For more information, visit the USCOTS web page: http://www.causeweb.org/uscots/
Statistics Education Research Journal, 6(1), 67-75, http://www.stat.auckland.ac.nz/serj
© International Association for Statistical Education (IASE/ISI), May, 2007
68
FORTHCOMING IASE CONFERENCES
SRTL-5
THE FIFTH INTERNATIONAL RESEARCH FORUM ON STATISTICAL
REASONING, THINKING, AND LITERACY
Coventry, UK, August 11 - 17, 2007
Reasoning about Statistical Inference:
Innovative Ways of Connecting Chance and Data
The
Forum’s
focus will be on
informal ideas of
inference rather than on formal methods of estimation and tests of significance. This topic
is emerging from the presentations and discussions at SRTL-3 and 4 and is a topic of
current interest to many researchers as well as teachers of statistics. As new courses and
curricula are developed, a greater role for informal types of statistical inference is
anticipated, introduced early, revisited often, and developed through use of simulation
and technological tools. Papers will address reasoning about statistical inference at all
levels of education including the professional development of elementary and secondary
teachers.
TOPICS
Submitted research papers address questions such as the following:
1. What are the simplest forms of statistical inference that students can understand?
2. How does reasoning about statistical inference develop from the simplest forms
(informal) to the more complex ones (formal)?
3. How can instructional tasks and technological tools be used to promote the
understanding of statistical inference?
4. What are sequences of activities that can help student develop a conceptual
understanding of statistical inference?
5. What types of misconceptions are found in students’ reasoning about statistical
inference?
6. What types of foundational knowledge and reasoning are needed for students to
understand and reason about statistical inference?
7. How do students develop an understanding of the language used in describing
statistical inference (e.g., significance, confidence)?
8. How does an understanding of statistical inference connect and effect
understanding of other statistical concepts?
9. What are useful items and questions to use to assess understanding of statistical
inference?
LOCAL SRTL-5 ORGANIZERS
Janet Ainley,
[email protected]
Dave Pratt,
[email protected]
For more information visit the SRTL-5 website: http://srtl.stat.auckland.ac.nz/
69
IASE SATELLITE CONFERENCE ON
ASSESSING STUDENT LEARNING IN STATISTICS
Guimaraes, Portugal, August 19-21, 2007
The meeting will be held on 19-21 August 2007 in Guimarães, Portugal, immediately
prior to ISI 56 in Lisbon. The fascinating historic city of Guimarães is about 50 km from
Portugal’s second largest city, Oporto. This Satellite will involve papers on many aspects
of assessing student learning in statistics. Over 40 papers will be presented along with a
number of posters and discussions of examination questions. Proceedings will be
available on CD and free at the publication page of IASE.
Non-participants must register by 31 May 2007.
CONFERENCE COMMITTEE
Brian Phillips (Australia) (Joint Chair and Joint Chief Editor)
[email protected]
Beth Chance (USA) (Joint Chair)
[email protected]
Allan Rossman (USA)
[email protected]
Ginger Rowell (USA)
[email protected]
Gilberte Schuyten (Belgium)
[email protected]
Larry Weldon (Canada) (Joint Chief Editor)
[email protected]
Local Organiser: Bruno C. de Sousa (Portugal)
[email protected]
For more information visit the website:
http://www.stat.auckland.ac.nz/~iase/conferences.php?show=iasesat07
ISI-56
THE 2007 SESSION OF THE INTERNATIONAL STATISTICAL INSTITUTE
Lisboa, Portugal, August 22 – 29, 2007
The 56th Session of the International Statistical
Institute (ISI) will be held in Lisboa, Portugal. As it does
at each major ISI conference, IASE will be organizing
about 10 statistics education sessions for ISI-56. Please
check the website at http://www.isi2007.com.pt/ for more
information, and contact the session organizers below if
you would like to offer to speak in one of the sessions.
IASE SPONSORED IPMS (ORGANIZERS, PRESENTERS, DISCUSSANTS)
IPM37 Research on Reasoning about Distribution
Organizer: Joan Garfield, USA
Presenters: Rolf Biehler, Germany; Jane Watson, Australia; Chris Reading,
Australia
Discussants: Roxy Peck and Beth Chance, USA
IPM38 How Modern Technologies Have Changed the Curriculum in Introductory
Courses
Organizers: Lucette Carter, France; Catherine Pardoux, France
Presenters: Cecily Peters, Malaysia; Brigitte Chaput, France; Mathilde
Mougeot, France
Discussants: Carmen Capilla, Spain; Robert Gould, USA
70
IPM39 Preparing Teachers of Statistics
Organizer: Allan Rossman, USA
Presenters: Carmen Batanero, Spain; Penelope Bidgood, UK; Verônica
Yumi Kataoka, Brazil; Madhuri Mulekar, USA
Discussant: Jerry Moreno, USA
IPM40 Research on the Use of Simulation in Teaching Statistics and Probability
Organizer: Rolf Biehler
Presenters: Nicolas Christou, Ivo D. Dinov and Juana Sanchez, USA;
Joachim Engel, Germany; Andrew Zieffler and Joan B. Garfield, USA
Discussant: Andrej Blejec, Slovenia
IPM41 Optimizing Internet-Based Resources for Teaching Statistics
Organizers: Roxy Peck, USA; Ginger Holmes Rowell, USA
Presenters: Mary Townsend, Canada; Iddo Gal and Dani Ben-Zvi, Israel;
Dennis Pearl, USA
Discussant: Irena Ograjensek, Slovenia
IPM42 Observational Studies, Confounding, and Multivariate Thinking
Organizer: Milo Schield, USA
Presenters: Donald Rubin, USA; Nancy Wermuth, Sweden and David Cox,
UK; James Nicholson, Jim Ridgway, and Sean McCusker, UK; John
Harraway, New Zealand
IPM43 Teaching of Official Statistics
Organizer: Sharleen Forbes, New Zealand
IPM44 Teaching of Survey Statistics
Organizer: Steve Heeringa, USA
Presenters: Don Royce, Canada; James J. Brown, UK; Marie-Christine
Ponsonnet, France; Giulio Ghellini, Italy
Discussant: Graham Kalton, USA
IPM45 Studying Variability Through Sports Phenomena
Organizer: Brian Phillips, Australia
Presenters: Tim Swartz, Canada; Stephen Clarke, Australia; Phil Everson,
USA; Kaznori Yamaguch, Michiko Watanabe, and Fumitake Sakaori, Japan
Discussant: Larry Weldon, Canada
IPM46 Use of Symbolic Computing Systems in Teaching Statistics
Organizer: Zaven Karian, USA
IASE ORGANIZING COMMITTEE
Allan J. Rossman (USA)
[email protected]
Gilberte Schuyten (Belgium)
[email protected]
Chris Wild (New Zealand)
[email protected]
For more information visit the ISI 56 website at http://www.isi2007.com.pt/ or
contact members of IASE OC.
71
JOINT ICMI /IASE STUDY
STATISTICS EDUCATION IN SCHOOL MATHEMATICS:
CHALLENGES FOR TEACHING AND TEACHER EDUCATION
Monterrey, Mexico, June 30 to July 4, 2008
The International Commission on Mathematical
Instruction (ICMI, http://www.mathunion.org/ICMI/) and
the International Association for Statistical Education
(IASE, http://www.stat.auckland.ac.nz/~iase/) are pleased
to announce the Joint ICMI/IASE Study Statistics
Education in School Mathematics: Challenges for Teaching
and Teacher Education. The conference is co-sponsored by
the American Statistical Association and endorsed by the
Interamerican Statistical Institute, Mexican Statistical Association and the International
Statistical Literacy Project.
Following the tradition of ICMI Studies, this Study will comprise two parts: the Joint
Study Conference and the production of the Joint Study book. The Joint Study
Conference will be merged with the IASE 2008 Round Table Conference.
The Joint Study Conference (ICMI Study and IASE Round Table Conference) will
take place at the Instituto Tecnológico y de Estudios Superiores, Monterrey, Mexico
(http://www.mty.itesm.mx/), from June 30 to July 4, 2008. Participation in the
Conference is only by invitation, based on a submitted contribution and a refereeing
process. Accepted papers will be presented in the Conference and will appear in the
Proceedings that will be published by ICMI and IASE as a CD-ROM and on the Internet.
The second part of the Joint Study – the Joint Study book – will be produced after the
conference and will be published in the ICMI Study Series. Participation in the Joint
Study Conference does not automatically assure participation in the book, because a
second selection and rewriting of selected papers will be made after the conference.
Proposed papers for contributions to the Joint Study Conference should be submitted
by e-mail no later than October 1, 2007, to the IPC Study Chair (Carmen Batanero,
[email protected]). Papers should be relevant to the Joint Study focus and research
questions, as described in the Discussion Document (which is available at the Joint Study
Website (http://www.ugr.es/~icmi/iase_study/). Guidelines for preparing and submitting
the paper are also available in the Discussion Document. Please address questions to
Carmen Batanero,
[email protected] or Joan Garfield,
[email protected] .
INTERNATIONAL PROGRAMME COMMITTEE
Carmen Batanero (Spain)
Bernard Hodgson (Canada, ICMI representative)
Allan Rossman (USA, IASE representative)
Armando Albert (México, ITSM representative)
Dani Ben-Zvi (Israel)
Gail Burrill (USA)
Doreen Connor (UK)
Joachim Engel (Germany)
Joan Garfield (USA)
Jun Li (China)
Maria Gabriella Ottaviani (Italy)
Maxine Pfannkuch (New Zealand)
Mokaeane Victor Polaki (Lesotho)
72
Chris Reading (Australia)
LOCAL ORGANISING COMMITTEE
Blanca Ruiz (Chair)
Ernesto Sánchez
Tomás Sánchez
Armando Albert
More information is available from Carmen Batanero,
[email protected] or from
http://www.ugr.es/~icmi/iase_study/
ICOTS-8
DATA AND CONTEXT IN STATISTICS EDUCATION:
TOWARDS AN EVIDENCE-BASED SOCIETY
Ljubljana, Slovenia, 11-16 July 2010
CONFERENCE THEME
The realization that data is preferable to anecdote or
intuition as a basis for robust decision making is spreading
through many professions and sections of society. More and
more, people want to see “the evidence”. Statistical
methodology and modelling are increasingly pervading the
research fabrics of all fields that advance by employing
empirical enquiry. And because the root purpose of statistics
is to extract insight and meaning about real contexts using
data, statistics educators are increasingly realizing that this cannot be modelled by
teachers without the use of rich, real contexts. It is important that data and contexts
pervade statistical learning and teaching, to help students understand the nature and value
of the statistical sciences, and to facilitate their learning. Successful learning processes
involve data and contexts that are meaningful to students. These can be relevant to
everyday life or to disciplines as varied as psychology, biology, business, sociology,
engineering, the health sciences and statistics itself. But many questions remain about the
myriad ways in which we can exploit context to achieve our educational goals. We also
must look hard at how well we use the data and contexts that should be guiding our own
educational practices.
EVIDENCE-BASED PRACTICE IN OTHER DISCIPLINES: SOME EXAMPLES
Statisticians are often essential contributors in research teams in many disciplines and
examples drawn from these contexts can enrich and facilitate the teaching of statistics.
Interaction between statistics educators, statisticians and researchers in a relevant
specialization can contribute significantly to the rich, real contextual and data resources
that are of such value in both motivating and assisting statistical learning. Trends in
medicine and other health sciences are governed by data, and evidence-based medicine is
taught now in all medical schools. Data from the biological sciences provide information
for resolving problems on environmental and ecological issues. The six-sigma revolution
uses statistical quality control methods to monitor and improve industrial and engineering
processes resulting in evidence-based decision making in industry. National statistics
offices and international agencies contribute to evidence-based decision making in
government and on public policy by collecting, collating, analysing and presenting data
to populations at large and to governments in particular.
73
EVIDENCE-BASED PRACTICE IN STATISTICS EDUCATION
Evidence-based practice should also be employed in statistics education itself. How
do we use context when teaching about variability, probability, inference and modelling?
How do we interpret data from surveys, questionnaires or interviews and how are these
related to the research hypotheses? To what extent are conclusions valid and reliable? Are
we dealing with and explaining risk appropriately? Only with the answers to these
intriguing questions will we be able to make informed decisions as we strive to reach an
evidence-based society. Education ideas are shared on the web, through international and
national projects, programmes, workshops and conferences in statistics education where
diverse innovations are shared. The impact and relevance of new ideas are assessed and
often adopted by others in their own teaching. Reports on the successes of recent statistics
education programmes in South Africa and Latin America reflect the impact of the two
ICOTS conferences and provide helpful ideas for other countries.
THE INTERNATIONAL PROGRAMME COMMITTEE EXECUTIVE
IPC Chair: John Harraway
Programme Chair: Roxy Peck
Information Manager: John Shanks
Scientific Secretary: Helen MacGillivray
Editor Proceedings: Alan McLean
LOCAL ORGANISING COMMITTEE
LOC Chair: Andrej Blejec
For more information visit the ICOTS-8 website: http://ICOTS8.org
74
OTHER FORTHCOMING CONFERENCES
2007 JOINT STATISTICAL MEETINGS
Salt Lake City UT, USA, July 29 - August 2, 2007
The 2007 Joint Statistical Meetings will be held July 29 - August 2, 2007 at the Salt
Palace Convention Center located at 100 South West Temple, Salt Lake City, Utah
84101.
JSM (the Joint Statistical Meetings) is the largest gathering of statisticians held in
North America. It is held jointly with the American Statistical Association, the
International Biometric Society (ENAR and WNAR), the Institute of Mathematical
Statistics, and the Statistical Society of Canada. Attended by over 5000 people, activities
at the meeting include oral presentations, panel sessions, poster presentations, continuing
education courses, an exhibit hall (with state-of-the-art statistical products and
opportunities), career placement service, society and section business meetings,
committee meetings, social activities, and networking opportunities. Salt Lake City is the
host city for JSM 2007 and offers a wide range of possibilities for sharing time with
friends and colleagues. For information, contact
[email protected]
Website: http://www.amstat.org/meetings/jsm/2007/
JOINT SOCR (STATISTICS ONLINE COMPUTATIONAL RESOURCE)
CAUSEWAY CONTINUING EDUCATION WORKSHOP 2007
UCLA, Los Angeles CA, USA, 6-8 August 2007
The 2007 joint SOCR/CAUSEway continuing education
workshop aims at demonstrating the functionality, utilization
and assessment of the current UCLA, SOCR and CAUSEweb
resources. This workshop will be of most value to AP teachers
and college instructors of probability and statistics classes who
have interests in exploring novel IT-based approaches for
enhancing statistics education. The workshop will provide an
interactive forum for the exchange of ideas and recommendations for strategies to
integrate computers, modern pedagogical approaches, the Internet and new student
assessment techniques.
For further information:
http://wiki.stat.ucla.edu/socr/index.php/SOCR_Events_Aug2007/
75
9TH INTERNATIONAL CONFERENCE OF THE MATHEMATICS
EDUCATION INTO THE 21ST CENTURY PROJECT
MATHEMATICS EDUCATION IN A GLOBAL COMMUNITY
Charlotte NC, USA, September 7 - 13, 2007
The Mathematics Education into the 21st Century Project was
founded in 1986 and is dedicated to the planning, writing and
disseminating of innovative ideas and materials in Mathematics and
Statistics Education. The next conference is planned for September
7 - 13, 2007 in Charlotte, North Carolina. The chairman of the Local Organising
Committee is Dr. David K. Pugalee, of the University of North Carolina Charlotte. The
title of the conference is “Mathematics Education in a Global Community.” Papers are
invited on all innovative aspects of mathematics education. Our conferences are
renowned for their friendly and productive working atmosphere. They are attended by
innovative teachers and mathematics educators from all over the world, 25 countries were
represented at our last conference for example.
For more information:
Alan Rogerson,
[email protected]
Website: http://math.unipa.it/~grim/21project.htm