Education Longitudinal Study of 2002 (ELS:2002/12) Third Follow-Up Field Test Report
Education Longitudinal Study of 2002 (ELS:2002/12) Third Follow-Up Field Test Report
Education Longitudinal Study of 2002 (ELS:2002/12) Third Follow-Up Field Test Report
The Working Paper Series was initiated to promote the sharing of the valuable work experience
and knowledge reflected in these preliminary reports. These reports are viewed as works in
progress, and have not undergone a rigorous review for consistency with NCES Statistical
Standards prior to inclusion in the Working Paper Series.
May 2012
Steven J. Ingels
Daniel J. Pratt
Donna M. Jewell
Tiffany Mattox
Ben Dalton
Jeffrey Rosen
Erich Lauff
Jason Hill
RTI International
Elise Christopher
Project Officer
National Center for Education Statistics
The National Center for Education Statistics (NCES) is the primary federal entity for collecting, analyzing, and
reporting data related to education in the United States and other nations. It fulfills a congressional mandate to
collect, collate, analyze, and report full and complete statistics on the condition of education in the United States;
conduct and publish reports and specialized analyses of the meaning and significance of such statistics; assist state
and local education agencies in improving their statistical systems; and review and report on education activities in
foreign countries.
NCES activities are designed to address high-priority education data needs; provide consistent, reliable, complete,
and accurate indicators of education status and trends; and report timely, useful, and high-quality data to the U.S.
Department of Education, the Congress, the states, and other education policymakers, practitioners, data users, and
the general public.
We strive to make our products available in a variety of formats and in language that is appropriate to a variety of
audiences. You, as our customer, are the best judge of our success in communicating information effectively. If you
have any comments or suggestions about this or any other NCES product or report, we would like to hear from you.
Please direct your comments to
National Center for Education Statistics, IES, U.S. Department of Education
1990 K Street NW
Washington, DC 20006-5651
May 2012
Content Contact
Elise Christopher
(202) 502–7899
[email protected]
ii
Foreword
In addition to official NCES publications, NCES staff and individuals commissioned by
NCES produce preliminary research reports that include analyses of survey results and
presentations of technical, methodological, and statistical evaluation issues.
The Working Paper Series was initiated to promote the sharing of the valuable work
experience and knowledge reflected in these preliminary reports. These reports are viewed as
works in progress and have not undergone a rigorous review for consistency with NCES
Statistical Standards prior to inclusion in the Working Paper Series. Copies of working papers
can be downloaded as PDF files from the NCES Electronic Catalog
(http://nces.ed.gov/pubsearch).
iii
This page intentionally left blank.
iv
Acknowledgments
We thank all the individuals who participated in any or all four rounds of the Education
Longitudinal Study of 2002 (ELS:2002), including the thousands of students, parents, and school
personnel who generously gave their time and approval to provide the data that the study reports.
The authors of this report also thank the many individuals who assisted in the planning of
the ELS:2002 third follow-up. We are particularly indebted to the ELS:2002 third follow-up
Technical Review Panel, which met to review plans for the study, helped refine them, and
provided important suggestions to help guide development of the instrumentation. The following
individuals serve as the nonfederal members of the ELS:2002 third follow-up Technical Review
Panel: Sara Goldrick-Rab, Robert Gonyea, Donald Heller, Robert Lent, Amaury Nora, Randall
Olsen, Aaron Pallas, Kent Phillippe, Michael Shanahan, and Marvin Titus.
v
This page intentionally left blank.
vi
Table of Contents
Foreword ....................................................................................................................................... iii
Acknowledgments ..........................................................................................................................v
Table of Contents ........................................................................................................................ vii
List of Tables ................................................................................................................................ ix
List of Figures ............................................................................................................................... xi
Chapter 1. Introduction ................................................................................................................1
1.1 Historical Background: NCES Secondary Longitudinal Studies Program .......................1
1.1.1 National Longitudinal Study of the High School Class of 1972 ..........................2
1.1.2 High School and Beyond ......................................................................................4
1.1.3 National Education Longitudinal Study of 1988 ..................................................5
1.1.4 High School Longitudinal Study of 2009 .............................................................5
1.2 Education Longitudinal Study of 2002 .............................................................................6
1.2.1 ELS:2002 Research and Policy Issues ..................................................................6
1.2.2 Overview of Data Collections...............................................................................7
1.2.3 ELS:2002 Study Design........................................................................................8
Base Year (2002) ..................................................................................................9
First Follow-up (2004) ........................................................................................10
Second Follow-up (2006) ...................................................................................11
Third Follow-up Field Test (2011) and Main Study (2012) ...............................12
Chapter 2. Field Test Survey Design and Preparation .............................................................15
2.1 ELS:2002 Third Follow-up Data Collection Instrumentation ........................................15
2.1.1 Instrument Development Goals and Constraints ................................................15
2.1.2 Instrument Development Process .......................................................................16
2.2 Pre-Data Collection Activities ........................................................................................17
2.2.1 Training of Data Collectors ................................................................................17
2.2.2 Locating and Tracing Activities .........................................................................18
Chapter 3. Data Collection Procedures and Results.................................................................21
3.1 Web/CATI Data Collection Procedures .........................................................................21
3.1.1 Study Website and Help Desk ............................................................................21
3.1.2 Interviewing ........................................................................................................22
3.2 Interview Data Collection Outcomes ..............................................................................25
3.3 Results of the Propensity Modeling Experiment ............................................................33
Chapter 4. Field Test Questionnaire Timing and Data Quality ..............................................43
4.1 Questionnaire Timing and Item Analyses ......................................................................43
4.1.1 Questionnaire Timing .........................................................................................43
vii
Table of Contents
Appendixes
A: ELS:2002 Technical Review Panel Summary ....................................................................... A-1
B: ELS:2002 Data Collection Materials ......................................................................................B-1
C: ELS:2002 Third Follow-up Field Test Questionnaire ............................................................C-1
D: ELS:2002 Cognitive Testing Reports, 2010 and 2011 .......................................................... D-1
E: ELS:2002 Scale Reliability Analyses...................................................................................... E-2
F: ELS:2002 Third Follow-up Field Test Codebooks ................................................................. F-1
viii
List of Tables
Number Page
1. Student locating snapshot, by dropout status and second follow-up response
status: 2011 ................................................................................................................. 27
2. Batch processing confirmation/update rates, by tracing source for 2010 and
2011: 2011 .................................................................................................................. 28
3. Address update participation by incentive group: 2011 ............................................. 29
4. Interview completion rate, by address update and incentive: 2011 ............................ 29
5. Sample members requiring intensive tracing procedures, by dropout status and
second follow-up response status: 2011 ..................................................................... 30
6. Interview completion rate, by address update and incentive: 2011 ............................ 30
7. Distribution of completed interviews, by data collection phase: 2011 ....................... 31
8. Number of cases and percentage of completed interviews by data collection
phase: 2011 ................................................................................................................. 31
9. Number and average of calls, by dropout status and prior- and current-round
response status and mode of interview: 2011 ............................................................. 32
10. Refusal and refusal conversion rates, by dropout status and prior-round
response status: 2011 .................................................................................................. 33
11. Distribution of third follow-up propensity cases by prior response status: 2011 ....... 36
12. Number of cases and percentage of completed interviews by data collection
phase: 2011 ................................................................................................................. 37
13. ELS:2002 field test treatment schedule ...................................................................... 37
14. Response rates by propensity and experimental group ............................................... 38
15. Estimates of select survey variables by propensity group (p < .05): 2011 ................. 39
16. Unit-level bias analyses: 2011 .................................................................................... 40
17. Questionnaire timing of ELS:2002 third follow-up field test questionnaire, by
response mode and questionnaire section: 2011 ......................................................... 44
18. Low-frequency responses to ELS:2002 third follow-up field test questionnaire
items: 2011 .................................................................................................................. 47
19. Items with “please specify” text: Methodological probes and items with an
“other” option: ELS:2002 third follow-up field test: 2011 ......................................... 50
20. Interview-reinterview agreement for categorical items on the ELS:2002 third
follow-up field test reliability reinterview: 2011 ........................................................ 53
ix
List of Tables
x
List of Figures
Number Page
1. Longitudinal design for the NCES high school cohorts: 1972–2016 ........................... 3
2. ELS:2002/12 tracing procedures: 2011 ...................................................................... 19
3. ELS:2002/12 website home page: 2011 ..................................................................... 22
4. Overall locating and interviewing rates at the time work stopped.............................. 26
xi
List of Figures
xii
Chapter 1.
Introduction
The Education Longitudinal Study of 2002 (ELS:2002) is conducted by RTI
International—a not-for-profit university-affiliated research organization—for the National
Center for Education Statistics (NCES), a part of the Institute of Education Sciences in the U.S.
Department of Education. This report describes the methodologies and results of the third
follow-up (ELS:2002/12) field test which was conducted in the summer of 2011.The field test
report is divided into six chapters:
• Chapter 1: Introduction
• Chapter 2: Field Test Survey Design and Preparation
• Chapter 3: Data Collection Procedures and Results
• Chapter 4: Field Test Questionnaire Timing and Data Quality
• Chapter 5: Survey Control Systems and Data Processing
• Chapter 6: Summary of Recommendations for the Full-scale Study
There also are six appendixes, which include proceedings of the two Technical Review
Panel meetings (appendix A), recruitment materials (field test letters and scripts) (appendix B), a
hardcopy facsimile of the electronic questionnaire (appendix C), a report on two sets of cognitive
interviews (appendix D), supplemental data on psychological scales (appendix E), and a
codebook with response frequencies (appendix F). Chapter 1 addresses two broad areas: it
provides information on the historical background of ELS:2002—the predecessor and successor
NCES Secondary Longitudinal Studies—and an overview of the design and purposes of
ELS:2002.
1
Chapter 1. Introduction
Reform Act of 2002 replaced the former Office of Educational Research and Improvement with
the Institute of Education Sciences, in which NCES is now housed.
The Secondary Longitudinal Studies program consists of three completed and two
ongoing studies. Completed studies are the National Longitudinal Study of the High School
Class of 1972 (NLS:72), the High School and Beyond (HS&B) longitudinal study of 1980, and
the National Education Longitudinal Study of 1988 (NELS:88). The two continuing studies are
ELS:2002 and the High School Longitudinal Study of 2009 (HSLS:09).
For ELS:2002—the fourth longitudinal study in the series—base-year, first, and second
follow-up data are now available; the ELS:2002 third follow-up full-scale study will take place
in the second half of 2012. For HSLS:09—the fifth study in the series—base-year data have now
been released, and first follow-up data will be collected in the spring of 2012. Taken together,
these studies describe the educational experiences of students from five decades—the 1970s,
1980s, 1990s, 2000s, and 2010s—and also provide bases for further understanding the correlates
of educational success in the United States.
Figure 1 includes a temporal representation of these five longitudinal education studies
and highlights their component and comparison points for the time frame 1972–2016.
1
For documentation of the NLS:72 project, see Riccobono et al. (1981) and Tourangeau et al. (1987).
2
Figure 1. Longitudinal design for the NCES high school cohorts: 1972–2016
12+14 F5 32
12+13 31
12+12 PST 30
12+11 PST 29
12+10 28
F4
12+9 27
F4-PST F3-PST
12+8 26
12+7 F4 25
F3
Year in School
12+6 24
12+5 23
Age
12+4 F3 F2-PST 22
F3
F2
12+3 21
12+2 F2 F1 F3 20
F2
12+1 F2
HST CU-P/S-HST 19
12 F1 BY-SA-P F2-SA-P-T-A-HST- 18
F1-SA-A-D F1-SA-A-
11
BY-SA HSES-D F1-SA-P- 17
3
HST-D
A-C-D 16
10
BY-SA-P-A BY-SA-P-T-A-L-F
9 F1-SA-T-A-HSES-D 15
BY-SA-P-A-C
8 14
BY-SA-P-T-A 13
7
2016
1972
1973
1974
1975
1976
1977
1978
1979
1980
1981
1982
1983
1984
1985
1986
1987
1988
1989
1990
1991
1992
1993
1994
1995
1996
1997
1998
1999
2000
2001
2002
2003
2004
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
Year of Data Collection
NLS:72=National Longitudinal Study of the High School Class of 1972 BY=Base-year data collection P=Parent survey PST=Postsecondary transcript
Chapter 1. Introduction
HS&B=High School and Beyond: 1980 F1=1st follow-up data collection T=Teacher survey C=Counselor questionnaire
NELS:88=National Education Longitudinal Study of 1988 F2=2nd follow-up data collection A=Administrator survey P/S=Parent or student
ELS:2002=Education Longitudinal Study of 2002 F3=3rd follow-up data collection L=Library/media center survey HSES=HS effectiveness study
HSLS:09=High School Longitudinal Study of 2009 F4=4th follow-up data collection F=Facilities checklist D=Dropout survey
F5=5th follow-up data collection CU=College update
HST=High school transcript
SA=Student assessment
Chapter 1. Introduction
A wide variety of data were collected in the NLS:72 surveys. For example, in addition to
background information about the student and his or her family, the base-year and follow-up
surveys collected data on each respondent’s educational activities (e.g., schools attended, grades
received, and degree of satisfaction with educational institutions). Participants were also asked
about their work experiences, periods of unemployment, job satisfaction, military service,
marital status, and children. Attitudinal information on self-concept, goals, community
involvement, and personal evaluations of educational activities were also included in the study.
The ELS:2002 sophomore cohort has no parallel in NLS:72, but the ELS:2002 and
NLS:72 senior cohorts can be compared.
2
For a summation of the HS&B sophomore cohort study, see Zahs et al. (1995). For further information on HS&B, see the
NCES website: http://nces.ed.gov/surveys/hsb/.
4
Chapter 1. Introduction
ELS:2002 data can be compared with the HS&B senior (1980) and sophomore (1980,
1982) cohorts; and indeed, the overall ELS:2002 design is modeled on the HS&B sophomore
cohort (both begin at 10th grade and returned to [mostly] the same schools 2 years later).
3
The entire compass of NELS:88, from its baseline through its final follow-up in 2000, is described in Curtin et al. (2002).
NCES keeps an updated version of the NELS:88 bibliography on its website. The bibliography encompasses both project
documentation and research articles, monographs, dissertations, and paper presentations employing NELS:88 as well as
ELS:2002 data (see http://nces.ed.gov/bibliography).
4
Types of schools that were excluded from the sample based on the HSLS:09 eligibility definition are described as part of the
discussion of the target population in the HSLS:09 Base-Year Data File Documentation (see chapter 3, section 3.2.1), Ingels et al.
(2011).
5
Chapter 1. Introduction
of 2013, which will be a survey of the cohort’s postsecondary plans and decisions. High school
transcripts will be collected in the 2013–14 academic year, and a second follow-up will take
place in 2015, when most sample members will be 2 years beyond high school graduation, or in
2016, when most sample members will be 3 years beyond high school graduation. Further
follow-ups are contemplated, to at least 2021.
The core research questions for HSLS:09 explore secondary to postsecondary transition
plans and the evolution of those plans; the paths into and out of science, technology, engineering,
and mathematics; and the educational and social experiences that affect these shifts.
Specific grade-based comparisons between HSLS:09 and ELS:2002 will not be possible,
because HSLS:09 is a study of entering 9th-graders (a cohort that will not be freshened at later
grades) and ELS:2002 is a study of spring-term (2002) high school sophomores and spring-term
(2004) high school seniors.
6
Chapter 1. Introduction
7
Chapter 1. Introduction
to student learning, predictors of dropping out, and high school correlates of students’ access to
and persistence and attainment in postsecondary education, and their entry into the workforce.
In the spring term of 2002, the base year of the study, high school sophomores were
assessed in reading and mathematics and surveyed in a national sample of high schools with 10th
grades. Their parents, teachers, principals, and librarians were surveyed as well.
In the first follow-up 2 years later (spring term 2004), base-year students who remained
in their base-year schools were resurveyed and tested in mathematics, along with a freshening
sample that makes the study representative of spring-term 2004 high school seniors nationwide.
Students who had transferred to a different school, had switched to a homeschool environment,
graduated early, or who had dropped out were administered a questionnaire. Also in the first
follow-up, academic transcripts were requested for all students who participated in either the
base year or the first follow-up. The transcripts normally 5 cover 4 years of coursework—for
students who were seniors in 2004, typically 9th through 12th grade. School course offerings
information for the base-year schools was also collected.
The second follow-up, in 2006, took place 2 years after modal graduation from high
school, and consisted of an electronically administered questionnaire, supplemented by matches
to various educational and postsecondary financial aid records. The administration of the third
and final follow-up will take place in the summer and fall of 2012, 6 years after modal
graduation from high school, followed by a postsecondary educational transcripts collection.
5
Although the goal is to collect four complete years of transcript data, this goal is not achieved for students who fall behind
modal grade progression—the transcripts collected for students who were held back, or who dropped out, are incomplete.
8
Chapter 1. Introduction
high school antecedents to later outcomes. For students who continue to higher education,
researchers can use ELS:2002 to measure the effects of their high school careers on subsequent
access to postsecondary institutions, their choices of institutions and programs, and, as time goes
on, their postsecondary persistence, attainment, and eventual entry into the labor force and adult
roles. For students who go directly into the workforce (whether as dropouts or high school
graduates), ELS:2002 can help to determine how well high schools have prepared these students
for the labor market and how they fare within it.
Key elements in the ELS:2002 longitudinal design are summarized by wave below.
6
Except where indicated otherwise, race/ethnicity is reported as follows: Black includes African American, Hispanic includes
Latino, Asian includes Native Hawaiian or Other Pacific Islander, and American Indian includes Alaska Native. All race
categories exclude individuals of Hispanic or Latino origin.
9
Chapter 1. Introduction
processes and mathematics coursetaking. The student questionnaire gathered information about
the student’s background, school experiences and activities, plans and goals for the future,
employment and out-of-school experiences, language background, and attitudes toward learning.
One parent of each participating sophomore was asked to respond to a parent survey. The
parent questionnaire was designed to gauge parental aspirations for their child, home background
and the home education support system, the child’s educational history prior to 10th grade, and
parental interactions with and opinions about the student’s school. For each sampled student, an
English teacher and a mathematics teacher were also selected to participate in a teacher survey.
The teacher questionnaire collected the teacher’s evaluations of the student and provided
information about the teacher’s background and activities. The head librarian or media center
director at each school was asked to complete a library media center questionnaire, which
inquired into the school’s library media center facility, staffing, technological resources,
collections and expenditures, and scheduling and transactions. Finally, the facilities checklist was
a brief observational form completed for each school, which collected information about the
condition of school buildings and facilities.
10
Chapter 1. Introduction
populations and the major analytical subgroups, the full-scale sample encompasses the following
types of students in the spring of 2004:
• ELS:2002 base-year sophomores enrolled (in either the 12th grade or some other
grade) in the school in which they were originally sampled;
• ELS:2002 base-year sophomores who dropped out of school prior to first follow-up
(2004) data collection;
• ELS:2002 base-year sophomores who finished high school early, including those who
graduated from high school early and those who did not graduate because they
achieved alternative certification (e.g., exam-certified equivalency such as a GED);
• ELS:2002 base-year sophomores who transferred out of the school in which they
were originally sampled (including homeschooled students);
• ELS:2002 base-year sample sophomores who were deemed unable to participate
directly during the base year as a result of severe disability or insufficient command
of the English language such that they could not complete a questionnaire; and
• students at the ELS:2002 base-year sample school who were enrolled in the 12th
grade in the spring term of 2004 but who were not in 10th grade in the United States
during the 2001–02 school year. In spring term 2002, such students may have been
out of the country, been enrolled in school in the United States in a grade other than
10th, had an extended illness or injury, been homeschooled, been institutionalized, or
temporarily dropped out of school. These students comprised the first follow-up
“freshening” sample.
Although all groups in the sample as categorized above were eligible to complete a
questionnaire, different instruments were tailored to different study populations. The guiding
intuition was to provide a core of items to which all sample members would respond,
supplemented by items specific to the circumstances of a particular group (such as dropouts, for
example, for whom questions about their current school situation would not be relevant).
11
Chapter 1. Introduction
• The second follow-up also captures three distinct (and sometimes alternating or
combined) transitions:
o transition from high school to the work force;
o transition from high school to postsecondary education; and
o transition to adult roles.
The second follow-up in the spring of 2006 employed a web-based self-administered
instrument with CATI and CAPI data collection for nonresponse follow-up. The focus of the
interview was on transition to postsecondary education and labor force participation. Out of a
sample of about 15,900 cases, about 14,200 sample members completed interviews.
The ELS:2002 second follow-up provides data to map and understand a key transition:
the path from high school to adulthood, as seen in the labor market, postsecondary education,
family formation, and civic engagement. The second follow-up collected information to enable
researchers and policymakers to better understand issues of postsecondary educational access
and choice. Thus, a major focus of the second follow-up was the postsecondary decision-making
process as reflected in applications to college and initial postsecondary enrollment histories.
ELS:2002, unlike studies that sample only postsecondary students, is uniquely positioned to
address these issues because it tracks respondents who attended postsecondary institutions before
they enrolled. Additionally, it follows students who did not attend college and thus provides
information on reasons students did not attend. The second follow-up also provides information
about high school completion (for students who dropped out or were held back), as well as
information about the status of dropouts and students who have obtained an alternative
credential, such as the GED. For non–college-bound students, the second follow-up mapped the
transition into the labor market. In addition to its focus on postsecondary (or sometimes
secondary) education and work experiences, the second follow-up survey also obtained
information about family formation. Indeed, workforce data are collected on everyone, including
those who are also enrolled in postsecondary education, thus supplying workforce data on all
cohort members who are employed full- or part-time. For further information about ELS:2002
base year through second follow-up, see Ingels et al. 2007
12
Chapter 1. Introduction
conducted, a formidable locating task remains, given that the age of the cohort marks a period of
high mobility, and given that the previous interview took place 6 years earlier. Finally, the 2011
field test also afforded a limited experiment into case response propensity modeling, monitoring
of data collection, and case prioritization with a focus on minimizing nonresponse bias and
maximizing the stability of estimates. (This is an alternative strategy to the tradition of
monitoring through response rates and assuming that a high response rate will necessarily result
in low bias.) Protocols and procedures for the ELS:2002/12 postsecondary transcripts study will
be evaluated in a separate, pilot study. The results of the 2011 field test, and recommendations
for the main study, are set out in the various chapters and appendixes that follow.
Main Study. The third follow-up sample design is straightforward: all members of the
second follow-up sample, regardless of their 2006 participation status, will be pursued in the
third follow-up, with the exception of the deceased. In addition, cases comprising study
withdrawals who have asked to not be recontacted and those who did not respond in any round,
although technically eligible, will not be fielded. Some fielded cases will be found to be out of
scope for the 2012 round—interviews will not be attempted if the sample member is incarcerated
or institutionalized or out of the country.
An approximately 35-minute web-based electronic questionnaire will be employed, with
the same questionnaire also available for CATI and field interviews. The focus of the
questionnaire records final outcomes across three topical areas:
• employment and career outcomes;
• education outcomes; and
• other outcomes reflective of attaining to adult status such as family formation, life
and career values, and civic engagement.
Products of the full-scale study will include a volume of Data File Documentation (DFD)
that covers instrumentation; sample design; data collection results and methods; data preparation
and planning; weighting, imputation, and design effects; and data file contents. Appendixes to
the DFD will include a facsimile of the electronic questionnaire and a hard-copy codebook with
response frequencies. There will also be a “First Look” report that sets out some basic
descriptive findings, and public and restricted use data sets, with suitable analysis systems.
13
Chapter 1. Introduction
14
Chapter 2.
Field Test Survey Design and Preparation
Chapter 2 reports on key elements of survey design and preparation for the 2011 field
test: instrument development (including its goals and constraints and the instrument development
process), and pre–data collection activities (chiefly recruitment and training of data collectors).
15
Chapter 2. Field Test Survey Design and Preparation
for all but the most salient experience and attitude questions. The third follow-up interview will
be supplemented with linkages to various administrative records, including postsecondary
educational transcripts. It is therefore critical that the questionnaire identify the names, locations,
and Integrated Postsecondary Education Data System codes for all postsecondary institutions
attended by sample members since high school, but not necessary to collect information on the
questionnaire that can be obtained from transcripts.
As a general principle, the 2012 content should focus on the unique features and special
strengths of the study by gathering outcomes for the areas in which it has already invested by
collecting critical antecedents. Information that uniquely fulfills the ELS:2002 design should be
collected in preference to duplicating the effort of other surveys that may be better suited than
ELS:2002 to tackle a particular issue. For example, the Department of Labor’s National
Longitudinal Surveys, youth cohorts (NLSY) is better equipped to collect complete labor market
event histories, because the NLSY interview is longer and the periodicity is one of far more
frequent interviews. The Beginning Postsecondary Students (BPS) study is being specifically
redesigned to optimally capture persistence in postsecondary education and in field of study.
BPS is better able to accommodate the fact of and to study nontraditional late entrants into
postsecondary education, while ELS:2002 is better able than BPS to examine issues of choice
because its cohort includes, prospectively, both those who do and who do not enter
postsecondary education. Indeed, postsecondary educational choice was the keynote of the
ELS:2002 second follow-up round in 2006, when the modal cohort age was 20, just as
educational attainment and early career experience will be the keynotes of the 2012 third follow-
up, when the cohort modal age is 26. Items that have been asked in prior NCES studies, such as
the National Education Longitudinal Study of 1988, have important extra value if repeated in
ELS:2002 in that they support cross-cohort analysis. However, these potential comparison items
should not be permitted to supplant more useful new constructs and items for the current study, if
the new constructs and items work longitudinally and promise fresh insights.
16
Chapter 2. Field Test Survey Design and Preparation
career theory items that were developed by Professor Lent; the second series (pre–full-scale
study) focused on potential new items concerning financial aid and economic literacy.
Fifth, before and after the field test, draft instruments were reviewed by the ELS:2002/12
Technical Review Panel, a specially appointed independent group of substantive,
methodological, and technical experts. A summary of the two panel meetings (held in September
2010 and November 2011) is found in appendix A. Sixth, justifications were written for
questionnaire items for Office of Management and Budget (OMB) review.
Seventh, actual development of the questionnaires, including specification, routing,
programming, and testing, took place within an RTI proprietary system, the Hatteras Survey
Engine and Survey editor. (A description of Hatteras appears in chapter 5.) Eighth, a field test
was conducted to test questionnaire and assessment items; that field test is documented in this
report, which includes recommendations for the main study instruments and procedures.
A hardcopy version of the electronic field test questionnaire is found in appendix C. The
quality of the questionnaire data is addressed in chapter 4.
17
Chapter 2. Field Test Survey Design and Preparation
participated in telephone interviewer project training. The project supervisor training included an
overview of the study, active listening techniques expected of interviewing staff, problem
resolution, and other specific project procedures and protocols.
Telephone Interviewers. As the primary point of contact with sample members, TIs
were responsible for gaining cooperation from and conducting interviews with sample members,
avoiding interview refusals, and addressing the concerns of reluctant sample members. Because
of this integral role in the study, TIs received a project manual and 12 hours of training that
included
• an overview of ELS:2002/12;
• an in-depth review of the questionnaire;
• hands-on practice administering the telephone interview;
• review of appropriate conversational interviewing techniques; and
• review of proper methods in handling inbound calls.
To conclude the training and verify that the material had been learned, all telephone interviewers
were certified by successfully conducting mock telephone interviews and by providing
satisfactory responses to the study’s FAQs.
In addition to the training described above, QCSs and TIs were also cross-trained as help
desk agents who assisted any sample members who had questions or problems while completing
web interviews or called in to complete the telephone interview.
Tracing Staff. Tracing staff (tracers) used intensive measures, described in section 2.2.2,
to locate sample members who lacked good telephone contact information. Tracers attended a
comprehensive 16-hour training session led by RTI tracing managers and covered all tracing
procedures. ELS:2002/12 tracers were cross-trained as TIs on the same project.
Additionally, weekly quality circle meetings were routinely conducted as an extension of
the training program for continual quality improvement. Quality circle meetings are discussed in
greater detail in section 3.1.
18
Chapter 2. Field Test Survey Design and Preparation
Located Nonrespondent/
Not located Located Unable to Locate
Intensive
Intensive tracing
tracing
Not Located operations Not Located
operations
(TOPS-2)
(TOPS-1)
Interview
Yes
Respondent
NOTE: CATI = computer-assisted telephone interviewing. CPS = Central Processing System. NCOA = National
Change of Address.
SOURCE: U.S. Department of Education, National Center for Education Statistics, Education Longitudinal Study of
2002 Second Follow-up (ELS:2002/12).
Batch Tracing. Batch database searches were conducted on all sample members to
update contact information in preparation for mailing activities. These searches used the U.S.
Department of Education’s Central Processing System and the U.S. Postal Service (USPS)
National Change of Address databases. Then, just prior to the start of outbound telephone
interviewing, all sample members were sent to Phone Append, which searches more than 400
million landline, voice over Internet protocol, and wireless numbers in the United States, Puerto
Rico, and Canada. All information obtained from these sources was compared with the
information previously available from the ELS:2002/12 second follow-up locator database to
identify any new contact information.
Mailings. An initial letter announcing the data collection was mailed on July 12, 2011,
by USPS first-class mail. The mailing to all sample members included the following:
• a letter, signed by both the project director and the NCES project officer, that
announced the start of data collection;
• information about the incentive;
• a link to the study website;
• login credentials for accessing the web interview;
• the ELS e-mail address and toll-free help desk number; and
• a brochure about ELS:2002/12.
19
Chapter 2. Field Test Survey Design and Preparation
In addition to the letter, an e-mail was sent to sample members 2 days later, on July 14,
2011. Both the letter and the e-mail encouraged sample members to complete the survey online
during the early response period. Additional mailings during this early period included a
reminder postcard and letter and additional e-mail reminders to encourage early interview
response. Once outbound telephone interview efforts began, periodic reminder mailings and e-
mails were sent to sample members throughout the course of data collection. Copies of letters
and brochures can be found in appendix B.
CATI Locating and Pre-intensive Tracing. Once outbound telephone interviewing
began, telephone interviewers conducted limited tracing and locating activities, as needed. The
telephone number believed to be the best known number for contacting the sample member was
attempted first. If the sample member could not be reached at that number after several attempts,
any other numbers associated with the sample member, including parent and other contacts, were
called. The interviewers attempted to gather locating information for the sample member from
the contact who answered the call. If the sample member could not be located, the case was sent
for intensive interactive tracing by RTI’s Tracing Operations (TOPS).
Intensive Tracing. Cases that could not be located through batch tracing or CATI
locating efforts underwent intensive tracing by RTI’s TOPS. Using a number of public domain
and proprietary databases, TOPS uses a two-tiered strategy. The first tier (TOPS-1) used Social
Security numbers (SSNs) to search for sample members in consumer databases such as
FastData’s SSN search and Experian which contains current address and telephone listings for
the majority of consumers with credit histories. If a search generated a new telephone number for
the sample member, tracers attempted to confirm the information by speaking with the sample
member or with someone else who could confirm the information. If the number was confirmed,
the case was sent back to CATI for telephone interviewing. This first level of effort minimized
the time that cases were in tracing and unavailable for CATI efforts.
If cases were not located (or locating information not confirmed) at the end of TOPS-1,
the cases underwent a more intensive level of tracing in TOPS-2, which included calls to other
possible sources of information, such as directory assistance, alumni offices, and contacts with
neighbors or landlords. Whenever any of these sources provided information that indicated that a
sample member was not available for the study (e.g., deceased, incarcerated, or out of the
country), no further contact efforts were made.
20
Chapter 3.
Data Collection Procedures and Results
Data collection procedures and outcomes for the field test will be covered in this chapter.
The procedures section describes the phases of data collection and modes of interview
administration. The outcomes section includes the locating rate and response rates by data
collection phase and mode as well as high school dropout status and second follow-up response
status. The response propensity experiment procedures and results are described in the final
section.
21
Chapter 3. Data Collection Procedures and Results
Sample members were provided with a toll-free telephone number, which was answered
by help desk agents. Help desk staff were available to sample members who had questions or
technical issues related to completion of the web interview. For each call received, staff
confirmed contact information for the sample member and recorded a description of the problem
and resolution. If technical difficulties prevented sample members from completing the web
interview, help desk staff were able to complete a telephone interview. Two common types of
help desk incidents were requests for login credentials and requests to complete the interview
over the telephone. To minimize the need for help desk assistance, a “Forgot Password?” link
was included on the study website and the need to disable pop-up blockers to launch the survey
was eliminated.
3.1.2 Interviewing
Data collection for the ELS:2002/12 field test interview consisted of three phases:
1. Phase 1: Early Response. This phase began with the start of data collection in July
2011 and lasted approximately 3 weeks until August 1, 2011. Sample members were
encouraged to complete the survey over the Web during this phase. The telephone
interview was available to sample members who contacted the help desk, but no
outbound telephone calls were made. Sample members who completed the interview
were eligible to receive an incentive of $25 or $45 based on their response propensity
experiment group assignment (see section 3.2 for more information on the
22
Chapter 3. Data Collection Procedures and Results
experiment). The sample members were divided into two groups, high propensity and
low propensity. Sample members in the high-propensity group were offered $25
while half of the low-propensity cases were offered $45 (experimental group) and
half were offered $25 (control group).
2. Phase 2: CATI Production. The computer-assisted telephone interviewing (CATI)
production phase began on August 1, 2011. During this phase, interviewers called to
encourage sample members to complete the interview by telephone or on the Web.
Sample members who completed the interview during the production phase were
eligible to receive the same incentives as during the early response phase.
3. Phase 3: Increased Incentive. The increased incentive phase began on September
13, 2011. During this phase, the incentive values were increased to $35 for high-
propensity cases and low-propensity control cases and $55 for low-propensity
experimental cases. Sample members were alerted with a letter and e-mail, and by
interviewers if successful contact was made.
Sample members could complete the interview on the Web or by telephone throughout
the entire data collection period. For the telephone interviewer, the interview screens were
identical to those in the web interviews completed by respondents, except that instructions on
how to administer each question were visible at the top of each screen for telephone interviews.
Following are details of the administration of the interview through the various modes.
Web Interviews. Sample members were informed of the web interview in all project
communications including mailings, e-mails, and telephone contacts. During the early response
period (the first 3 weeks of data collection), only web interviews were completed unless sample
members initiated a telephone interview by calling the help desk or sending an e-mail asking to
be called. Reminder mailings and e-mails were sent throughout all phases of data collection to
encourage sample members to complete the interview online. The website was accessible
24 hours a day, 7 days a week, throughout the data collection period, providing sample members
with the option to complete the interview online at any time.
Telephone Interviews. Outbound telephone follow-up began on August 1, 2011, after
the 3-week early response period ended. Telephone interviewers attempted to locate, gain
cooperation from, and interview sample members who had not yet completed the interview.
Interviewers encouraged sample members to complete the interview by telephone; however, the
web survey remained available throughout data collection for sample members who preferred
that option. Sample members who did express a preference to complete a web interview were
called back 5 days later for follow-up if the interview had not yet been completed.
The CATI Case Management System (CATI-CMS) included an automated call scheduler
that assigned cases to interviewers by case priority, time of day, day of week, existence of
previously scheduled appointments, and type of case. Case assignment was designed to
maximize the likelihood of contacting and interviewing sample members, and cases were
assigned to various queues accordingly. For example, the CATI-CMS included queues for new
cases that had not been called, Spanish-language cases, initial refusals, and various appointment
23
Chapter 3. Data Collection Procedures and Results
queues. In addition, available telephone numbers for each case were automatically prioritized for
the interviewers. As new telephone numbers were added—as a result of CATI tracing, other
tracing efforts, and information from other sources such as respondent e-mails or help desk call-
ins—available telephone numbers were reprioritized based on the new information.
Some cases required special treatment. To gain cooperation from those sample members
who initially refused to participate (and from contacts such as parents and roommates who acted
as gatekeepers to the sample member), interviewers were trained in refusal conversion
techniques. For example, a telephone interviewer will recontact the person and acknowledge the
issue or concern the person conveys and listen carefully. He or she will then speak to the specific
concern and also explain the importance of the study and the importance that individuals similar
to the sample member are represented in the data.
Help Desk Operations. Throughout ELS:2002/12 field test data collection, telephone
interviewers and quality control supervisors served as help desk agents for the toll-free help line.
They were available to assist sample members who had questions or problems accessing and
completing the web interview or who called in to complete the telephone interview. If technical
difficulties prevented sample members from completing the web interview, sample members
were encouraged to complete a telephone interview.
Quality Circle Meetings. Weekly quality circle meetings were conducted to serve as an
essential feedback loop for ensuring that project staff and call center staff were communicating
on a regular basis about the goals of the study and addressing challenges encountered along the
way. These meetings provided a forum for discussing elements of the instrument design and
interview cooperation tactics, motivating the group toward the goals of the study, and obtaining
feedback on data collection issues. Weekly quality circle meetings for telephone staff were held
at the call center. Issues discussed at these meetings were documented in meeting notes and
stored where staff could access them as needed. The interviewers were informed of counts of
interview completions to date, general data collection issues and issues specific to the survey
instrument, and project staff responses to questions from interviewers.
Throughout the study, a variety of issues were addressed at the quality circle meetings
that reinforced specific content from training and contributed to prompt problem solving. Some
of the issues covered in these meetings included the following:
• clarification of questions and item responses and reinforcement of positive
interviewing techniques;
• methods of gaining cooperation from sample members and gatekeepers (e.g., spouses,
parents, and roommates);
• problem sheets submitted during contacting and interviewing;
• the importance of providing and reviewing detailed case comments;
• data security protocols; and
• study progress and general morale boosting.
24
Chapter 3. Data Collection Procedures and Results
Telephone interviewers and their supervisors were given the opportunity to ask questions
in quality circle meetings, and as needs were identified, additional training topics were
highlighted and addressed in subsequent meetings.
7
A partial interview is defined as any case where the respondent began the interview, completed it through the high school
completion section, but broke off and did not return to complete the interview.
25
Chapter 3. Data Collection Procedures and Results
Figure 4. Overall locating and interviewing rates at the time work stopped
Sample
n = 1056
Located
Respondent Exclusions
Nonrespondent
n = 602 n = 17
n =87
Locating Results and Interview Response Rates. Overall snapshot locating rates and
response rates for the ELS:2002/12 interview were determined at the point that data collection
ended early. Thus, the rates are only intended to serve as a comparison between groups and not
indicative of final rates expected from a data collection that runs its full course. The snapshot
locating and response rates varied by high school dropout status and prior-round response status,
as shown in table 1. As alluded to above, the locating rate was defined as the percentage of cases
in which an interview was completed, the sample member was successfully contacted, or the
contact information in the study database was not proven out of date or incorrect during contact
attempts. The snapshot locate rate was significantly higher for those who never dropped out
compared to those who ever dropped out of high school (t = 2.83, p < .05). Regarding response
rate, of field test sample members who ever dropped out, the snapshot response rate was
significantly lower compared to those who never dropped out (t = 3.49, p < .05).
26
Chapter 3. Data Collection Procedures and Results
Table 1. Student locating snapshot, by dropout status and second follow-up response status:
2011
Located snapshot Responding sample members
ELS:2002 dropout status and Total Percent of Percent Percent
second follow-up response status eligible Number total Number of located of total
Total 1,039 689 66.3 602 87.4 57.9
Dropout status
Ever dropped out 95 50 52.6 39 78.0 41.1
Not ever dropped out 944 639 67.7 563 88.1 59.6
Second follow-up response status
ELS:2002/06 respondent 771 560 72.6 502 89.6 65.1
ELS:2002/06 nonrespondent 268 129 48.1 100 77.5 37.3
NOTE: Numbers may not sum to total because of rounding.
SOURCE: U.S. Department of Education, National Center for Education Statistics, Education Longitudinal Study of
2002 Third Follow-up (ELS:2002/12) Field Test.
In addition to differences by high school dropout status, snapshot locate and response
rates differed by prior-round (ELS:2002/06) response. The snapshot locate rate was significantly
higher for those who responded to ELS:2002/06 (second follow-up) compared to those who did
not respond to the second follow-up (t = 7.10, p < .05). Interview snapshot response rates also
differed with a significantly higher percentage of second follow-up respondents completing the
third follow-up field test interview in contrast to second follow-up nonrespondents (t = 8.14,
p < .05).
A concerted effort must be made in the full-scale study to convince parents of the value
of the study so they cooperate and share information with their children. It was learned that in
some cases parents continue to act as gatekeepers for the sample members even though the
sample members are several years removed from being minors (15 percent of refusals in the field
test were by someone other than the sample member). The full-scale study procedures include
dual mailings to both the sample members and their parents. These procedures will ensure at
both the panel maintenance and data collection stages that parents are aware of the procedures
for the third follow-up. Direct mail and e-mail contacts with parents will also allow parents to
provide updated contact information for their young adults. Also, because some phone numbers
available for sample members are numbers for their parents’ homes, CATI procedures will be in
place to guide interviewers on how to appropriately ask for and record new contact information
for sample members from parents. Successful contacts with parents will be an important part of
interviewer training for the full-scale study.
Locate Rates by Source of Batch Update. Prior to the ELS:2002/12 field test data
collection, batch matching was conducted for the 1,056 sample members. This effort confirmed
contact information or provided new contact information for several hundred field test cases, as
shown in table 2. The highest proportion of matches was obtained through FirstData (National
Change of Address and Phone Append), 62 percent of cases sent in 2010 and 58 percent in 2011.
The smallest proportion, 20 percent of cases sent in 2010 and 19 percent in 2011, were matched
27
Chapter 3. Data Collection Procedures and Results
through the Central Processing System (CPS). Because CPS relies on postsecondary student loan
application data, it is unsurprising that CPS matched the fewest records because only 65 percent
of ELS:2002/06 respondents reported attending postsecondary education courses in the previous
interview, and it had also been 8 years since the majority of sample members completed high
school so the numbers attending postsecondary institutions were likely to be low. Also, not all
postsecondary students fill out the application for federal student aid and those students would
not appear in the CPS data.
Table 2. Batch processing confirmation/update rates, by tracing source for 2010 and 2011: 2011
Address Update Mailing Results. The last panel maintenance address update mailing
for the cohort, in which sample members and one parent are asked to confirm or update contact
information, was conducted in fall/winter of 2010. Address updates or confirmations were
received from 230 field test sample members (22 percent) in response to the request, which was
sent via mail and e-mail. The response was double the response seen in the 2007 and 2008
address update efforts (11 percent and 10 percent, respectively). An incentive experiment was
conducted during the 2010 panel maintenance effort to assess whether offering a $10 incentive
would increase the response rate for an address update mailing. The sample was split in half. One
half was offered $10 if the sample member or a parent confirmed or updated contact information,
while the other half was offered no incentive. The experiment results showed that the $10 group
had a significantly higher address update response rate (25 percent) than the $0 group (20
percent, one-tailed t = 1.90, p < .05), as shown in Table 3. Based on the positive experiment
results, ELS:2002/12 plans include offering a $10 incentive to full-scale sample members if they
or a parent updates or confirms contact information when the fall 2011/winter 2012 address
update mailing is conducted.
28
Chapter 3. Data Collection Procedures and Results
Participated
1
Type of incentive Eligible sample Number Percent!
$0 495 99 20.0!
$10 489 124 25.4!
1
Eligible sample includes sample members for whom an address was known and the mailing was not returned as
undeliverable.
SOURCE: U.S. Department of Education, National Center for Education Statistics, 2002 Education Longitudinal Study
(ELS:2002/12) Third Follow-up Field Test.
Sample Members Requiring Intensive Tracing. Overall, about 160 (16 percent) of the
1,039 eligible field test sample members required intensive tracing (table 5) before data
collection was closed early. Thirty-three percent of sample members who ever dropped out of
high school required intensive tracing, compared with 14 percent of sample members who never
dropped out. Thirty-six percent of second follow-up nonrespondents required intensive tracing,
compared with 9 percent of second follow-up respondents. Of the cases located through intensive
tracing before data collection ended early, about 73 percent completed the ELS:2002/12
interview.
29
Chapter 3. Data Collection Procedures and Results
Table 5. Sample members requiring intensive tracing procedures, by dropout status and
second follow-up response status: 2011
Dropout status and second follow-up response Cases requiring intensive tracing
status Total Number Percent
Total 1,039 163 15.7
Dropout status
Ever dropped out 95 31 32.6
Never dropped out 944 132 14.0
Second follow-up response status
ELS:2002/06 respondent 771 66 8.6
ELS:2002/06 nonrespondent 268 97 36.2
NOTE: Numbers may not sum to total because of rounding.
SOURCE: U.S. Department of Education, National Center for Education Statistics, Education Longitudinal Study of
2002 Third Follow-up (ELS:2002/12) Field Test.
Interviewed
Number Percent
Web without telephone contacts 212 35.2
Web with telephone contacts 286 47.5
Telephone interview 104 17.3
SOURCE: U.S. Department of Education, National Center for Education Statistics, Education Longitudinal Study of
2002 Third Follow-up (ELS:2002/12) Field Test.
30
Chapter 3. Data Collection Procedures and Results
Interviewed
Number Percent
Eligible 1,039 †
Respondent 602 57.9
1
Early response phase 217 36.0
Computer-assisted telephone interview phase 307 51.0
2
Increased incentive phase 78 12.9
† Not applicable.
1
Of the 217 respondents, 212 were complete and 5 were partially complete.
2
Of the 78 respondents, 76 were complete and 2 were partially complete.
SOURCE: U.S. Department of Education, National Center for Education Statistics, Education Longitudinal Study of
2002 Third Follow-up (ELS:2002/12) Field Test.
Each phase of data collection can be evaluated individually. The early response phase of
data collection yielded 217 cases that were either complete or partially complete. Of the 217
cases, 212 were fully complete, translating to a 20 percent response rate (212 completed out of
1,039 eligible) (table 8). The next phase of data collection yielded a 37 percent response rate
(307 completed out of 827 eligible for the second phase). The third phase yielded 78 cases that
were either complete or partially complete. Of the 78 cases, 76 were fully complete, translating
to a 15 percent response rate (78 completed out of 520 eligible by the third phase). As mentioned
previously, the target yield was met prior to the third phase, but it was conducted briefly to test
procedures and determine whether additional yield would result from those procedures.
Table 8. Number of cases and percentage of completed interviews by data collection phase:
2011
Telephone Interviewer Hours and Call Counts. During the course of ELS:2002/12
data collection 100 telephone interviews were completed. On average, telephone interviews that
were completed in a single session took 41.3 minutes to administer. Most of the telephone
interview hours not spent administering an interview were spent on case management activities
such as locating and contacting sample members, prompting sample members to complete
interviews, reviewing call history, scheduling appointments for callbacks, and entering detailed
comments and suggestions to assist with reaching and interviewing sample members. Some of
the time was also spent responding to incoming help desk calls.
31
Chapter 3. Data Collection Procedures and Results
On average, 13 calls were made per ELS:2002/12 field test sample member during the
data collection period. 8 Of the sample, cases that completed the current-round (the third follow-
up) interview required an average of 7 calls, while current-round nonrespondents received an
average of 21 calls during the interviewing period. There were no significant differences in call
counts between respondents who completed interviews over the telephone and respondents who
completed interviews over the Web but with phone prompting. The average number of telephone
calls is shown in table 9.
Table 9. Number and average of calls, by dropout status and prior- and current-round response
status and mode of interview: 2011
Average number
Response status and mode Number of cases Number of calls of calls
Total 1,039 13,527 13.0
Dropout status
Ever dropped out 95 1,361 14.3
Never dropped out 944 12,166 12.9
Second follow-up response status
ELS:2002/06 respondent 771 9,708 12.6
ELS:2002/06 nonrespondent 268 3,819 14.3
Current-round response status
Respondent 602 4,233 7.0
Web interviews 498 3,168 6.4
Excluding early response 287 3,168 11.0
Telephone interviews 104 1,065 10.2
Nonrespondent and exclusions 454 9,367 20.6
NOTE: Numbers may not sum to total because of rounding.
SOURCE: U.S. Department of Education, National Center for Education Statistics, Education Longitudinal Study of
2002 Third Follow-up (ELS:2002/12) Field Test.
Averting and Converting Refusals. Refusal aversion and conversion techniques were
integrated into telephone interviewer training and were reinforced throughout data collection in
quality circle meetings. Interviewers were encouraged to share their experiences gaining sample
member cooperation and seek guidance from the group. Sample members who refused to
complete the interview were placed in a separate queue and worked by a subset of interviewers
selected for additional refusal conversion training. Overall, 10 percent of eligible cases ever
refused. Of these refusals, 21 percent subsequently completed the interview (table 10) after
refusal conversion attempts were made. In the previous round (second follow-up), 13 percent of
eligible cases ever refused and, of these, 8 percent subsequently completed the interview.
8
This includes sample members who required no call attempts.
32
Chapter 3. Data Collection Procedures and Results
Table 10. Refusal and refusal conversion rates, by dropout status and prior-round response
status: 2011
Ever refused
interview Interviewed, given refusal
Total Percent Percent Percent of
eligible Number of total Number of refused total eligible
Total 1,039 103 9.9 22 21.4 2.1
Dropout status
Ever dropped out 95 6 6.3 1 16.7 1.1
Never dropped out 944 97 10.3 21 21.6 2.2
Prior-round response status
ELS:2002/06 respondent 771 69 8.9 18 26.1 2.3
ELS:2002/06 nonrespondent 268 34 12.7 4 11.8 1.5
SOURCE: U.S. Department of Education, National Center for Education Statistics, Education Longitudinal Study of
2002 Third Follow-up (ELS:2002/12) Field Test.
Debriefing Meeting. At the conclusion of the ELS:2002/12 data collection, project staff
held a debriefing meeting with interviewers to learn more about their experiences. The
interviewers reported positive overall experiences working on the study and provided
suggestions for improving full-scale data collection. With regard to training, interviewers
suggested additional practice on the CATI-CMS, occupation coder, and refusal conversion.
Based on their interactions with sample members and other contacts, interviewers also provided
examples of successful techniques for gaining cooperation to incorporate into training for full-
scale data collection. Interviewers reported that they received the resources necessary to
administer the interview successfully, and identified specific questions that sometimes presented
challenges for survey respondents. Interviewers were generally appreciative of the support they
received through monitored interviews and quality circle meetings. Project staff prepared a
summary of the debriefing meeting for consideration when planning the full-scale survey.
33
Chapter 3. Data Collection Procedures and Results
Traditionally in survey data collections, potential bias has been addressed through
weighting adjustments after the conclusion of the data collection period. However, using
responsive design, the methodology attempts to determine whether the risk of bias can be
effectively addressed prior to data collection. Using propensity models to predict response in the
second follow-up field test, sample members with a low likelihood of responding (low
propensity) were identified prior to the third-follow-up field test data collection and were
subsequently targeted for special interventions in an effort to reduce the potential overall bias
introduced into final survey estimates.
The most important assumption evaluated in the field test was whether low-propensity
cases are fundamentally different from high-propensity cases, particularly in the way they
respond to survey questions. If differences in estimates between low- and high-propensity cases
do exist and are large enough, survey estimates are likely to be affected. Thus, low-propensity
cases could contribute to nonresponse bias. 9 Therefore, bringing more low-propensity cases into
the respondent pool could have several additional advantages. First, more low-propensity cases
means an expanded and richer imputation donor pool for traditionally underrepresented cases.
Second, more low-propensity cases being interviewed could result in more precise weights
simply because more data are collected. Finally, targeting resources strategically to cases likely
to yield more value in terms of bias is an efficient approach to data collection.
With this in mind, the goal of the approach was to reduce nonresponse bias by employing
a different methodology for low likelihood of response cases prior to the start of data collection.
The experiment implemented in the field test evaluated the ability to identify low-propensity
cases a priori, the effectiveness of an increased cash incentive on low-propensity cases, how
survey responses may differ between high- and low-propensity cases, and how low-propensity
cases may contribute potential bias in final estimates.
Experimental Process. To assign a case’s response propensity for the third follow-up
field test, using a logistic regression model we estimated their response propensity in the second
follow-up field test by using the sample member’s known ELS:2002 second follow-up field test
response status as the dependent variable. As independent variables, a range of information
known for all respondents and nonrespondents from each prior wave of the longitudinal field test
including information from batch tracing activities was examined for significance. The following
variables were considered as predictors of a field test sample member’s second follow-up field
test response outcome:
• base-year response status;
• first follow-up response status;
• whether the respondent ever refused;
• whether the respondent ever scheduled an appointment to complete the interview;
9
For a more extensive explanation of the relationship between a case’s response propensity and nonresponse bias, please see
Peytchev et al. (2010).
34
Chapter 3. Data Collection Procedures and Results
10
The field test sample size was 1,056 with 17 exclusions (see section 3.2) resulting in 1,039 cases pursued in the third follow-up
field test data collection.
35
Table 11. Distribution of third follow-up propensity cases by prior response status: 2011
Table 12 shows the distribution of the case propensities across some demographic
characteristics of interest. From the data in table 11, the overall sample percentages were similar
in the high-propensity groups and in the overall sample. There is no obvious over- or
underrepresentation in any one subgroup of cases.
Treatment for Low-Propensity Cases. The basic premise of the response propensity
approach is to identify low-propensity cases prior to data collection and assign to them a
treatment, such as a cash incentive, to encourage response. The treatment for ELS:2002/12 field
test low-propensity cases was a higher incentive of $45 at the start of data collection (weeks 1–9)
that increased to $55 at week 10. High-propensity and control group cases were offered $25 until
week 10 of data collection, when the incentive increased to $35. Table 13 outlines the timing and
levels of the different incentives.
Results of the Experiment. The predictive model developed ahead of the field test data
collection successfully informed the eventual response outcome for sample members. As shown
in table 14, the response rate of the high-propensity group (67 percent) was significantly higher
than that of the low-propensity control group (45 percent) (χ2 = 34.9; p < .0001).
Table 12. Number of cases and percentage of completed interviews by data collection phase:
2011
37
Chapter 3. Data Collection Procedures and Results
Percentage Percentage
respondents N nonrespondents N Total
Total eligible 1,039
High propensity 67.4 351 32.6 170 521
Low propensity (control) 45.4 118 54.6 142 260
Experimental group 51.6 133 49.5 125 258
Control group 45.4 118 54.6 142 260
NOTE: Of the 1,056 field test sample cases, 1,039 were eligible for the third follow-up collection. See section 3.2 of
the report for information on eligibility.
SOURCE: U.S. Department of Education, National Center for Education Statistics, Education Longitudinal Study of
2002 Third Follow-up (ELS:2002/12) Field Test.
In examining the effect of the higher incentive treatment for low-propensity cases, a
numerical difference in participation (52 percent for treatment cases and 45 percent for control
cases) was observed. However, the difference was not statistically significant (z = 1.41, p = .08).
One of the analytical goals of the field test was to determine whether high- and low-
propensity cases in ELS:2002 answer survey questions differently (i.e., exhibit item-level
differences). Following is a list of select survey variables likely of high value to analysts
showing significant differences between propensity groups (p < .05). Their estimates can be seen
in table 15.
• Whether respondent has earned a regular high school diploma
• Whether respondent has attended a postsecondary institution
• Highest level of postsecondary school attended
• Postsecondary attainment (specifically, postsecondary attendance with no credential,
and bachelor’s degree)
• Educational expectations (specifically, 1- or 2-year degree, master’s, and PhD)
• Current marital status (specifically, partnered/living with a significant other)
• Whether respondent is currently registered to vote
38
Chapter 3. Data Collection Procedures and Results
Table 15. Estimates of select survey variables by propensity group (p < .05): 2011
Low High
propensity propensity
Group (percent) (percent)
Has a regular high school diploma 91.0 95.8
No postsecondary attendance 12.4 6.2
Highest postsecondary institution attended is a 4-year 53.7 64.3
postsecondary attendance but no credential 44.2 29.4
Highest postsecondary credential earned is a bachelor’s 26.6 39.9
Highest level of education expected: complete 1- or 2-year program 17.5 10.9
Highest level of education expected: Master’s 22.8 30.5
Highest level of education expected: PhD, professional doctorate 8.9 16.1
Current marital status is living with significant other 13.7 7.5
Currently registered to vote 79.8 89.4
SOURCE: U.S. Department of Education, National Center for Education Statistics, Education Longitudinal Study of
2002 Third Follow-up (ELS:2002/12) Field Test.
Other variables showing significant differences between propensity groups (p < .05)
include the following:
• Currently taking courses at voc/tech/trade school
• Currently working and attending postsecondary school
• Debt/asset ratio is in highest quartile
Other variables with differences between propensity groups approaching significance
(p < .1) include the following:
• If sold assets and paid debt: break even
• Debt/asset ratio is in middle two quartiles
• Voted in presidential election
• Voted in state or local election
Variables showing no significant differences between propensity groups include the following:
• Currently working at full-time job
• Currently working at part-time job
• Currently taking courses at 2- or 4-year college
• Not currently working and not currently attending postsecondary school
• Currently working but not attending postsecondary school
• Currently attending postsecondary school but not working
• Timing of high school credential
• Taken out any education loans
• Amount owed on education loans
• Number of hours working per week
• Ever unemployed since January 2008
39
Chapter 3. Data Collection Procedures and Results
Bias is defined as the weighted absolute difference between an estimate calculated using
respondents only and an estimate calculated using nonrespondents only. The weight for a bias is
defined as the nonresponse rate associated with the bias calculation. A relative bias is calculated
by dividing the bias by the estimate calculated using both respondents and nonrespondents.
Propensity Modeling Conclusions. The response propensity model successfully
predicted response outcome. The inclusion of low-propensity cases showed an apparent
reduction in unit-level biases. Including more low-propensity cases in the data may reduce bias
and may help improve final estimates because low-propensity cases appear to be different in
terms of their survey responses. Though the results were not statistically significant, a higher
incentive amount produced a numerically higher response rate for low-propensity cases. It is not
known how the experiment would have concluded if the field test data collection had continued 2
additional months to its scheduled end. The data collection ended two months early because the
interview yield goal was met. If data collection had continued, more cases would have been
traced and located, and more cases would have been successfully reached and interviewed.
There are important lessons learned for the full-scale study. First, from an operational
standpoint, strategic case targeting can be implemented well for ELS:2002/12. Interventions
were effectively delivered to targeted cases providing confidence that this or a similar process
could also be implemented successfully on the much larger full-scale study. However, this field
40
Chapter 3. Data Collection Procedures and Results
test experiment revealed only small differences in measured bias; therefore, propensity to
respond may not be the best method for strategic case targeting. New approaches for bias
reduction should be considered for the full-scale study. New statistics, such as the Mahalanobis
distance function, within a responsive design, should be considered for the full-scale study.
Although case propensity may end up not providing sufficient reductions in bias in the
ELS:2002/12 data, the process of strategic case targeting for the purpose of improved data
quality will be implemented in the study for several important reasons. Prior research has found
that nonresponse bias is related to how and with whom new interviews are obtained during
nonresponse follow-up (Peytchev et al., 2009). Specifically, data collection actions during
nonresponse follow-up can be detrimental to data quality by bringing in sample members who
resemble those most likely to respond (Schouten et al., 2009). In the full-scale study, the goal for
data collection will be to conduct nonresponse follow-up to maximize the chances of a fully
representative respondent pool. Chapter 6 outlines full-scale study recommendations based on
the results discussed above.
41
Chapter 3. Data Collection Procedures and Results
42
Chapter 4.
Field Test Questionnaire Timing
and Data Quality
This chapter presents findings from analyses of questionnaire timing, item timing, item
frequency distributions, “please specify” text, reliability reinterviews, and the reliability and
dimensionality of scales formed from sets of items. In addition to the findings, recommendations
for the full-scale study are included. All recommendations may not necessarily be implemented,
owing to survey constraints and National Center for Education Statistics (NCES) priorities.
43
Chapter 4. Field Test Questionnaire Timing and Data Quality
Table 17. Questionnaire timing of ELS:2002 third follow-up field test questionnaire, by response
mode and questionnaire section: 2011
No significant differences in the average completion time were found in the overall
questionnaire or in any section when comparing by mode of administration (online or by
computer-assisted telephone interview [CATI]) (table 1). Questionnaire timing therefore does
not suffer from mode effects and does not require adjustments specific to the mode of
administration.
44
Chapter 4. Field Test Questionnaire Timing and Data Quality
field test respondents averaged 26 seconds answering it. The question may be
conceptually difficult because it asks respondents to judge various combinations of
who controls “what” and “how” they perform their job. Nevertheless, the TRP was
generally supportive of including this item in the full-scale questionnaire as a measure
of career-goal realization; and, given that this item was borrowed from the fourth
follow-up of NELS:88, F3JOBDESCRB will be retained as is for the full-scale
questionnaire to preserve potential trend analyses.
• F3MLTCURBRCH—This item asks about the current military branch in which a
respondent serves, if he or she indicated service in multiple branches. With only four
response options, respondents took an average of 80 seconds to answer. However,
this question only applied to two respondents, so the lengthy response is likely
attributable to a single case (although more respondents would be expected to answer
this item in the full-scale sample). After discussing this item at the post–field test
TRP meeting, this item was deemed to be of relatively lesser importance and will not
be included in the full-scale instrument.
• F3RFUND—Asking about retirement savings plans, this item had five response
options and averaged 24 seconds. The question wording may contribute to the lengthy
time, given the long question stem that distinguishes between employer-based and
individual retirement plans and lists seven specific types of plans. Additionally, the
response options are not aligned with the question’s emphasis on the distinction
between employer and individual plans. Instead, the response options distinguish
personal plans from spouse or partner plans and ask about different combinations of
personal and spouse/partner planning—including introducing the term “joint plan” in
the second response option before it has been defined in the third response option.
Furthermore, a response option appears to be missing (spouse/partner has own plan as
well as joint plan with respondent). To address the problems associated with the field
test response options, this question will be recast in the full-scale instrument (when
administered to respondents who are currently married or living with a significant
other) as a series of three more-appropriately ordered yes/no questions: “Do you and
your [spouse/partner] have any of these plans jointly?”; “Do you have any of these
plans on your own separate from your [spouse/partner]?”; and “Does your
[spouse/partner] have any of these plans separate from you?” Full-scale survey
respondents who are not currently married or living with a significant other will
simply be asked, “Do you have any savings in these types of plans?”
• F3DONATE—Asking about the frequency of voluntary contributions, this item
averaged 20 seconds despite five response options presented as a 1–5 scale (“never”
to “always”). The question is unclear about the inclusion of religious organizations
and donations of goods. Also, the never/always scale is not a concrete time span (i.e.,
annually, monthly, or weekly) and has no connection to financial amounts, which is a
common metric of financial contributions. This item (borrowed from the UniLOA
[University Learning Outcomes Assessment]) was discussed at the post–field test
TRP meeting, and was ultimately retained for the full-scale instrument because of (1)
the promising distribution of responses it produced during the field test; (2) the strong
psychometric properties of UniLOA items; (3) its relative brevity; and (4) the lack of
another suitable well-tested item to use in its place.
45
Chapter 4. Field Test Questionnaire Timing and Data Quality
11
There are reasonable exceptions, particularly in a longitudinal study. If one is looking for the change of some phenomenon
over time, it is appropriate to repeat a prior question even when most respondents will not be expected to report it. For all
practical purposes within ELS:2002/12, however, nonzero variability is a legitimate minimum standard.
46
Chapter 4. Field Test Questionnaire Timing and Data Quality
Table 18. Low-frequency responses to ELS:2002 third follow-up field test questionnaire items:
2011
47
Chapter 4. Field Test Questionnaire Timing and Data Quality
Ta b le 18. Lo w-fre q u e n c y re s p o n s e s to ELS :2002 th ird fo llo w-u p fie ld te s t q u e s tio n n a ire ite m s :
2011—c o n tin u e d
48
Chapter 4. Field Test Questionnaire Timing and Data Quality
Several items seem to add little to the questionnaire. A question about why a respondent
did not complete a postsecondary degree (F2WHYNOCRED13—because of school or program
closure or lost accreditation) had very few positive responses. This item will not be included in
the full-scale questionnaire because it does not add sufficient information to the already-long list
of other reasons (F3WHYNOCRED1-12) that the questionnaire asks. F3PARHOME, asking
about living in parent’s home, had one low-frequency response that is a coding error.
Nine of the low-frequency items ask about income received from various sources by the
respondent or the respondent’s spouse (e.g., F3INCOMSIN6 and the F3INCOMMAR_R
series—the F3INCOMMAR_R series is administered to respondents who are married or
partnered, while the F3INCOMSIN series is administered to all other respondents). Income
received from Social Security and from veterans benefits, both of which are expected to be rare
at this age range, are uncommon in the field test responses, as is income from child support and
unspecified other income.
The F3INCOMMAR_R series has an additional problem; that is, it not only separately
asks if the respondent or the respondent’s spouse/partner received income from each source, but
also asks whether “neither” the respondent nor spouse/partner received income from each source
(not all of the F3INCOMMAR_R items are shown in table 2 because frequencies were not low
for some of the items). This involves double-questioning on the same issue, because a respondent
who failed to affirmatively check, for example, either the “respondent” or “spouse/partner” box
on the form is implying that neither received such income. As a result of these difficulties, the
parallel sets of field test items F3INCOMSIN and F3INCOMMAR will not be included in the
full-scale instrument.
The final two sets of questions with low-frequency responses involve stressful life events
and the importance of specific goals or values. Both series of items are part of longer series in
which broader distributions of responses occur. In addition, for the life event items,
distinguishing “happened once” from “happened more than once” serves to route respondents to
a subsequent set of questions about, respectively, when the single event happened or when the
first and then the most recent event happened. Given this integration with other items, no
changes were made to the field test questionnaire when producing the full-scale questionnaire.
The values questions are likewise embedded in a series involving broader response distributions,
and are also part of a series that has been asked in prior rounds of ELS:2002, as well as in the
National Education Longitudinal Study of 1988. Although 3 of the 14 items in this series will be
dropped from the full-scale instrument in hopes of shortening the overall average interview
length, the remaining 11 “value” items will be preserved to maintain longitudinal and cross-
cohort comparability.
As a final note for this section, even though it does not result in low response rates, it
should be noted that the field test response categories for a question about highest degree ever
expected (F3EDEXP) do not match those for a later question about the degree needed for the job
49
Chapter 4. Field Test Questionnaire Timing and Data Quality
the respondent expects to hold at age 30 (F3OCC30ED). This not only precludes direct
comparison between those responses—an obvious analytical use of the data—but may even
confuse respondents to the questionnaire. Thus, these response categories will be aligned in their
corresponding full-scale versions.
50
Chapter 4. Field Test Questionnaire Timing and Data Quality
question. For both, the question stem was “Has any of your student loan debt been…”, and
individual items were “paid off by you,” “paid off by your family,” and “forgiven by a loan
forgiveness program?” For the first item series (F3LOANPAID), response options were “none of
the debt,” “some of the debt,” and “all of the debt.” This item was presented to respondents who
owed nothing on their loans. For the second item series (F3LOANPAIDALT), response options
were “none of the debt” and “some of the debt” only. This item was presented to respondents
who still owed money on their loans. A greater percentage of respondents indicated that they had
difficulty with the first item (10 percent) than the second item (6 percent). The “please specify”
text indicated that a number of respondents with loan amounts (who therefore received
F3LOANPAIDALT) said they were still enrolled or were within the grace period before
repayment was required, and therefore did not have to pay anything. However, the questionnaire
routes respondents to F3LOANPAIDALT even if they indicated on a prior question
(F3LOANPAY) that they pay nothing each month. Because of the issues described here, and
because several questions on debt are included later in the instrument, neither version of these
items (F3LOANPAID or F3LOANPAIDALT) will be retained in the full-scale instrument.
The other four methodological probes covered job and family issues. For the question
about professional certification and state and industry licensure (F3LICENSE), 3 percent of field
test respondents indicated a problem answering. The “please specify” text indicated that some
respondents had difficulty understanding what was meant by a “professional certification” or a
“state or industry license” (for example, one asked if a teaching license was included). We
recommend adding examples to this question, such as teaching license or computer certification.
For the job training question (F3JTRAIN), 3 percent of respondents indicated a problem
answering. The “please specify” responses did not indicate any consistent issue, but this item
was used in the reliability reinterview discussed in section 4.3 below (the discussion there
critiques the lengthy, confusing wording of this item). For the question about marital or partner
status when the respondent’s child was born (F3BIOCHPART), 5 percent indicated a problem
answering. There were only three responses supplied to the “please specify” request, and they
did not reveal a consistent problem. One respondent indicated some confusion about the meaning
of “partnered with,” because that person was not married but was living with the child’s other
parent. For the full-scale instrument, “partnered with” will be removed from the question
wording because existing research indicates not being married (as opposed to not being married
or partnered) is the primary distinction in determining whether having a child might interfere
with other education/employment-related activities. For the question about a parent living with
the respondent (F3PARHOME), fewer than 2 percent of field test respondents indicated a
problem, and only one vague response to “please specify” was provided; therefore, no changes to
this item will be made for the full-scale instrument.
51
Chapter 4. Field Test Questionnaire Timing and Data Quality
52
Chapter 4. Field Test Questionnaire Timing and Data Quality
Categorical Items. Table 20 presents the categorical items with response options of
yes/no or other two-category responses. It shows percentage agreement; Cramer’s V, a measure
of the strength of association, which ranges from 0 to 1, with 1 indicating perfect association;
and the statistical significance of the association between the original interview and reinterview,
where cell sizes were sufficiently large to produce valid chi-squared statistics.
Table 20. Interview-reinterview agreement for categorical items on the ELS:2002 third follow-up
field test reliability reinterview: 2011
53
Chapter 4. Field Test Questionnaire Timing and Data Quality
Ta b le 20. In te rvie w-re in te rvie w a g re e m e n t fo r c a te g o ric a l ite m s o n th e ELS :2002 th ird fo llo w-u p
fie ld te s t re lia b ility re in te rvie w: 2011—c o n tin u e d
For the categorical variables, the percentage agreement ranged from 67 percent to 98
percent, with 20 of the 25 items having matched responses in at least 85 percent of the cases.
Five items had percentage agreement between 55 percent and 84 percent, and none had a
percentage agreement below 55.
Of the five items with moderate reliability, the first is a question about whether work and
school activities at the time of the interview are the same as they were in the last week of June
2011, which is designed (along with follow-up questions) to anchor all of the respondent’s
activities at the same time point for comparability across cases during analysis. This item had a
respectable percent agreement of 72 percent, but a low Cramer’s V of 0.30 (the lowest of all
reinterview items). Although a simple yes/no question, the anchoring of this question at a
specific week, even a recent week, may make recall difficult for some respondents.
Alternatively, the number of different items involved in “work or school activities” asked about
in this question makes it quite possible that one or more of the activities that the respondent is
comparing his or her past to could have changed. For this reason, and the importance of
anchoring activities to a specific time point, no changes to this item will be made for the full-
scale instrument.
The next moderate-reliability item asks “In the last 12 months, have you participated in a
formal training program offered by an employer or a union that helped you learn or improve the
skills needed to do your job?” Although the percent agreement is also respectable at 73 percent,
and there is a statistically significant association between the original and reinterview responses,
Cramer’s V is relatively low at 0.45. As another simple factual question, we might expect
percent agreement to be higher. However, the 12-month reference window used by the
reinterview is by definition slightly different from the 12-month reference window used in the
original interview and may account for some of the disagreement. We therefore implemented no
changes.
The remaining three items with moderate reliability are part of a series which asks about
certain events occurring since January 2005 (about a 5.5-year time span). The items in this series
that have high reliability ask about events that the respondent experienced (such as becoming
seriously ill) or highly significant events involving parents or guardians (such as death or
divorce). In contrast, two of the three moderate reliability items ask about less-significant events
54
Chapter 4. Field Test Questionnaire Timing and Data Quality
that others experience: close relative or friend died (percent agreement = 76, Cramer’s V = 0.59)
and family member became seriously ill or disabled (percent agreement = 67, Cramer’s
V = 0.50). The ambiguity of “close relative or friend” and, particularly, of “family member,”
may be problematic. The third item in this series with moderate reliability asks whether
respondent lost a job (percent agreement = 84, Cramer’s V = 0.61), and does not have a clear
explanation for its somewhat lower percent agreement, given that it involves both the respondent
and is likely a very memorable event. Thus, given that the percent agreement is not particularly
low, no changes to this item will be made for the full-scale instrument.
Continuous Items. Because percent agreement is not as meaningful when the number of
possible responses is large, results are presented separately for continuous variables (all
addressing household membership) in table 21. Table 21 reports percentage agreement; a
measure of association (Pearson’s correlation coefficient r, which ranges from -1 to 1); and the
statistical significance of an F statistic from a bivariate regression between the original and
reinterviewed responses.
Table 21. Interview-reinterview agreement for continuous items on the ELS:2002 third follow-up
field test reliability reinterview: 2011
Pearson’s Statistical
Percent correlation significance
Item agreement coefficient (r) (F test)
Number of spouses living with respondent 100.0 1.00 †
Number of marriage-like partners living with respondent 100.0 1.00 †
Number of mothers or female guardians living with respondent 84.9 0.69 F < 0.0001
Number of fathers or male guardians living with respondent 89.7 0.79 F < 0.0001
Number of friends or roommates living with respondent 80.8 0.87 F < 0.0001
Number of brothers or sisters living with respondent 92.3 0.89 F < 0.0001
Number of children living with respondent 96.3 0.94 F < 0.0001
Number of others living with respondent 100.0 1.00 †
† Not applicable
SOURCE: U.S. Department of Education, Institute of Education Sciences, National Center for Education Statistics,
Education Longitudinal Study of 2002 Third Follow-up Field Test (ELS:2002/12).
Table 21 shows the eight continuous items that were asked in reinterviews. Percentage
agreement ranges from 81 percent to 100 percent, with six of the eight items showing agreement
above 85 percent. The two moderate-agreement items are attached to questions about the number
of mothers or female guardians living with the respondent (just under 85 percent agreement, with
Pearson’s r at 0.69) and the number of friends or roommates living with the respondent (81
percent agreement, Pearson’s r 0.87). The F tests of bivariate regressions between the original
and reinterview responses for both items were statistically significant. Given these relatively
good results, and no obvious problems in question construction, no changes to either item will be
made for the full-scale instrument.
55
Chapter 4. Field Test Questionnaire Timing and Data Quality
12
Social cognitive career theory (for an influential statement of this approach, see Lent et al. [1994]) builds on the work of
Bandura and others in social learning theory. Self-efficacy beliefs are seen as interacting with outcome expectations and goals.
The theory is frequently applied in the context of career choice and development, including the explanation of subgroup
differences.
13
Nunally and Bernstein (1994) suggest that an internal consistency coefficient of 0.80 is sufficient in research contexts where
the purpose of scaling is group-level comparison and research. Where clinical decisions or individual judgments are the main
focus, reliabilities of 0.90 and above may be required.
56
Chapter 4. Field Test Questionnaire Timing and Data Quality
on the results provided in the appendix, no items were recommended for deletion from any scale,
on the basis of this initial analysis.
Table 22. Reliability analysis of questionnaire scales: Standardized alpha
Because, however, the six social cognitive career theory scales (i.e., all but the college
satisfaction scale) were newly constructed specifically for ELS:2002/12, additional analyses
were conducted to ensure that they would prove efficacious. For further evidence that items and
constructs were functioning well psychometrically, exploratory and confirmatory factor analyses
were undertaken.
Theoretically, these six item sets should have corresponded to six correlated but
relatively distinct constructs: occupational self-efficacy, interest, supports, benefits, satisfaction,
and commitment. As already noted, all item sets produced acceptable internal consistency
reliability estimates. However, when all items were pooled, three of these item sets (interest,
benefits [outcome expectations], and satisfaction) tended to load on the same underlying factor.
That is, rather than distinguishing between them, participants tended to respond to them as if
they were all reflecting a single, larger construct of positive perceptions of one’s job. The three
other item sets (support, satisfaction, commitment [persistence intention]) tended to load on
correlated but reasonably distinct factors.
The distribution of self-efficacy was highly skewed and kurtotic, with most participants
feeling highly confident about their ability to perform their job duties and few feeling insecure.
This non-normal distribution likely attenuated the correlations between self-efficacy and the
other constructs (the correlations were much lower than those typically found in the literature).
Support scores were also quite skewed and kurtotic, although support did still yield plausible
correlations with most other constructs.
The three social cognitive career theory scales that seem most justifiable to include are
support, satisfaction, and commitment. All three lend themselves to trait-like measurement (i.e.,
they do not need to be linked to specific types of jobs to be assessed adequately). The items
which comprise the remaining SCCT scales (self-efficacy, interest, and benefits [outcome
expectations]) will not be included in the full-scale instrument.
57
Chapter 4. Field Test Questionnaire Timing and Data Quality
58
Chapter 5.
Survey Control Systems and
Data Processing
5.1 System Design, Development, and Testing
Systems and processes used in the Education Longitudinal Study of 2002 (ELS:2002)
third follow-up field test were designed and developed to test and identify areas of improvement
in preparation for the full-scale study. The smaller scope of the field test provides opportunities
to test systems and processes which often can provide greater efficiencies in the full-scale study.
The following are the major systems that were used for the field test:
• Integrated Management System (IMS) (a comprehensive tool used to exchange files
between RTI and the National Center for Education Statistics (NCES), and to provide
access to a centralized repository of data collection reports);
• Survey Control System (SCS) (the central repository of the status of each activity for
each case in the study);
• Hatteras Survey Editor (a web-based application used to develop the ELS:2002 third
follow-up instrument);
• Web Survey (a web-based application used to administer the ELS:2002 third follow-
up instrument);
• Computer-assisted telephone interview Case Management System (CATI-CMS) (a
system that assigns cases to be called and provides telephone interviewers with the
appropriate screens and scripts to be used during the contacting and locating phase of
CATI); and
• ELS:2002 sample member website (public website hosted at NCES and used to
disseminate information, collect sample data, and administer the ELS:2002 survey).
Each system is described in detail in this chapter.
Systems were developed using the full system development life cycle process. Each
system makes necessary safeguards to handle personally identifying information (PII). Systems
such as IMS, Hatteras, and CATI-CMS are standard RTI systems used successfully on other
NCES studies (e.g., High School Longitudinal Study of 2009, Baccalaureate and Beyond), and
were developed using the latest software tools such as Microsoft.NET and Microsoft SQL Server
database management system.
Processing of PII by all systems was developed in accordance with the Federal
Information Processing Standards (FIPS) moderate security standard. Movement of the data
59
Chapter 5. Survey Control Systems and Data Processing
containing PII between the locations was handled in accordance with security requirements. In
compliance with the FIPS 140.2 standards, data were encrypted when moved between locations,
and decrypted once successfully reaching the destination. In addition to security, these automated
systems were developed to handle the need of moving data and files efficiently.
60
Chapter 5. Survey Control Systems and Data Processing
layout or behavior, programmers entered custom programming code (HTML, Javascript, and
C#.NET script) into the Hatteras custom code interface. This code was stored in the SQL Server
database along with the instrument specifications for compilation by the survey execution
engine.
61
Chapter 5. Survey Control Systems and Data Processing
62
Chapter 6.
Summary of Recommendations
for the Full-Scale Study
Methodologies and systems to accommodate the standard features of the Education
Longitudinal Study of 2002 (ELS:2002) third follow-up—a mixed mode (web, computer-
assisted telephone interview [CATI], and in the full-scale study, computer-assisted personal
interview [CAPI]) survey of a dispersed youth sample—have been thoroughly tested in the prior
study round, as well as in related longitudinal surveys of postsecondary populations, such as
Beginning Postsecondary Students and Baccalaureate and Beyond. The function of the field test
was more to confirm the efficacy of these standard locating practices and data collection methods
than to break new ground that deviated from them.
Nonetheless, despite the familiar nature of the task, ELS:2002 third follow-up represents
in many ways an extreme case of data collection difficulty, hence its own unique challenge. A
sample of the modal age of 26 is highly mobile. Unlike any other current National Center for
Education Statistics (NCES) study, there is a gap of a full 6 years between the second and third
follow-up rounds of ELS:2002. Moreover, the ELS:2002 sample is not an “elite” sample such as
baccalaureate recipients, but includes the full sociodemographic range of the young adult
population, including difficult-to-survey groups such as high school dropouts. Panel maintenance
activities (chiefly address updates) between the rounds do help to mitigate the negative effects of
this long gap in the interview schedule, but cannot take the place of a full-fledged interview
contact. A further challenge is to obtain an extremely high (90 percent or above) response rate.
Below, recommendations are provided for both pre-data collection (e.g., locating) and
data collection procedures, and the general types of modifications that were applied in the
questionnaire revision for the full-scale study are summarized. All recommendations that are
made in this chapter will be implemented.
63
Chapter 6. Summary of Recommendations for the Full-Scale Study
64
Chapter 6. Summary of Recommendations for the Full-Scale Study
interviews. Revisions and deletions were also informed by Technical Review Panel
recommendations, and driven by reexamination of policy relevance or theoretical cogency of
prime constructs, in deliberations between RTI and NCES staff.
65
Chapter 6. Summary of Recommendations for the Full-Scale Study
66
References
Cronbach, L.J. (1951). “Coefficient Alpha and the Internal Structure of Tests.” Psychometrika,
16: 297-334.
Curtin, T.R., Ingels, S.J., Wu, S., and Heuer, R. (2002). NELS:88 Base Year to Fourth Follow-up
Data File User’s Manual (NCES 2002-323). U.S. Department of Education, National Center
for Education Statistics. Washington, DC: U.S. Government Printing Office.
Groves, R. M., & Heeringa, S. (2006). Responsive design for household surveys: tools for
actively controlling survey errors and costs. Journal of the Royal Statistical Society
Series A: Statistics in Society, 169(Part 3), 439-457.
Ingels, S.J., Pratt, D.J., Wilson, D., Burns, L.J., Currivan, D., Rogers, J.E., and Hubbard-
Bednasz, S. (2007). Education Longitudinal Study of 2002 (ELS:2002) Base-Year to Second
Follow-up Data File Documentation (NCES 2008-347). U.S. Department of Education,
Institute of Education Sciences. Washington, DC: National Center for Education Statistics.
Ingels, S.J., Pratt, D.J. Herget, D.R., Burns, L.J., Dever, J.A., Ottem, R., Rogers, J.E., Jin, Y., and
Leinwand, S. (2011). High School Longitudinal Study of 2009 (HSLS:09). Base-Year Data
File Documentation. (NCES 2011-328). U.S. Department of Education, Institute of
Education Sciences. Washington, DC: National Center for Education Statistics.
Lent, R.W., Brown, S.D. & Hackett, C. (1994), “Toward a Unifying Social Cognitive Theory of
Career and Academic Interest, Choice, and Performance.” Journal of Vocational Behaviour,
45, 79-122.
Nunally, J.C., and Bernstein, I.H. (1994). Psychometric Theory. 3rd ed. New York: McGraw-
Hill.
Peytchev, A., Baxter, R. K., & Carley-Baxter, L. R. (2009). Not All Survey Effort is Equal:
Reduction of Nonresponse Bias and Nonresponse Error. Public Opinion Quarterly, 73(4),
785-806.
Peytchev, A., Riley, S., Rosen, J.A., Murphy, J.J. and Lindblad, M. (2010). Reduction of
Nonresponse Bias in Surveys through Case prioritization. Survey Research Methods, 4(1),
21-29.
Research Triangle Institute (2008). SUDAAN Language Manual, Release 10.0. Research Triangle
Park, NC: Research Triangle Institute.
Riccobono, J.A., Henderson, L.B., Burkheimer, G.J., Place, C., and Levinsohn, J.R. (1981).
National Longitudinal Study: Base Year (1972) through Fourth Follow-up (1979) Data File
User’s Manual. U.S. Department of Education, National Center for Education Statistics.
Washington, DC: U.S. Government Printing Office.
Schouten, B., Cobben, F., & Bethlehem, J. (2009). Indicators for the representativeness of survey
response. Survey Methodology, 35(1), 101-114.
67
References
Tourangeau, R., Sebring, P., Campbell, B., Glusberg, M., Spencer, B.D., and Singleton, M.
(1987). The National Longitudinal Study of the High School Class of 1972 (NLS:72) Fifth
Follow-up (1986) Data File User’s Manual (NCES 87-406c). U.S. Department of Education,
National Center for Education Statistics. Washington, DC: U.S. Government Printing Office.
Zahs, D., Pedlow, S., Morrissey, M., Marnell, P., and Nichols, B. (1995). High School and
Beyond Fourth Follow-up Methodology Report (NCES 95-426). U.S. Department of
Education, National Center for Education Statistics. Washington, DC: U.S. Government
Printing Office.
68