NICE Real-World Evidence Framework-1
NICE Real-World Evidence Framework-1
NICE Real-World Evidence Framework-1
framework
Corporate document
Published: 23 June 2022
www.nice.org.uk/corporate/ecd9
Contents
Overview .................................................................................................................................. 4
Background ......................................................................................................................................... 10
Introduction......................................................................................................................................... 30
Introduction......................................................................................................................................... 44
Introduction......................................................................................................................................... 58
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 2 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Analysis ............................................................................................................................................... 71
Reporting............................................................................................................................................. 81
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 3 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Overview
Key messages
• Real-world data can improve our understanding of health and social care delivery,
patient health and experiences, and the effects of interventions on patient and system
outcomes in routine settings.
• As described in the NICE strategy 2021 to 2026 we want to use real-world data to
resolve gaps in knowledge and drive forward access to innovations for patients.
- identifying when real-world data can be used to reduce uncertainties and improve
guidance
- clearly describing best practices for planning, conducting and reporting real-world
evidence studies to improve the quality and transparency of evidence.
• The framework aims to improve the quality of real-world evidence informing our
guidance. It does not set minimum acceptable standards for the quality of evidence.
Users should refer to relevant NICE manuals for further information on how
recommendations are made (see the section on NICE guidance).
- ensure data is of good provenance, relevant and of sufficient quality to answer the
research question
- generate evidence in a transparent way and with integrity from study planning
through to study conduct and reporting
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 4 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
- use analytical methods that minimise the risk of bias and characterise uncertainty.
• The framework provides in-depth guidance and tools to support the implementation of
these core principles across different uses of real-world evidence. It is structured as
follows:
• The framework is a living framework that will be updated periodically to reflect user
feedback, learnings from implementation including exemplar case studies,
developments in real-world evidence methodology, and to extend its scope to include
additional guidance on priority topics.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 5 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Table 1
Summary of key considerations in planning, conducting and reporting real-world
evidence studies
Stage of
evidence Key considerations
generation
• Plan the study in advance and make protocols (including a data analysis
plan) publicly available
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 6 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Stage of
evidence Key considerations
generation
- data curation
• Real-world data is already widely used to inform NICE guidance to, for example:
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 7 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
• Real-world data that represents the population of interest is NICE's preferred source
of evidence for most of these applications. Such data is regularly used for these
purposes in NICE guidance, but its use could be more commonplace, especially of
routinely collected data.
• Randomised controlled trials are the preferred source of evidence on the effects of
interventions. Randomisation ensures that any differences in baseline characteristics
between groups are because of chance. Blinding (if applied) prevents knowledge of
treatment allocation from influencing behaviours. However, randomised trials are
sometimes unavailable or are not directly relevant to decisions about patient care in
the NHS.
- funding is not available for a trial (for example, when the intervention is already
used in routine practice).
• Even if randomised evidence is available, it may not be sufficient for decision making
in the NHS for several reasons including:
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 8 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
- there are major differences in patient behaviours, care pathways or settings that
differ from implementation in routine practice
- follow up is limited
• Non-randomised studies are already widely used to estimate the effects of medical
devices and procedures and public health interventions, for which trials are less
common. They are becoming more widely used in initial assessments of medicines, as
more are granted regulatory approval based on uncontrolled single-arm trials. Finally,
the increased focus on the lifecycle evaluation of technologies and lived experiences
of patients relies on non-randomised studies after initial approvals. The most common
non-randomised studies using real-world data to assess comparative effects are
observational cohort studies and single-arm trials with real-world external control.
• Real-world data could be used more routinely to fill evidence gaps and speed up
patient access. For this promise to be realised, real-world evidence studies must be
performed transparently and with integrity, use fit-for-purpose data, and address the
key risks of bias.
• We are communicating our view on best practices for the conduct of real-world
evidence studies to ensure they are generated transparently and are of good quality.
This is essential to improving trust in real-world evidence studies and their use in
decision making.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 9 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
As described in the NICE strategy 2021 to 2026, we want to use real-world data to resolve
gaps in knowledge and drive forward access to innovations for patients.
We developed the real-world evidence framework to help deliver on this ambition. It does
this by:
• identifying when real-world data can be used to reduce uncertainties and improve
guidance
• clearly describing best practices for planning, conducting and reporting real-world
evidence studies to improve the quality and transparency of evidence.
The framework aims to improve the quality of real-world evidence informing our guidance.
It does not set minimum acceptable standards for the quality of evidence. Users should
refer to relevant NICE manuals for further information on how recommendations are made
(see the section on uses of real-world data in NICE guidance).
The framework is mainly targeted at those developing evidence to inform NICE guidance.
It is also relevant to patients, those collecting data, and reviewers of evidence.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 10 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
sources are observational (or non-interventional), that is, any interventions (or exposures)
are not determined by a study protocol. Instead, medical interventions are decided by
patients and healthcare professionals. And in public health or social care, interventions
may be determined by individual behaviours, environmental exposures or policy makers.
Some interventional studies, such as pragmatic clinical trials, can also produce real-world
evidence. Such trials may also make use of real-world data sources to design trials, recruit
participants or collect outcome data. For more information, see the UK Medicines and
Healthcare products Regulatory Agency's (MHRA) guideline on randomised controlled
trials using real-world data to support regulatory decisions.
Real-world data can be quantitative or qualitative. Common data types include patient
demographics, health behaviours, medical history, clinical outcomes (including patient-
reported outcomes), patient or user experiences, resource use, costs, omics, laboratory
measurements, imaging, free text, test results and patient-generated data. We consider
both national data collections and international data when making recommendations.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 11 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Table 2
Common sources of real-world data
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 12 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 13 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 14 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
NICE guidance
NICE has several guidance products that use the best available evidence to develop
recommendations that guide decisions in health, public health and social care, including:
• guidelines for clinical, social care and public health topics, which offer advisory
guidance to health and social care professionals
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 15 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Guidelines are developed internally by NICE. Technology evaluations are usually informed
by company submissions but may also use evidence submitted by manufacturers or other
stakeholders or research commissioned from independent academic centres.
The processes and methods for technology evaluations differ across NICE's programmes.
The Technology Appraisal Programme evaluates mostly medicines (including highly
specialised technologies) but can also include medical devices and diagnostics. The
Technology Appraisal and Diagnostic Guidance Programmes both consider the cost
effectiveness of medical technologies. The Medical Technologies Evaluation Programme
evaluates medical technologies including medical devices, digital health technologies and
diagnostics that are expected to be cost-saving or cost-neutral and uses cost-
consequence analysis considering patient and system outcomes. The Interventional
Procedures Programme evaluates the efficacy and safety of interventional procedures
without analysis of cost.
When NICE recommends a treatment 'as an option' through its Technology Appraisal
Programme, the NHS must make sure it is available within 3 months (unless otherwise
specified) of its date of publication. If a technology is potentially cost effective but there is
substantial and resolvable uncertainty about its value, it can be recommended for use in a
managed access agreement. After a specified period of collecting real-world data, the
technology is reassessed through the Technology Appraisal Programme. Selected devices,
diagnostic or digital technologies that are recommended in NICE guidance and are likely to
be affordable and produce cost savings within 3 years of adoption can be funded through
NHS England's MedTech funding mandate.
Methods and process manuals have been developed for different NICE programmes. Users
of this framework should consult these manuals as appropriate:
• Developing NICE guidelines: the manual explains the processes and methods used to
develop and update NICE guidelines, the guidance that NICE develops covering topics
across clinical care, social care and public health.
• NICE's health technology evaluations manual describes the methods and processes
for developing health technology evaluation, including for the Diagnostics Assessment
Programme, the Medical Technologies Evaluation Programme, the Highly Specialised
Technologies Programme, and the Technology Appraisal Programme.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 16 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
NICE's evidence standards framework for digital health technologies sets out what good
levels of evidence for digital health technologies look like. It is aimed at innovators and
commissioners of digital health technologies.
Real-world data is already used across NICE programmes to generate different types of
evidence, especially for questions that are not about the effects of interventions.
Examples from previous NICE guidance include:
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 17 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
- resource use and costs: NICE medical technologies guidance on HeartFlow FFRCT
for estimating fractional flow reserve from coronary CT angiography used cost
data on coronary revascularisation from NHS reference costs
• developing and validating digital health technologies including prognostic models: see
the NICE evidence standards framework for digital health technologies for further
information
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 18 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
adjuvant chemotherapy decisions in early breast cancer reported results from several
prospective observational studies.
Real-world data can also be used to assess the applicability of trial results to patients in
the NHS or even to estimate intervention effects (for further information, see the section
on estimating intervention effects using real-world data).
While real-world evidence is already widely used for many of these types of evidence
(Leahy et al. 2020, Makady et al. 2018), its use could be more commonplace. When data is
representative of the target population and of sufficient quality it may be the preferred
source of data. Background event rates or natural history data from trials may sometimes
overestimate or underestimate event rates in the target population because of selective
recruitment (Bloudek et al. 2021). In some cases, there may be value in performing studies
using routinely collected data rather than relying on published evidence that has lower
applicability to the research question.
However, randomised controlled trials are not always available or may not be sufficient to
address the research question of interest.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 19 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Randomised controlled trials may be especially difficult to do for rare diseases, innovative
and complex technologies, or in certain populations.
Similarly, high-quality randomised controlled trials can be challenging for medical devices
and interventional procedures because of the difficulty of blinding, the importance of
learning effects, changes to standard of care making the choice of comparator
challenging, changes to the characteristics of the technology over time that may impact
on performance, and limited research capacity or access to funding (Bernard et al. 2014).
Even if trials are available, they may not be directly applicable to the research question or
to routine care in the NHS because of:
• use of comparators that do not represent the standard of care in the NHS (including
placebo control)
• limited follow up
• learning effects (that is, the effect of an intervention changes over time as users
become more experienced)
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 20 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Some of these challenges, such as the use of comparators that do not represent the
standard of care in the NHS, can potentially be addressed through other approaches such
as network meta-analysis under certain assumptions about the comparability of the trials.
See the NICE Decision Support Unit report on sources and synthesis of evidence for
further information.
Real-world evidence can also be generated from randomised controlled trials that use
real-world data in their design or for measuring outcomes, such as pragmatic clinical trials.
Such trials may provide substantial value in combining the internal validity from
randomisation with the greater generalisability of data from routine practice. The UK
MHRA has published guidance on producing real-world evidence from randomised
controlled trials.
Real-world evidence
Real-world data can be used to contextualise randomised trials, to estimate effects of
interventions in the absence of trials, or to complement trials to answer a broader range of
questions about the impacts of interventions in routine settings.
Contextualisation
Contextualisation involves assessing whether the results from trials will translate well to
the target population in the NHS. While this is an important use of real-world data across
NICE programmes, NICE may require the collection of further data through managed
access arrangements for medicines that are potentially cost effective and if uncertainties
can be addressed through further data collection. This data is often used to understand
the relevance of trials to the NHS.
Real-world data has been used in NICE guidance to contextualise clinical trials including
for:
• differences in eligible population in the NHS, treatment pathways, care settings and
outcomes: NICE technology appraisal guidance on pegcetacoplan for treating
paroxysmal nocturnal haemoglobinuria used UK registry data to show that urinary
haemoglobin levels in UK practice were in line with the eligibility threshold for the
randomised controlled trial
• modelling the relationship between surrogate outcomes and final outcomes (including
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 21 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
• measuring the use of, and adherence to, interventions: NICE medical technologies
guidance on Sleepio to treat insomnia and insomnia symptoms used data on usage
collected from the app or website
NICE technology appraisal guidance on osimertinib for treating EGFR T790M mutation-
positive advanced non-small-cell lung cancer used data from the Systemic Anti-Cancer
Therapy (SACT) dataset to assess the relevance of results from the AURA3 trial to NHS
patients. In particular, SACT data was used to compare:
• overall survival
Estimation
Real-world data can be used to better understand the effects of an intervention over its
life cycle. The potential uses of real-world data for estimating effects of interventions
depend on the stage in their life cycle.
For new interventions (for example, those with recent marketing authorisation in the UK),
there will be limited real-world data on their use and outcomes in the NHS. The uses of
real-world data include:
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 22 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
• creating a comparator arm (that is, external control) to estimate effects against a
single-arm trial or to add to controls from a randomised controlled trial: NICE highly
specialised technologies guidance on metreleptin for treating lipodystrophy used a
natural history study to form an external control to a single-arm trial
• using data from early access to medicines schemes: NICE technology appraisal
guidance on berotralstat for preventing recurrent attacks of hereditary angioedema
included early access to medicines scheme data to reduce uncertainty around long-
term outcomes
Once medical technologies are used routinely or in pilot projects, the opportunities for
real-world data are greater and include:
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 23 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
• estimating effects on final outcomes of interest (rather than surrogate outcomes) and
over longer time periods
• incorporating into evidence synthesis, for example, informing priors, increasing power
or filling evidence gaps in a network meta-analysis (NICE Decision Support Unit report
on sources and synthesis of evidence, Sarri et al. 2020).
These studies have demonstrated that high-quality non-randomised studies can produce
valid estimates of relative treatment effects in many, but certainly not all, situations. There
are some common design principles that improve the likelihood of valid estimates
including:
• the use of active comparators (alternative interventions for the same or similar
indication, usually of the same modality) and
• comparing new users (or initiators) of interventions rather than those who have been
using an intervention for some time (prevalent users).
Validity may also depend on other factors including the characteristics of the disease, type
of outcome (objective clinical outcomes are preferred), the treatment landscape, and data
content and quality.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 24 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Therefore, concerns about the integrity and trustworthiness of the resulting evidence (for
example, resulting from data dredging or cherry-picking) need to be addressed. Concerns
about the legitimate use of data have been highlighted by the retraction of high-profile
studies about the effectiveness of repurposed medicines for treating COVID-19 from
prominent medical journals.
• registering the study protocol before implementing the study (see the Real-World
Evidence Transparency Initiative)
• reporting checklists or tools (see Enhancing the Quality and Transparency of Health
Research [EQUATOR] network)
• requiring author statements to confirm the integrity of data access and study conduct
(see learning from a retraction by the editors of the Lancet Group, 2020)
• providing access to data through secure data environments and maintaining audit
trails (see the Department of Health and Social Care's report on better, broader, safer:
using health data for research and analysis).
See guidance on planning, conducting and reporting real-world evidence studies in the
section on conduct of quantitative real-world evidence studies to generate real-world
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 25 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
evidence.
Other variables may be collected at an insufficiently granular level. For instance, a study
may need knowledge of a specific drug or medical device, but the data may include only
drug or device class. Similarly, a study may need to distinguish between haemorrhagic and
ischaemic strokes while a data source may contain data on all strokes without further
detail. Even if relevant items are collected with the needed granularity, the data may be
missing or inaccurate, which can cause information bias. In addition, there may be
variation in data-recording practices and quality across centres or individuals, and in the
quality management processes for different sources of data.
In addition to the availability of data on relevant study elements, the relevance of a given
data source to a research question may be affected by several factors. This includes the
representativeness of the study sample and similarities in treatment patterns and
healthcare delivery to routine care in the NHS, the timeliness of data, sample size and
length of follow up. The key questions are whether the data is sufficient to produce robust
estimates relevant to the decision problem and whether results are expected to translate
or generalise to the target population in the NHS.
Risk of bias
Studies using real-world data are at risk of bias from a number of sources, depending on
the use case. We describe key risks of bias that threaten validity in individual real-world
evidence studies below. Detailed descriptions of risks of bias in non-randomised studies
are available, such as the European Network of Centres for Pharmacoepidemiology and
Pharmacovigilance (ENCEPP) guide on methodological standards in
pharmacoepidemiology and chapter 25 in the Cochrane handbook for systematic reviews
of interventions.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 26 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Selection bias
In non-comparative studies, selection bias can occur if the people studied are not
representative of the target population. This might result from non-random sampling of the
source population, non-response to a questionnaire, or differences in behaviours and
outcomes of those who volunteer to be part of research studies.
• including prevalent users of a technology compared with non-users (users who had
already experienced the event or not tolerated the intervention would be excluded
from analysis)
• excluding a period of follow up in which the outcome cannot occur (known as immortal
time bias for survival outcomes)
• selection into the study based on a characteristic (for example, admission to hospital)
that is related to the intervention and outcome.
A common cause of selection bias at study exit is loss to follow up. Selection bias can also
be caused by excluding participants from analysis, such as those with missing data.
Information bias
Information bias may result from missing or inaccurate data on population eligibility
criteria, interventions or exposures, outcomes and covariates (as relevant). These
limitations may occur because of low data quality, care patterns or data collection
processes. They may also result from misspecification of the follow-up period.
The consequences of these issues depend on factors including the study type, whether
limitations vary across intervention groups, whether they are random or systematic (that
is, the missing data mechanism), the magnitude of the limitation and in which variables
they occur. One common cause of differential misclassification across groups is detection
bias. This occurs when the processes of care differ according to intervention status such
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 27 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
that outcomes are more likely to identified in 1 group than in another. See the section on
measurement error and misclassification for further information.
Confounding
Confounding occurs when there are common causes of the choice of intervention and the
outcome. This is expected to be common in healthcare because healthcare professionals
and patients make decisions about treatment initiation and continuation based on their
expectations of benefits and risks (known as confounding by indication or channelling
bias). Confounding bias may be intractable when comparing treatments with different
indications and across types of intervention (for example, interventional procedure
compared with drug treatment) and for studies of environmental exposures.
Bias may also arise because of inappropriate adjustment for covariates, for example, if a
study controls for covariates on the causal pathway (such as blood pressure in the effect
of anti-hypertensive medication on stroke), colliders (a variable influenced independently
by both the exposure and the outcome), or instruments (defined as a variable that is
associated with the exposure but unrelated with the outcome except through the
exposure).
External validity refers to how well the findings from the analytical sample apply to the
target population of interest. Study findings may be intended to be applied to a target
population from which the study sample was drawn ('generalisability'), or to another target
population, from which the study sample was not derived ('transportability').
Differences can occur between the study sample and target population for factors that
affect outcomes on the scale of estimation (for example, relative versus absolute effects).
These may include differences in patient or disease characteristics, healthcare settings,
staff experience, treatment types and clinical pathways. Further differences may result
from patient exclusions, drop out and data missingness in the analytical sample.
Methods to assess and adjust for some elements of external validity bias (those relating to
differences in patient characteristics in studies of comparative treatment effects) are
discussed in the section on addressing external validity bias.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 28 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Reverse causation (or protopathic bias) occurs when the intervention is a result of the
outcome or a symptom of the outcome. This is most problematic in conditions with long
latency periods such as several cancers. If present, this is a severe form of bias with major
implications for internal validity.
Biases may also result from the statistical analysis of data (for example, model
misspecification).
When assessing the body of literature on a research question there are further concerns
about publication bias because of non-reporting of real-world evidence studies, especially
if they show null results (Chan et al. 2014).
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 29 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Key messages
• Transparent and reproducible generation of real-world evidence is essential to
improve trust in the evidence and enable reviewers to critically appraise studies.
- Use analytical methods that minimise the risk of bias and characterise
uncertainty.
• The required level of evidence may depend on the application and various
contextual factors (see the section on considerations for the quality and
acceptability of real-world evidence). Users should refer to relevant NICE
manuals for further information on how recommendations are made.
Introduction
The following principles underpin the conduct of all real-world evidence studies:
• Ensure data is of good and known provenance, relevant and of sufficient quality to
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 30 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
• Generate evidence in a transparent way and with integrity from study planning
through to study conduct and reporting.
• Use analytical methods that minimise the risk of bias and characterise uncertainty.
The focus here is currently on real-world evidence studies of quantitative data. However,
several aspects of planning, conducting and reporting that we describe are also applicable
to qualitative studies. For aspects that differ, recognised methods of analysing,
synthesising, and presenting qualitative evidence should be applied.
Patients should be consulted throughout all aspects of study planning and conduct.
The contribution of a particular type of evidence will vary across applications depending
on the key drivers of uncertainty (that is, the evidence gap). For instance, in oncology,
assumptions around long-term outcomes such as overall survival and the applicability of
global trials to the NHS are often key (Morrell et al. 2018). In cost-effectiveness or cost-
comparison models, a number of different parameters could be important determinants of
cost effectiveness including event incidence, prevalence, natural history of disease, test
performance, costs or quality of life.
In general, non-randomised studies of clinical effects will need higher levels of rigour and
transparency than simple characterisation studies. Estimates of clinical effectiveness are
usually a key driver of recommendations and non-randomised studies can be at risk of
bias.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 31 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
The contextual factors that influence the acceptability of evidence include the level of
decision uncertainty, disease prevalence, impact on health inequalities and the possibility
of generating high-quality evidence. Users should refer to the relevant NICE manual for
further information on how recommendations are made (see the section on NICE
guidance).
Common challenges in the evaluation of medical devices and interventional studies using
real-world data include:
• limited integrated national data collections of medical device use and outcomes
• lack of granularity in many routinely collected data sources to identify specific devices
(and unique device identifiers) or procedures
These challenges are not universal and there are ongoing improvements to the availability
of high-quality data collections for medical devices and procedures including registries
and electronic health record systems. When possible the highest quality data should be
used.
• small sample sizes or the need to combine multiple sources of data with different data
models and data collection processes
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 32 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Study planning
• the target quantity that is to be estimated, for example, disease prevalence or average
effect of adhering to an intervention on overall survival.
Patient outcomes should reflect how a patient feels, functions or how long a patient lives.
This includes clinical outcomes such as survival as well as patient-reported outcomes.
Outcomes should be reliable and valid for the context of use. Choice of outcomes may be
supported by high-quality core outcome sets such as those listed in the Core Outcome
Measures in Effectiveness Trials (COMET) database.
The target quantity to be estimated should address the overall research question of
interest. For example, prevalence can reflect the quantity of a population who might need
access to services at a point in time. It represents a function of incidence and duration of
the condition; this may be useful for public health planning. Incidence captures rates of
events across different subgroups or those with different exposures but assumes a
constant rate across defined time intervals. Plausibility of the average rate should
therefore be considered.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 33 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Planning studies before conduct improves the quality of studies and can reduce the risk of
developers performing multiple analyses and selecting those producing the most
favourable results.
When planning the study, developers should consider any equality or diversity issues that
should be addressed in design, analysis, or interpretation of the study.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 34 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
• pre-specification of search strategy and defined criteria for selection and prioritisation
of datasets
• expert consultation to inform the search strategy and selection criteria and to highlight
known suitable datasets
• an online search and systematic literature search, and correspondence with lead
authors of relevant publications, when necessary, to gain information on access to and
suitability of potential data sources
• a direct search of data sources. In the UK this may be supported by registries of data
sources such as the Health Data Research UK Innovation Gateway
• a flow diagram outlining the total number of potential data sources identified and the
number excluded and reasons why (including for reasons of poor data suitability and
feasibility of access).
This approach can be informed by the considerations outlined in the section on assessing
data suitability or by following external guidance (Hall et al. 2012, Gatto et al. 2021).
The efforts made to identify data sources should be proportional to the overall importance
of the study. We also recognise that currently, registries of data sources are not always
available or may have limited metadata.
Data should be accessed and used in accordance with local law, governance
arrangements, codes of practice and requirements of the data controller. In the UK, the
Health Research Authority (HRA) provides guidance around research and use of data in
accordance with the UK Policy Framework for Health and Social Care Research.
Making early contact with data controllers and data processors is prudent to ensure data
are available when needed. Developers should ensure they have appropriate ethical (or
other) approval for the research study if needed. Developers should also create a plan for
sharing data with independent researchers and NICE collaborating centres, when
appropriate.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 35 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Data collection
For some use cases, primary data collection may be needed. Examples include:
• additional data collection to complement an existing data source, for example, adding
a quality-of-life questionnaire to a patient registry or performing a subsample
validation study
• a health survey.
When planning primary data collection, consider how to implement this collection in a
patient-centred manner while minimising the burden on patients and healthcare
professionals. Assess the feasibility of additional data collection before proceeding.
Sampling methods reduce the burden of data collection but can introduce selection bias.
Methods such as simple random sampling support external validity but tend to be feasible
only when the target population is small and homogenous. Alternative sampling techniques
are available, for example:
• stratified selection divides the target population into subgroups based on important
characteristics, such as prognostic factors or treatment effect modifiers, sampling
from each strata to ensure representation of all important subgroups
• balanced sampling for site selection considers important variation across sites in the
target population. Recruitment focuses on sufficient representation of sites within
each subgroup. Potential sites are ranked, allowing for quick identification of
replacements due to non-participation
Data collection should follow a predefined protocol and quality assurance processes
should be put in place to ensure the integrity and consistency of data collection. This also
applies to the extraction of structured information in retrospective chart reviews or when
using data science methods to derive structured data elements from already collected
data sources.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 36 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Data should be collected, stored, processed and deleted in accordance with the current
data protection laws with appropriate transparency information provided and safeguards
implemented. Approvals from the HRA or local organisation review and agreement as
appropriate should be in place. When appropriate, consent from participants should be
provided.
Please refer to Health Research Authority guidance on governance requirements and data
protection regulation for research and non-research use of healthcare data.
Study conduct
• sample size
• the structure of the data including data hierarchies or clustering (for example, patients
may be clustered within hospitals or data may be collected on a patient at multiple
timepoints)
Diagnostic checks should be used to assess the appropriateness of the selected statistical
model, if relevant. The appropriate checks will depend on the purpose of the study and
methods used.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 37 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Further information on the design and analysis of comparative effect studies is provided in
the methods section.
The risk of bias from using a particular data source will be informed by the information
considered during data suitability assessment.
Sensitivity analyses should reflect areas with the greatest concerns about risk of bias, or
when data curation or analytical decisions were made despite notable uncertainty.
Common considerations include:
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 38 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
If concerns about residual bias remain high and impact on the ability to make
recommendations, developers could consider using quantitative bias analysis. These
methods provide quantitative estimates of the impact of bias on study results (Lash et al.
2014). If external data on bias is incorporated, this should be identified in a transparent
and systematic way. For parameters of economic models including relative effects,
sensitivity analysis may consider the impact of bias on cost effectiveness as well as the
parameter value.
For further information on quality assurance please see the Office for National Statistic's
Quality Assurance of Code for Analysis and Research and the UK Government's Aqua
Book. This may be supported by using validated analytical platforms.
Study reporting
Reporting of studies should be sufficient to enable an independent researcher with access
to the data to reproduce the study, interpret the results, and fully understand its strengths
and limitations. Several reporting checklists identify key reporting items for:
• observational studies (see the EQUATOR network for reporting checklists by study
design, and the Strengthening the Reporting of Observational Studies in Epidemiology
[STROBE] guidelines)
Also, the STaRT-RWE tool has been developed to help the presentation of study data,
methods and results across use cases.
Below we describe key issues across data sources, data curation, methods and results
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 39 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
• Ethical (or other) approval for the research study or explain why such approval was not
necessary.
• A statement that the data was accessed and used in accordance with approvals and
information governance requirements.
• A description of how others can access the data (that is, a data sharing statement; for
an example, see the BMJ policy on data sharing).
For each individual study, developers should provide information on the software used to
perform analyses including the version system and any external packages used. Ideally,
analytical code should follow best practice in code structure, formatting and comments
and be publicly available (for example, through a code repository such as GitHub) or made
available on request to enable reproduction. When human abstraction or artificial
intelligence tools are used to construct variables from unstructured data, the methods and
processes used should be clearly described and their validity documented.
It may not be feasible to provide fully open code in all situations, for instance, when using
proprietary software or identifiable personal information. Developers should provide clear
information on the methods used and their validity. They should also seek to provide
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 40 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
access to the algorithms necessary to replicate and validate the analyses on request, with
necessary intellectual property protections in place.
Trust in the integrity of study conduct can be further improved by providing evidence that
the study was done appropriately, for example, by showing an audit trail of the analysis, if
this is feasible. This could demonstrate, for instance, that developers prepared analysis
and finalised protocols before the relevant results were revealed (MacCoun and Perlmutter
2015).
Reporting on methods
Below we describe key items that should be reported. This information should be
presented for all analyses including subgroup and sensitivity analyses. Methods should be
consistent with the study protocol, and deviations should be identified and justified.
Study design
Clear operational definitions should be given for all study variables and details of follow up,
if relevant. Study variables typically include patient eligibility criteria, interventions or
exposures, outcomes and covariates.
• the operational definition of the variable including code lists and algorithms when
possible
- how code lists or algorithms have been developed and, when possible, validated.
• the time period over which information for each variable is sought, defined in relation
to an index date (for example, 12 months before starting treatment)
• the grace period between observations that are assumed to represent continued use
of an intervention, if relevant.
For studies of comparative effects, the process by which potential confounders were
identified should be described alongside assumptions about the causal relationships
between study variables.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 41 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
• for interventions, assumptions about the minimum time between intervention and
outcome occurrence (latency period) and the likely duration of effects (exposure-
effect window).
In longitudinal studies, this information can be usefully summarised using a study design
diagram (Schneeweiss et al. 2019). The Reproducible Evidence: Practices to Enhance and
Achieve Transparency (REPEAT) initiative's project page hosts the paper and design
diagram templates.
Statistical methods
The statistical methods used should be clearly described. Information should be sufficient
to:
• understand what methods were used and why they were chosen
• understand how the analysis addresses different risks of bias including selection bias,
information bias and, if relevant, confounding (also see the section on quality
appraisal).
Reporting results
The following information should be presented in all studies:
• flow (or patient attrition) diagrams to report number of patients at each stage of the
study from raw data to the final analytical sample with reasons for exclusion
• patient characteristics (including missing data) and details of follow up including event
rates (or other distributional information on outcomes). For comparative studies these
should be presented across groups or levels of exposure and, if relevant, before and
after adjustment
Results should include central-point estimates, measures of precision and other relevant
distributional information if needed. Results should be presented for the main analysis and
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 42 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
all subgroup and sensitivity analyses. It should be clear which of these analyses were pre-
specified and which were not. For analyses that use adjustment to deal with confounding,
unadjusted results should also be presented.
Ensure that information in figures and tables cannot inadvertently identify patients. The
Office for National Statistics has guidance on maintaining confidentiality when
disseminating health statistics.
• following advice on writing understandable scientific material (see Gopen and Swann
1990, Greene 2013)
• labelling tables, graphs, and other non-text content clearly and explaining how to
interpret them.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 43 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Key messages
• Transparent reporting of data sources is essential to ensure trust in the data
source and understand its fitness for purpose to address the research question
- Reporting on data sources should cover the characteristics of the data, data
collection, coverage and governance
• Data fitness for purpose can be summarised by the data quality and relevance
- data quality relates to the completeness and accuracy of key study variables
• The acceptability of a given data source may depend on the application and
various contextual factors.
Introduction
Data used to inform NICE guidance should be reported transparently and be of good
provenance and fit for purpose in relation to the research question. The primary aims of
this section of the framework are to:
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 44 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
• provide clear guidance to evidence developers about expectations for clear and
transparent reporting on data and its fitness for purpose
This section should be read alongside the section on conduct of quantitative real-world
evidence studies.
We do not define minimum standards for data suitability beyond that the data should be
used in accordance with national laws and regulations concerning data protection and
information governance (see the section on reporting on data sources). The
considerations for data suitability are broadly applicable across different types of real-
world data and use cases but are largely focused on quantitative studies.
The acceptability of a data source will depend on the use case, and contextual factors
(see the section on considerations for the quality and acceptability of real-world evidence
studies). We recognise the need for trade-offs between different characteristics of data
sources including quality, size, clinical detail and locality. International data may be
appropriate for some questions in the absence of sufficient national data or when results
are expected to translate well between settings. We also recognise that there may be
challenges in identifying or collecting the highest quality evidence in some applications
including in rare diseases and for some medical devices and interventional procedures
(see the section on challenges in generating real-world evidence).
We do not request a particular format for the overall presentation of this information.
However, we have developed the Data Suitability Assessment Tool (DataSAT) to help the
consistent and structured presentation of data suitability at the point of assessment. The
concepts presented in the tool may also help developers choose between potential data
sources and in performing feasibility studies, but this is not its primary purpose. The tool
template and example applications are presented in appendix 1.
Data provenance
A full understanding of data provenance is essential to create trust in the use of data and
understand its fitness for purpose for a given application. In this section we present data
provenance considerations across 4 themes: basic characteristics of the data source, data
collection, coverage and governance.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 45 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Many real-world evidence studies will combine more than 1 data source, either by data
linkage or data pooling. Data linkage is often done to extend the information available on
individual patients, for example, by combining data from a prospective observational
cohort study with hospital discharge or mortality records, or patient-generated health
data. Data pooling is used to extend sample size or coverage of data and is common in
studies of rare diseases.
The reporting of data sources should primarily refer to the combined data used for the
research study. However, important differences between contributing datasets should be
clearly described.
Common data models are used to standardise the structure and sometimes coding
systems of different data sources. If data has been converted to a common data model,
the model and its version should be reported and full details of the mapping made
available, including any information loss. This information is essential to allow the study to
be reproduced.
Common data models can also support the use of federated data networks. These allow
individual patient health data to stay under the protection of partnering data holders who
will run standardised analyses before results are aggregated across datasets. Reporting of
federated data networks should be sufficient to understand the process of recruiting data
partners, feasibility assessments, and the common analytical framework used.
While complete and accurate data linkage will improve the quality and value of data,
imperfect linkage could exclude patient records or lead to data misclassification.
Therefore, when multiple sources of data are linked the following information should be
reported:
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 46 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Data collection
An understanding of a data source requires knowledge of the purpose and methods of
data collection.
• whether the data was routinely collected or collected for a specific research purpose
(or a combination)
• which types of data were collected, for example, clinical diagnoses, tests, procedures
and prescriptions
• how these were coded or recorded, for example, using ICD-10 codes for clinical
diagnoses, or free text data on cancer stage or biomarkers
• how data was collected, for example, directly by healthcare professionals in clinical
examinations, by remote monitoring or by administrative staff. If data is captured by a
digital health technology, the validity of the technology should be reported
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 47 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
- changes to coding systems (for example, the switch from Read v2 to SNOMED-CT
codes in UK primary care). Information on any mapping between coding systems
should be made available
• quality assurance processes for data collection that were in place (including training or
blinded review)
Any differences between data providers in how and what data were collected, and its
quality, should be described. This is especially important when data sources are pooled
from different systems and across countries.
Data coverage
Providing clear information on data coverage is essential, including the population, care
settings, geography and time. Such information has important implications for data
relevance that can inform later assessments of data suitability.
• the extent to which the data source captures the target population:
- if a data source does not include the full target population, the representativeness
of the data captured should be noted
- this should distinguish between care settings (for example, primary care
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 48 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
compared with secondary care), type of providers (for example, specialist medical
centres compared with general hospitals) and other factors when relevant
- if information was collected outside of the health or social care system, this
should be described (for instance, remote monitoring of activities of daily living).
• the geographical coverage of the data including countries and regions, if relevant
Data governance
Information about data governance is important for understanding the maturity of data and
its reliability. This should include the following information:
• data documentation including items such as a data dictionary and data model
• details of the quality assurance and data management process including audit.
Data quality
Limitations to data quality include missing data, measurement error, misclassification and
incorrect reporting of dates. These issues can apply to all study variables including patient
eligibility criteria, outcomes, interventions or exposures, and covariates. They can create
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 49 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
information biases that cause real-world evidence studies to produce biased estimates.
Transparent reporting of data quality is essential for reviewers to understand the risk of
bias and whether it has been adequately addressed through data analysis or explored
through sensitivity analysis. We focus on 2 main aspects of data quality: completeness and
accuracy.
Information on completeness and accuracy should be provided for all key study variables.
Study variables can be constructed by combining multiple data elements, including both
structured data and unstructured data, and may come from different linked data sources.
The complexity of these study variables will vary according to the data sources and
applications. For instance, in some applications an asthma exacerbation may be identified
from a single data field (such as the response to a questionnaire), while in others it may
need to be constructed from combinations of diagnostic codes, prescriptions, tests, free
text or other data.
As described in the section on study reporting, it is essential that clear and unambiguous
definitions are given for each study variable including types of data, code lists, extraction
from unstructured data, and time periods, when possible. These operational definitions
including code lists should be made available to others and reused, if appropriate. The
validity of an existing code list should be reviewed before use. When unstructured data is
used, information should be provided on data extraction and the reliability of these
methods.
These considerations also apply to data from digital health technologies producing
patient-generated data, including patient-reported outcomes and digital biomarkers.
Further information on the validity of data generated from the technology and user
accessibility should be provided.
To interpret study results, further information is needed on reasons for data missingness
and inaccuracy and whether these are random or systematic. For comparative studies, it is
important to understand the extent to which missingness or inaccuracy differ across
intervention groups. The section on addressing information bias has further information on
methods for dealing with missing data, measurement error and misclassification. We have
not set minimum thresholds for data completeness or accuracy because the acceptable
levels will depend on the application (see the section on considerations for the quality and
acceptability of real-world evidence studies).
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 50 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Completeness
Data completeness refers to the percentage of records without missing data at a given
time point. It does not provide information on the accuracy of that data. The percentage is
often easily calculated from the data source and should be calculated before excluding
relevant data or imputation. For outcomes such as experiencing a myocardial infarction,
issues of data missingness should be clearly distinguished from misclassification. For
binary variables, the absence of an event (when it has occurred) may be best summarised
as a data accuracy issue (misclassification due to false negatives).
Accuracy
Measuring accuracy, or how closely the data resemble reality, depends on the type of
variable. Below we describe common metrics of accuracy for different types of variables:
• continuous or count variables (mean error, mean absolute error, mean squared error)
• time-to-event variables (difference between actual time of event and recorded time of
event).
• comparison with an established gold standard source (for example, UK Office for
National Statistics mortality records)
These approaches may be taken for a subset of the analytical population or be based on a
previous study in the same or similar population and data source.
These gold standard approaches are not always possible or feasible. Other approaches
that can show approximate accuracy include:
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 51 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
• checking consistency (agreement in patient status in records across the data sources)
• assessing conformance (whether the recording of data elements is consistent with the
data source specifications)
• Describing the methods and processes used to quantify accuracy including any
assumptions made. When this is based on previous studies, the applicability to the
present analysis should be discussed, and may consider differences in study variable
definitions, populations, data sources, time periods or other relevant considerations.
Data relevance
The second component of data fitness for purpose is data relevancy. Key questions of
data relevancy are whether:
• the data provides sufficient information to produce robust and relevant results
The assessment of data relevancy should be informed by the information provided in the
section on data provenance.
NICE prefers data relating directly to the UK population that reflects current care in the
NHS. However, we recognise the potential value of international data if limited information
is available for the NHS or if results can be expected to translate well between settings. In
some applications there will be a trade-off between using local data and other important
characteristics of data including quality, recency, clinical detail, sample size and follow up.
International data is likely to be of particular value when an intervention has been available
in another country before becoming available in the UK, or in the context of rare diseases.
Similar considerations apply to using data from regional or specialist healthcare providers
within the NHS.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 52 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
We describe key aspects of data relevancy below distinguishing between data content,
coverage and characteristics.
Data content
There are 3 key considerations for understanding whether the data content is sufficient for
a research question:
• Does the data source contain sufficient data elements to enable appropriate
definitions of population eligibility criteria, outcomes, interventions and covariates, as
relevant?
• Are the data elements collected with sufficient granularity (or detail)?
To help understand whether data elements are sufficient, it is useful to first define the
target concept and judge the extent to which this can be proxied using real-world data.
The implications of insufficient data will vary depending on the study variable and use
case. Key endpoints necessary to answer the research questions should be available and
should be sufficiently objective and detailed to support an evaluation. Insufficient
information to define the population, interventions or outcomes appropriately will limit the
relevance of the research findings. Insufficient information on confounders will limit the
ability to produce valid findings.
The needed granularity of data will vary across research questions. For example, when
considering the effect of knee replacement on quality of life we may be interested in the
effect compared with physiotherapy alone, total versus partial knee replacement, or of
different implanted devices. Similarly, any stroke may be appropriate as an outcome for
some research questions, while others will need haemorrhagic and ischaemic strokes to
be separated.
Finally, we may be interested in the effect of knee replacement on quality of life at 1 year
after the procedure. In routinely collected data, the recording of such information does not
follow a strict protocol with measurements missing or taken at irregular time points.
Data coverage
The generalisability of research findings to patients in the NHS will depend on several
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 53 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
factors, including:
• the similarity in patient characteristics between the analytical sample and target
population
• changes in care pathways (including diagnostic tests) and outcomes over time.
The similarity of the analytical sample to the target population is especially important in
descriptive studies, such as those estimating disease prevalence. In comparative studies
this may be less important if the intervention effects are expected to transfer across
patients with different characteristics, and the emphasis should be on ensuring internal
validity. If there is substantial heterogeneity in treatment effects across subgroups,
similarity in patient characteristics becomes more important. Effect estimates on the
relative scale usually transfer better across subgroups than estimates on absolute scales
(Roberts and Prieto-Merino 2014). In other applications, such as prognostic modelling,
non-representative sampling may be preferred to ensure adequate representation of
important patient subgroups.
Consideration needs to be given to how any differences in the treatment pathways or care
settings seen in the analytical sample and the NHS may impact on the relevance of results.
This is especially important when using international data. Even within the NHS, the data
may relate to specific regions that are not representative of the country or focus on
specialist providers rather than all providers. Finally, changes to care pathways including
diagnostic tests as well as background trends in outcomes (such as mortality) may limit
the value of historical data even from the NHS. These issues need to be carefully
considered and reported when discussing the relevance of data for use in NICE guidance.
Data characteristics
The final category of data relevancy concerns the size of the analytical sample and the
length (and distribution) of follow up. The sample size should be large enough to produce
robust estimates. However, we recognise that sample size will always be limited in some
contexts. The follow up should be long enough for the outcomes of interest to have
occurred or accrued (for outcomes such as healthcare costs). The amount of data
available before the start of follow up may also be important to provide information on
confounders and identify new users of an intervention. Using data sources with a lower
time lag between data collection and availability for research may allow for longer follow
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 54 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 55 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 56 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Key messages
• Non-randomised studies can be used to provide evidence on comparative effects
in the absence of randomised controlled trials or to complement trial evidence to
answer a broader range of questions about the effects of interventions in routine
settings.
• Study design
- For studies using external control, select and curate data to minimise
differences between data sources including availability and operational
definitions of key study variables, data collection processes, patient
characteristics, treatment settings, care pathways, and time periods, and
consider the implications for study quality and relevance.
• Analysis
- Consider the impact of bias from informative censoring, missing data, and
measurement error and address appropriately, if needed.
- Assess the external validity of findings to the target population and consider
if adjustment methods are suitable or needed.
- Use sensitivity and bias analysis to assess the robustness of results to main
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 57 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
• Reporting
- Provide a study protocol and statistical analysis plan before performing final
analyses.
- Assess the risk of bias and relevance of the study to the research question.
• The acceptable quality of evidence may depend on the application and various
contextual factors (see the section on considerations for the quality and
acceptability of real-world evidence).
Introduction
We previously outlined principles for the robust and transparent conduct of quantitative
real-world evidence studies across different use cases. In this section we provide more
detailed recommendations for the conduct of studies of comparative effects using real-
world data. This includes traditional observational studies based on primary or secondary
data collection and trials in which real-world data is used to form an external control. We
do not provide specific considerations for purely interventional studies (whether
randomised or not) or external control studies using only interventional data. We focus
here on quantitative studies but recognise that qualitative evidence can play an important
role in improving our understanding of the value of interventions.
Randomised controlled trials are the preferred study design for estimating comparative
effects. Non-randomised evidence may add value if randomised controlled trials are
absent, not directly relevant to the research question or of poor quality (see the section on
uses and challenges of randomised controlled trials). They can also complement trial
evidence to answer a broader range of questions (see the section on estimating
intervention effects using real-world data).
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 58 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
The recommendations provided here are intended to improve the quality of real-world
studies of comparative effects, both in terms of methodological quality and validity, and
the transparency of study conduct. They were derived from best-practice guidance from
the published literature, international research consortia, and international regulatory and
payer bodies, and will be updated regularly in line with developing methodologies. They
build on NICE Decision Support Unit's technical support document 17, which presents
statistical methods for analysing observational data.
We recognise that not all studies will be able to meet all recommendations in full. The
ability to perform studies of the highest quality will depend on the availability of suitable
data (see the section on assessing data suitability) and characteristics of the condition
and intervention. Simpler methods may be appropriate for other applications including
assessing non-health outcomes like user experience or some system outcomes. In
addition, the acceptability and contribution of specific studies to decisions will depend on
the application as well as several contextual factors (see the section on considerations for
the quality and acceptability of real-world evidence studies).
Figure 1
Visual summary of key considerations for planning and reporting cohort studies using real-world data
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 59 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Overview
A large variety of study designs can be used to estimate the effects of interventions,
exposures or policies. The preferred study design will be context dependent. It may
depend on whether variation in the exposure is within individuals over time, between
individuals, or between other groups such as healthcare providers. In general, confidence
in non-randomised study results is strengthened if results are replicated using different
study designs or analytical methods, known as triangulation (Lawlor et al. 2016).
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 60 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Cohort studies
In cohort studies, individuals are identified based on their exposures and outcomes
compared during follow up. Usually, cohort studies will compare individuals subject to
different exposures from the same data source. However, they can also combine data from
different sources including from interventional and observational data sources. In this
case, the observational data is used to form an external control to the intervention used in
the trial. The trial will often be an uncontrolled single-arm trial but could also be an arm
from a controlled trial. External data can also be used to augment concurrent controls
within a randomised controlled trial.
External controls can also be formed from data from previous clinical trials. A potential
advantage of such studies is greater similarity in patient inclusion criteria, follow up and
outcome determination. Often only aggregate rather than individual patient-level data will
be available from previous trials. NICE Decision Support Unit's technical support document
18 describes methods for unanchored indirect comparisons with aggregated data.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 61 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
In the following study design and analysis sections, we focus on cohort studies including
those using external control from real-world data sources which are the most common
non-randomised study designs informing NICE guidance. Other study designs including
quasi-experimental designs or self-controlled studies may be relevant in some contexts as
outlined below.
Self-controlled studies
Self-controlled, or 'within-subject', designs make use of variation in exposure status within
individuals over time. These include case-crossover, self-controlled case series, and
variants of these designs. They are most appropriate for transient exposures with acute-
onset events (Hallas and Pottegard 2014). While primarily used in studies of adverse
effects of medicines (including vaccines), they have been used to assess the effects of
oncology medicines using the experiences of individuals on prior lines of therapy (Hatswell
and Sullivan 2020). This is most relevant if appropriate standard-of-care comparators are
not available.
Cross-sectional studies
In cross-sectional studies information on current exposures and outcomes is collected at a
single time point. While they can be used to estimate intervention effects, they are less
reliable than longitudinal studies (such as cohort studies) if there is need for a clear
temporal separation of exposures and outcomes.
Case-control studies
In case-control studies individuals are selected based on outcomes, and odds of
exposures are compared. Case-control studies embedded within an underlying cohort are
known as nested case-cohort studies. Case-control studies conducted within existing
database studies are generally not recommended because they use less information than
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 62 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
cohort studies (Schuemie et al. 2019). Case-control studies are most useful for rare
outcomes or if there is a need to collect further information on exposures, for example,
from manual medical record review or primary data collection.
Quasi-experimental studies
Quasi-experimental studies and natural experiments exploit external variation in exposure
across people or over time (an 'instrument') that is otherwise unrelated to the outcome to
estimate causal effects (Reeves et al. 2017, Matthay et al. 2019). Common quasi-
experimental methods include instrumental variable analysis, regression discontinuity,
interrupted time series and difference-in-difference estimation. They are frequently used
in public health settings when randomisation is not always feasible but have also been
used in medical technologies evaluations (see NICE medical technologies guidance on
Sleepio to treat insomnia and insomnia symptoms).
Examples of instruments that have been used in healthcare applications include variation
in physician treatment preferences or hospital formularies, genes, distance to healthcare
providers or geographic treatment rates, arbitrary thresholds for treatment access, or time
(for example, time of change to clinical guidelines that have immediate and substantial
impacts on care patterns).
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 63 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Study design
In this section we present study design considerations for cohort and external control
studies using real-world data. These approaches may also be useful for other non-
randomised study designs.
Studies should aim to emulate the target trial as closely as possible and, if this is not
possible, trade-offs should be clearly described. In some cases, a data source may not be
of sufficient relevance or quality to allow trial emulation. This can be particularly
problematic for studies using real-world data to form an external control because
differences in terms of patients, settings, care, data collection and time periods can limit
the comparability between the trial and the real-world data (Gray et al. 2020, Pocock
1976). Sometimes it will not be possible to adequately emulate a target trial with real-world
data and bespoke data collection may be needed.
The target trial can be defined across 7 dimensions: eligibility criteria, treatment
strategies, assignment procedure, follow-up period, outcomes, causal effect of interest
and analysis plan. We describe each dimension below and provide considerations for
those developing evidence to inform NICE guidance.
Eligibility criteria
For most studies, the eligibility criteria should mimic a hypothetical pragmatic trial by
reflecting the clinical pathways (including diagnostic tests) and patients seen in routine
care in the NHS. For external control studies, the focus should be on matching the
eligibility criteria from the interventional study rather than the broader target population.
As in a trial, eligibility criteria should be based on variables recorded before treatment
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 64 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
assignment.
Treatment strategies
Comparators that are for the same (or similar) treatment indication (that is, active
comparators) are preferred to comparison with those not receiving an intervention. Active
comparators reduce the risk of confounding by indication by ensuring greater similarity of
patients having different interventions. If routine follow-up procedures are similar across
interventions this also reduces the risk of detection bias. The active comparator should
ideally reflect established practice in the NHS.
For studies of interventions, new (or incident) user designs are generally preferred to
studies of prevalent users (those who have already been using the intervention for some
time) because of the lower risk of selection bias and better emulation of trial designs.
Prevalent users have, by definition, remained on-treatment and survived for some period
of follow up. When making use of already collected data, new users are typically defined
using an initial period in which the individual was not observed to use the intervention of
interest (known as the 'washout' period in pharmacoepidemiology). A further advantage of
new-user designs is the ability to estimate time-varying hazards from treatment initiation.
The inclusion of prevalent users may be needed if the effects of interventions are
cumulative, there are too few incident users in the data, or follow up is limited
(Vandenbroucke and Pearce 2015, Suissa et al. 2016).
Data on comparators would ideally come from the same period as the intervention as well
as from the same healthcare system and settings. This is to minimise any differences
between treatment groups resulting from differences in care access, pathways (including
diagnostic tests) or time-based trends in outcomes.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 65 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Assignment procedure
In some applications, individuals will meet eligibility criteria at multiple time points. For
example, they may start treatment more than once after a sufficient period without
exposure (or 'washout' period). There are several approaches to deal with this including
using only the first eligible time point, a random eligible time or all eligible time points
(Hernán and Robins 2016).
Follow-up period
The start and end of follow up must be defined. The start of follow up should ideally begin
at the same time at which all eligibility criteria are met and the intervention is assigned (or
just after). If a substantial latency period is expected between treatment initiation and
outcomes, it may be necessary to define an induction period before which outcomes are
not counted. This can reduce the risk of reverse causation, in which the outcome
influences the exposure.
The follow-up period should be long enough to capture the outcomes of interest but
should not exceed the period beyond which outcomes could be reasonably impacted by
the intervention (known as the exposure-effect window). Censoring events should be
clearly defined and will depend on the causal effect of interest.
Outcomes
Primary and secondary outcomes should be defined and can include both patient and
health system outcomes (such as resource use or costs). Patient outcomes should reflect
how a patient feels, functions, or how long a patient lives. This includes quality of life and
other patient-reported outcome measures. Objective clinical outcomes (such as survival)
are typically subject to a lower risk of bias than subjective outcomes if outcome detection
or reporting could be influenced by known treatment history.
For a surrogate outcome there should be good evidence that changes in the surrogate
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 66 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
outcome are causally associated with changes in the final patient outcomes of interest
(Ciani et al. 2017).
While outcome ascertainment is not blinded in observational data, analysts can be blinded
to outcomes before finalising the analysis plan (see the section on analysis).
Researchers should describe the causal effect of interest. Trials are usually designed to
estimate 1 of 2 causal effects: the effect of assignment to an intervention (intention-to-
treat) or the effect of adhering to treatment protocols (per-protocol). It is not usually
possible to estimate the effect of treatment assignment using observational data because
this is not typically recorded. However, it can be proxied using treatment initiation (the as-
started effect). The equivalent of the per-protocol effect is sometimes called the on-
treatment effect.
Analysis plan
The analysis plan should describe how the causal effect of interest is to be estimated,
taking into account intercurrent events. Intercurrent events are events occurring after
treatment initiation (such as treatment switching or non-adherence) that affect the
interpretation of the outcome of interest. This is supported by the estimand framework (for
further information, see ICH E9 [R1] addendum on estimands and sensitivity analysis in
clinical trials).
The relevance of intercurrent events will depend on the causal effect of interest. In an as-
started analysis, treatment discontinuation, switching or augmentation can usually be
ignored. However, if these changes are substantial there is a risk of increasing exposure
misclassification over time. In most cases this would bias estimates of effect towards the
null.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 67 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 68 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 69 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
The Women's Health Initiative randomised controlled trial showed that initiating
treatment with hormone replacement therapy increased the risk of coronary heart
disease in postmenopausal women. This contradicted earlier observational studies
that found a reduction in the risk of coronary heart disease. Hernán et al. 2008
followed a target trial approach, replicating as far as possible the Women's Health
Initiative trial using data from the Nurses' Health Study. They were able to show that
the difference in results between the trial and observational studies resulted from the
inclusion of prevalent users of hormone replacement therapy in the observational
cohort. These women had already survived a period of time on-treatment without
experiencing the outcome. Following a new-user design (as well as other principles of
the target trial approach) they were able to produce effect estimates consistent with
the trial.
Example 2: What is the optimal estimated glomerular filtration rate (eGFR) at which
to initiate dialysis treatment in people with advanced chronic kidney disease?
The IDEAL randomised controlled trial showed a modest reduction in mortality and
cardiovascular events for early versus late initiation of dialysis. The average eGFR
scores in the early and late treatment arms were 9.0 and 7.2 mL/min/1.73 m2,
respectively. There therefore remains considerable uncertainty about the optimal time
to initiate dialysis. Fu et al. 2021 emulated the IDEAL trial using data from the National
Swedish Renal Registry and were ability to produce similar results over the narrow
eGFR separation achieved in the trial. They were then able to extend the analysis to a
wider range of eGFR values to identify the optimal point at which to initiate dialysis
therapy.
Example 3: What is the effect of initiating treatment with fluticasone propionate plus
salmeterol (FP-SAL) versus 1) no FP-SAL or 2) salmeterol only on COPD
exacerbations in people with COPD?
The TORCH trial found that treatment with FP-SAL was associated with a reduction in
the risk of COPD exacerbations compared with no FP-SAL or salmeterol only.
However, the trial excluded adults aged above 80 years and those with asthma or mild
COPD. There is uncertainty about the extent to which results from the TORCH trial
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 70 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
apply to these patients. Wing et al. 2021 were able to replicate the findings of the
TORCH trial for COPD exacerbations using primary care data from Clinical Practice
Research Datalink in England for the comparison with salmeterol only but not with no
FP-SAL. This reflects the challenge in emulating a trial with placebo control. By
extending their analysis to a wider target population they were able to demonstrate
evidence of treatment effect heterogeneity by COPD severity but not by age or
asthma diagnosis.
Analysis
Developers should outline their assumptions about the causal relationships between
interventions, covariates and outcomes of interest. Ideally, this would be done using
causal diagrams known as directed acyclic graphs (Shrier and Platt 2008).
Inappropriate adjustment for covariates should be avoided. This may result from
controlling for variables on the causal pathway between exposure and outcomes
(overadjustment), colliders or instruments. Confounders that may change value over time
should be recorded before the index date, except when using statistical methods that
appropriately address time-varying confounding.
The selection of covariates may use advanced computational approaches such as machine
learning to identify a sufficient set of covariates, for example, when the number of
potential covariates is very large (Ali et al. 2019, Tazare et al. 2022). The use of these
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 71 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
methods should be clearly justified and their consistency with causal assumptions
examined. Choosing covariates based on statistical significance should be avoided.
Adjusted comparisons based on clear causal assumptions are preferred to naive (or
unadjusted) comparisons. Statistical approaches should be used to address confounding
and approximate randomisation (see the section on assignment procedure).
If there are many potential confounders, more complex methods such as multivariable
regression and propensity score (or disease risk score) methods are preferred. Propensity
scores give the probability of receiving an intervention based on observed covariates.
Several methods use propensity scores including matching, stratification, weighting and
regression (or combinations of these). General discussions of the strengths and
weaknesses of these different approaches can be found in Ali et al. 2019. The choice of
method should be justified and should be aligned with the causal effect of interest.
There is mixed evidence on the relative performance of regression and propensity score
methods for addressing confounding bias (Stürmer et al. 2006). However, using propensity
score methods may have advantages in terms of the transparency of study conduct:
• Propensity scores are developed without reference to outcome data, which can
reduce the risk of selective reporting of results when combined with strong research
governance processes.
Regression and propensity score methods may also exclude some participants to enhance
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 72 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
the similarity of people across intervention arms or levels. When using such methods,
trade-offs between internal validity, power and generalisability should be considered. For
studies of comparative effects, internal validity should generally be prioritised.
Time-varying confounders should typically not be adjusted for using the above methods. It
may be acceptable for on-treatment analyses if confounders that vary over time are not
affected by previous levels of the intervention but this is uncommon. G-methods including
marginal structural models with weighting are preferred (Pazzagli et al. 2017, Mansournia
et al. 2017). Adjustment for time-varying confounders requires high-quality data over the
whole follow-up period.
Various sensitivity and bias analyses can be used to adjust for bias because of residual
confounding or to explore its likely impact (see the section on assessing robustness of
studies). This may be informed by external data on confounder-outcome relationships or
data from a data-rich subsample of the analytical database, if available (Ali et al. 2019).
Negative controls (that is, outcomes that are not expected to be related to the
intervention) may also be useful (Lipsitch et al. 2010).
If there are multiple potential sources of suitable real-world data to provide external
control to trial data, developers should consider whether to estimate effects separately for
each data source or to increase power by pooling data sources. Data sources should only
be pooled when there is limited heterogeneity between sources in terms of coverage and
data quality. Individual estimates of effects for each data source should always be
provided.
External controls can also be used to supplement internal (or concurrent) controls in
randomised controlled trials. There are several methods available to combine internal and
external controls, which place different weight on the external data (NICE Decision
Support Unit report on sources and synthesis of evidence).
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 73 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Informative censoring
Censoring occurs in longitudinal studies if follow up ends before the outcome is fully
observed. It can happen because the data collection period ends (administrative
censoring), loss to follow up, occurrence of events such as treatment switching, non-
adherence, or death depending on the analysis. It may be induced by analytical strategies
such as cloning to avoid time-related biases in studies without active comparators (Hernán
and Robins 2016).
Censoring can create bias if it is informative (that is, it is related to the outcomes and
treatment assignment). For example, in on-treatment analyses, if people on an
experimental drug were less likely to adhere to the treatment protocol because of a
perceived lack of benefit this could lead to informative censoring. When modelling effects
on-treatment or dynamic treatment strategies, censoring because of treatment switching
is likely to be informative. Methods to address informative censoring are similar to those
for time-varying confounding such as marginal structural models with weighting or other
G-methods (Pazzagli et al. 2017). Methods for dealing with missing data may also be used
(see the section on missing data).
Missing data
The impact of missing data depends on the amount of missing data, the variables that
have missing data, and the missing data mechanism. Developers should compare patterns
of missingness across exposure groups and over time, if relevant, considering causes of
missingness and whether these are related to outcomes of interest. Missing data on
outcomes may arise for a number of reasons including non-response to questionnaires or
censoring.
If the amount of missing data is low and likely to be missing completely at random,
complete records analysis will be sufficient. Advanced methods for handling missing data
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 74 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
include imputation, inverse probability weighting and maximum likelihood estimation. Most
of these methods assume the missing data mechanism can be adequately modelled using
available data (that is, missing at random). If this is not the case, sensitivity or bias
analysis may be preferred (see the section on assessing robustness). A framework for
handling missing data is provided in Carpenter and Smuk 2021.
Random measurement error in exposures tends to (but does not always) bias estimates of
treatment effects towards the null (van Smeden et al. 2020). Random measurement error
in continuous outcomes reduces the precision of estimates but provides unbiased
estimates of comparative effects. For risk ratios and rate ratios, non-differential
misclassification of a categorical outcome provides unbiased estimates of comparative
effects when specificity is 100%, even if sensitivity is low. So, it is often recommended to
define outcome variables to achieve high specificity.
This section focuses on methods to assess and address external validity bias resulting
from differences in patient characteristics (for example, age, disease risk scores) between
the analytical sample and the target population. Importantly, differences in patient
characteristics may not be the only, or most important, sources of external validity bias.
Developers should also consider differences in: setting (for example, hospital type and
access to care), treatment (for example, dosage or mode of delivery, timing, comparator
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 75 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
therapies, concomitant and subsequent treatments) and outcomes (for example, follow-
up, measurements, or timing of measurements). Identifying a suitable data source, using a
target trial approach and using internally valid analysis methods remain the primary
approaches by which external validity can be achieved.
To assess external validity, an explicit definition of the target population is needed and
suitable reference information. Information can be drawn from published literature,
context-relevant guidelines, or bespoke analysis of data from the target population
alongside information gathered during the data suitability assessment.
To assess differences between the analytical sample and target population for patient
characteristics, several tests are available:
• averages and distributions of individual variables can be compared (for example, using
absolute standardised mean differences);
Methods to adjust for external validity bias are similar to those which adjust for
confounding bias, including matching, weighting, and outcome regression methods. These
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 76 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
• Regression methods model outcomes in the analytical sample and then standardise
model predictions to the distribution of covariates in the target population.
Degtiar and Rose 2022 provides further guidance on these methods, including approaches
for when only summary-level data is available for the target population. Adjustment
approaches are unlikely to perform well when the target population is poorly represented
in the analytical sample, that is, where there is insufficient overlap for important
covariates, or across strata of these variables. Successful application of these methods
also depends on good internal validity of analyses and consistency in measurements of
outcomes, treatments, and covariates across settings.
Where the sample is drawn from an entirely different population to the target population,
judgements of similarity will require stronger assumptions. Pre-specified, empirical
assessments of 'transportability' for the decision context could provide supportive
evidence (for example, see Ling et al. 2023). In all cases, sensitivity analyses are
recommended to explore potential violation of study assumptions (for example, see
Dahabreh et al. 2023, Nguyen et al. 2018).
For key risks of bias (for example, those arising because of unmeasured confounding,
missing data or measurement error in key variables), quantitative bias analysis may be
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 77 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
valuable. Quantitative bias analysis describes a set of techniques that can be used to:
• examine the extent to which bias would have to be present to change results or affect
a threshold for decision making, or
• estimate the direction, magnitude and uncertainty of bias associated with measures of
effect.
Methods that examine the extent to which bias would have to be present to change study
conclusions tend to be simpler and include the e-value approach. These approaches are
most useful when exploring a single unmeasured source of bias, however sources of bias
are often multiple and may interact. Developers should consider and pre-specify a
plausible level of bias in the parameter before application of these methods. More
sophisticated approaches look to model bias and incorporate it into the estimation of
effects (Lash et al. 2014). Bias parameters can be informed by external information or
data-rich subsamples of the analytical data source. The identification and validity of
external bias should be clearly described and justified.
Bias analysis may be particularly valuable in studies using real-world data external controls
if differences between data collection, settings and time may reduce comparability of
data. Panel 2 shows an example of bias analysis in practice, and table 3 shows examples
of sensitivity analysis.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 78 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Bias analysis methods were used to explore these risks. The e-value approach was
used to estimate the relative risk of an unobserved confounder between intervention
and mortality that would be needed to remove the treatment effect. The estimated
relative risk of 2.2 was substantially higher than for any observed confounders and
considered unlikely given the estimated imbalance for important but poorly captured
confounders.
For missing ECOG data they assumed the causes of missing data were non-random
and missing data values in the ceritinib arm were likely to be worse than expected
based on multiple imputation. They argued that no plausible assumptions about
missing data could explain the observed association between intervention and
mortality.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 79 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Table 3
Examples of sensitivity analyses to examine robustness of results to data curation,
study design, and analysis decisions
• On-treatment analyses
Exposure
misclassification • Vary exposure definitions including, if relevant, days' supply, grace
period, washout period, exposure-effect window and latency
Reverse • Introduce or change lag time between exposure end and start of
causation follow up for outcomes
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 80 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Reporting
We provide general principles for the transparent reporting and good conduct of real-
world evidence studies in the section on conduct of quantitative real-world evidence
studies. The following reporting considerations are especially important for comparative
effects studies:
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 81 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
• Justification of the use of real-world evidence. This should cover, as relevant, the
reasons for the absence of randomised evidence, the limitations of existing trials and
the ability to produce meaningful real-world evidence for the specific research
question.
• For each data source, provide the information needed to understand data provenance
and fitness for purpose (see the section on assessing data suitability).
• Justify the use of statistical method for addressing confounding and report methods
clearly (see appendix 3).
• Clearly describe the exclusion of patients from the original data to the final analysis,
including reasons for exclusion using patient flow (or attrition) diagrams.
• Present results for adjusted and unadjusted analyses and for all subgroup and
sensitivity and bias analyses.
Quality appraisal
Evidence developers should identify risks of bias at the study planning stage. These
should be described alongside how design and analytical methods have been used to
address them, and how robust the results are to deviations from assumptions in the main
analysis using sensitivity or bias analysis. This can be done for specific domains of bias
using the reporting methods in appendix 2. This information will help those completing (or
critically appraising) risk of bias tools. The preferred risk of bias tool for non-randomised
studies is the ROBINS-I (Sterne et al. 2016) but it should be recognised that it may not
cover all risks of bias (D'Andrea et al. 2021). It should be recognised that the uncertainty in
non-randomised studies will not typically be fully captured by the statistical uncertainty in
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 82 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 83 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Research question
Add the research question here.
Data provenance
Item Response
For each contributing data source provide the name, version and date of
Data sources
data cut. Provide links to their websites, if available.
Report which datasets were linked, how these were linked, and
Data linkage
performance characteristics of the linkage. Note whether linkage was
and data
done by a third party (such as NHS Digital).
pooling
Clearly describe which data sources were pooled.
Type of data Describe the types of data source (for example, electronic health record,
source registry, audit, survey).
Purpose of
Describe the main purpose of data collection (for example, clinical care,
data
reimbursement, device safety, research study).
collection
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 84 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Item Response
State the setting of care for each dataset used (for example, primary
Care setting care, secondary care, specialist health centres, social services, home
use [for wearable devices, or self-reported data on apps or websites]).
Geographical
State the geographical coverage of the data sources.
setting
Population State how much of the target population is represented by the dataset
coverage (for example, population representativeness or patient accrual).
Time period
State the time period covered by the data.
of data
Data Provide the details of the data controller and funding for each source.
governance Describe the information governance processes for data access and use.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 85 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Item Response
Data
management
plan and Note whether a data management plan, documentation of source quality
quality assurance methods is available with links to relevant documents.
assurance
methods
Data quality
Details of data quality should be provided for key study variables including population
eligibility criteria, outcomes, interventions or exposures, and covariates.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 86 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Data relevance
Please see recommendations for reporting data relevance.
Item Response
Describe the extent to which the analytical sample reflects the target
Population population. This should consider any data exclusions (for example,
because of missing data on key prognostic variables).
Care setting Describe how well the care settings reflect routine care in the NHS.
Note how the dataset met the requirements of the research question in
Availability of
terms of availability of the necessary data variables including key
key study
population eligibility criteria, outcomes, intervention and covariates
elements
(including confounders and effect modifiers).
State the extent to which the time period covered by the data provides
relevant information to decisions. This should cover any important
Study period
changes to care pathways (including tests) or background changes in
outcome rates.
Timing of Describe whether the timing of measurements meet the needs of the
measurements research question.
Note how the follow-up period available in the dataset is sufficient for
Follow up
assessing the outcomes.
Provide the sample size of the target population in the dataset and
Sample size
demonstrate that it is adequate to generate robust results.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 87 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Research question
What is the effect of the long-acting beta-2 agonist and inhaled corticosteroid
combination product fluticasone propionate plus salmeterol compared with no exposure or
exposure to salmeterol only in people with chronic obstructive pulmonary disease (COPD)?
Data provenance
Item Response
CPRD and HES are linked. Patients are identified in a centralised linkage
algorithm done by NHS digital. This uses an 8-step deterministic linkage
Data linkage algorithm based on 4 identifiers: NHS number, sex, date of birth and
and data postcode.
pooling
Linkage to HES data is possible for 75% of enrolled patients.
See information on linked data for CPRD.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 88 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Item Response
HES = England
Geographical
CPRD = a representative sample of UK general practices using Vision
setting
software. HES-linked CPRD data is available for England only.
CPRD GOLD has data for about 3 million currently registered people
Population
(around 4.74% of UK population). See CPRD data highlights
coverage
HES data covers all NHS Clinical Commissioning Groups in England.
Time period The CPRD-linked HES dataset covers from January 2000 to January
of data 2017.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 89 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Item Response
Fields in HES are derived from the NHS data model and the NHS data
Data
dictionary.
specification
CPRD GOLD data specification document.
Data
HES undertakes processing and data quality checks: The processing
management
cycle and HES data quality.
plan and
No data quality assurance information was identified for CPRD GOLD.
quality
However, records from individual general practices are assessed and
assurance
only included in CPRD after being deemed of research quality.
methods
Other
None.
documents
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 90 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Data quality
Previously
published
validation
CPRD study
diagnostic comparing
Positive
(Read v2) algorithms for
predictive
codes for identifying
value (PPV):
COPD (see people with
Population COPD Accuracy 87% (95%
codelist in COPD with
Confidence
supplementary physician
interval [CI]
material of review
78% to 92%)
Quint et al. questionnaire
2014) as gold
standard
(Quint et al.
2014)
Global
Initiative for
Chronic
Proportion of
Obstructive
patients with
Disease Lung Disease
Population Completeness missing 20%
severity (GOLD) stage
spirometry
derived from
data
spirometry
measurements
(see codelist)
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 91 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
CPRD
prescribing
CPRD
record
prescribing
Fluticasone matching
data is
Intervention propionate + definition of Accuracy n/a
expected to
salmeterol drug
be highly
treatment
accurate
determined by
codelist
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 92 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Any of the
following:
CPRD
diagnostic
(Read) code
for lower
respiratory
tract infection
or acute Previously
exacerbation published
of COPD validation
study
A prescription
comparing
of a COPD-
algorithms for PPV: 86%
specific
identifying (95% CI
antibiotic
people with 83% to
combined with
COPD COPD 88%)
Outcome oral Accuracy
exacerbation exacerbations Sensitivity:
corticosteroid
with 63% (95%
(OCS) for 5 to
physician CI 55% to
14 days
review 70%)
A record questionnaire
(Read code) of as gold
2 or more standard
respiratory (Rothnie et al.
symptoms of 2016)
AECOPD with
a prescription
of COPD-
specific
antibiotics
and/or OCS on
the same day.
See codelist
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 93 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Record in
Office for
National
ONS mortality
Statistics
records are
All-cause (ONS)
Outcome Accuracy the gold n/a
mortality mortality
standard data
statistics
for deaths
(centrally
linked to CPRD
data)
Proportion of
Reported
patients with
Covariate Alcohol directly in
Completeness missing data 30%
(confounder) intake CPRD (closest
on alcohol
to index date)
intake
Data relevance
Item Response
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 94 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Item Response
Treatment
The data represents routine practice in the NHS.
pathway
The longitudinal nature of the analysis allows for the research question
Timing of
to be answered. The date of entry is expected to reflect the actual
measurements
timing of clinical events well.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 95 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Selection bias at study entry can arise for several reasons including
selection of patients based on eligibility criteria related to the
exposure and outcome, or from deviations between the date the
patient meets eligibility criteria, the date treatment is assigned, and
Selection bias
the start of follow up. Common types of time-related bias are
at study entry
prevalent-user bias, lead time bias, immortal time bias and depletion
of susceptibles. Discuss the potential for selection bias at study entry
and how this was addressed or investigated through study design,
statistical analysis or sensitivity analysis.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 96 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 97 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Describe the potential for bias from missing data (this should be
informed by assessment of data suitability). Consider which variables
have missing data, whether this is random or systematic, and how it
differs across comparison groups.
Missing data Show how you have addressed risks of bias using statistical methods
(such as multiple imputation) and demonstrating their validity. If
missingness may not be explainable by observed variables or has
unknown mechanisms, sensitivity or bias analysis can be used to
explore the impact of different missing 'not at random' assumptions.
The study assesses the impact of initiating dialysis at different estimated glomerular
filtration rates (eGFR) on cardiovascular events and survival in people with advanced
chronic kidney disease. The study used data from the Swedish Renal Registry.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 98 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 99 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
• replicated the results of the IDEAL trial over the eGFR separation
observed in the trial.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms- Page 100 of
and-conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Reverse
Reverse causation is not expected to be a problem in this analysis.
causation
The study estimates the comparative effectiveness of alectinib versus ceritinib on survival
in people with ALK-positive non-small-cell lung cancer. The study uses real-world data on
ceritinib from Flatiron Health to form an external control to patients having alectinib in
phase 2 trials.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms- Page 101 of
and-conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Selection bias This was an as-started analysis with limited loss to follow up.
at study exit Censoring is not thought to be informative.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms- Page 102 of
and-conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms- Page 103 of
and-conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Reverse
Reverse causation is not expected to be a problem in this analysis.
causation
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms- Page 104 of
and-conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
• Number of observations in
Dividing the data into subsets, or
Stratification each stratum
strata for analysis
• Descriptive statistics and
results within each stratum
• Matching algorithm
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms- Page 105 of
and-conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms- Page 106 of
and-conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms- Page 107 of
and-conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
• Strength of association
between instrument and
intervention (for example,
odds ratio, risk difference)
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms- Page 108 of
and-conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms- Page 109 of
and-conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
• Pre-specification of point of
intervention effect (for
example, explanation needed
if point of analysis is not point
of intervention delivery)
• Number of pre-intervention,
post-intervention, and
Individuals or groups are used as
between-intervention data
their own controls and observed
(time) points, and the data
Interrupted over multiple time points. Effects
points contributing to
time series are observed by comparing
forecasting
outcome trends in the time period
before and after intervention. • Table comparing participant
characteristics and missing
data across each group
analysed (for example, before
and after intervention and for
defined subgroups)
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms- Page 110 of
and-conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms-and- Page 111 of
conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
• health charities
• healthcare professionals
• academia
We would like to thank everyone who took part in the development and review of the real-
world evidence framework.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms- Page 112 of
and-conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms- Page 113 of
and-conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
Update information
Minor changes since publication
March 2024:
• We added information about external validity bias to the section on risk of bias.
• We added a new section on assessing external validity bias and adjusting for
differences between the study sample and the target population.
July 2023:
• We updated links to recent case studies in the section on use of real-world evidence in
NICE guidance.
- federated data networks and how they relate to common data models in the
section on assessing data suitability
• We updated the link to the hospital episode statistics GDPR webpage in appendix 1.
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms- Page 114 of
and-conditions#notice-of-rights). 115
NICE real-world evidence framework (ECD9)
experimental studies.
We made minor changes to style and language throughout without changing the meaning.
ISBN: 978-1-4731-4640-2
© NICE 2024. All rights reserved. Subject to Notice of rights (https://www.nice.org.uk/terms- Page 115 of
and-conditions#notice-of-rights). 115