Missing outcome data in clinical trials may jeopardize the validity of the trial results and inferences for clinical practice. Although sick and preterm newborns are treated as a captive patient population during their stay in the NICUs, their long-term outcomes are often ascertained after discharge. This greatly increases the risk of attrition. We surveyed recently published perinatal and neonatal randomized trials in 7 high-impact general medical and pediatric journals to review the handling of missing primary outcome data and any choice of imputation methods. Of 87 eligible trials in this survey, 77 (89%) had incomplete primary outcome data. The missing outcome data were not discussed at all in 9 reports (12%). Most study teams restricted their main analysis to participants with complete information for the primary outcome (61 trials; 79%). Only 38 of the 77 teams (49%) performed sensitivity analyses using a variety of imputation methods. We conclude that the handling of missing primary outcome data was frequently inadequate in recent randomized perinatal and neonatal trials. To improve future approaches to missing outcome data, we discuss the strengths and limitations of different imputation methods, the appropriate estimation of sample size, and how to deal with data withdrawal. However, the best strategy to reduce bias from missing outcome data in perinatal and neonatal trials remains prevention. Investigators should anticipate and preempt missing data through careful study design, and closely monitor all incoming primary outcome data for completeness during the conduct of the trial.

Missing outcome data in randomized controlled trials (RCTs) may jeopardize the comparability between the comparison groups and compromise inferences about treatment effects.1 Missing data may arise in many ways in perinatal and neonatal trials. In the birth hospital or NICU, parents or guardians of participating infants may not respond to questions, fail to adhere to trial protocols, request withdrawal from protocols early, or refuse ongoing data collection. Although caregivers can easily assess in-hospital outcomes for sick and preterm newborn infants admitted to the ICUs, their important longer-term outcomes require ascertainment later in childhood and increase the risk of attrition. For example, nearly 40% of the primary outcome data were missing at age 5 years in a recent high-profile trial comparing general anesthesia with awake regional anesthesia for infants undergoing hernia repair.2 

Despite the large literature on prevention and handling of missing data,3,7 reporting and dealing with missing data remain inadequate in clinical trials.8,9 It is plausible that perinatal and neonatal trials are not exceptions. We surveyed recently published perinatal and neonatal RCTs in high-impact general medical and pediatric journals to review the handling of missing primary outcome data and use of imputation methods. We discuss the strengths and limitations of different imputation methods, and recommend strategies to prevent or minimize missing primary outcome data in future perinatal and neonatal RCTs.

We searched Medline (via PubMed) for RCTs that enrolled newborns or their birthing parents, reported outcomes in the children, and were published between January 1, 2020, and December 31, 2022. The search was limited to selected high-impact general medical journals (Lancet, New England Journal of Medicine, Journal of the American Medical Association, British Medical Journal), and pediatric specialty journals (JAMA Pediatrics, Pediatrics, Lancet Child and Adolescent Health). We did not intend to perform a systematic review of imputation practices in all recently published perinatal and neonatal trials. Instead, we examined a small sample of target journals for perinatal and neonatal RCTs with high impact factors to examine best-case scenarios of current imputation practices.

Our search strategy is summarized in Supplemental Table 5. The inclusion criteria for our survey were: Trials must have enrolled newborn infants randomized before 28 days after birth, and for trials that enrolled pregnant women, outcomes had to be measured in their children at any time after birth. Data extraction from eligible trials included the numbers of participants who were randomized and numbers included in primary analyses for primary outcomes. When studies provided information on missing primary outcome data for both mothers and children, we extracted both data sets and included the larger amount of missing primary outcome data in our analysis. For RCTs with missing primary outcome data we examined:

  1. whether authors reported the missing data in their results;

  2. what statistical method they used for the primary analysis; and

  3. whether they performed any sensitivity analysis accounting for the missing data, and if so, what imputation methods were used.

Data screening and extraction were conducted by authors in pairs (J.Z. and Y.L., G.L. and Q.M.) and in duplicate, with discrepancies resolved by discussion or adjudicated (L.T.). Continuous data were described by means and SDs, and categorical data by counts and percentages. We compared the trials published in general medical journals with those published in pediatric journals using χ2 or Fisher’s exact tests.

A total of 87 eligible perinatal and neonatal RCTs were included in this survey (Supplemental Fig 1). Of these, 77 (89%) trials had missing primary outcome data. The mean percentage of randomized participants who had missing primary outcome data were 11% (SD 16%). The majority of missing primary outcomes were dichotomous (n = 61, 79%), and the remaining missing outcomes were continuous (n = 16, 21%). The dichotomous outcomes included death, either as a single outcome or as component in a composite outcome (n = 21, 27%); infection (n = 9, 12%); and a variety of other endpoints. Infant weights were the most common continuous primary outcome (n = 9; 12%) to be missing.

Table 1 summarizes our findings in the 77 trials with missing primary outcome data. The authors of 9 RCTs (12%) did not report the missing data in their results or comment on the missing data. Sixty-one study teams (79%) used a complete-case analysis as their primary approach, which meant that they excluded participants with missing primary outcome data from the analysis. Only 38 (49%) trial reports included any sensitivity analysis using a variety of imputation methods (Table 1). Reported approaches to the main analysis differed between trials published in general medical journals when compared with those published in pediatric journals. Sensitivity analyses were used more often in the former (Table 1).

TABLE 1

Practice of Handling Missing Primary Outcome Data in 77 Perinatal and Neonatal RCTs Published Between 2020 and 2022 in 7 High-Profile Pediatric and General Journals

Practice of Handling Missing Outcome DataTotal Number of RCTs (N = 77), No. (%)RCTs in Pediatric Journals (N = 35), No. (%)RCTs in General Journals (N = 42), No. (%)P
Not reporting missing data in the results 9 (12) 7 (20) 2 (5) .07d 
Primary approach for intervention effect estimates 
 Complete-case analysis 61 (79) 23 (66) 38 (91) <.01d 
 Best-case scenario 1 (1) 1 (2) 
 Unclear 15 (20) 12 (34) 3 (7) 
Sensitivity analysis performed for intervention effect estimates 
 Yes 38 (49) 10 (29) 28 (67) <.01e 
 No 39 (51) 25 (71) 14 (33) 
Imputation methods used in 38 RCTs with sensitivity analysisa 
 Single imputationb 15 (40) 2 (20) 13 (46) .26d 
 Multiple imputation 22 (58) 4 (40) 18 (64) .27d 
 Inverse probability weighting method 3 (8) 2 (20) 1 (4) .16d 
 Model-based imputationc 2 (5) 2 (7) — 
 Unclear 3 (8) 2 (20) 1 (4) .16d 
Sensitivity analysis yields different results from primary approach 
 Yes 1 (3) 1 (4) .74d 
 No 31 (81) 8 (80) 23 (82) 
 Unclear 6 (16) 2 (20) 4 (14) 
Practice of Handling Missing Outcome DataTotal Number of RCTs (N = 77), No. (%)RCTs in Pediatric Journals (N = 35), No. (%)RCTs in General Journals (N = 42), No. (%)P
Not reporting missing data in the results 9 (12) 7 (20) 2 (5) .07d 
Primary approach for intervention effect estimates 
 Complete-case analysis 61 (79) 23 (66) 38 (91) <.01d 
 Best-case scenario 1 (1) 1 (2) 
 Unclear 15 (20) 12 (34) 3 (7) 
Sensitivity analysis performed for intervention effect estimates 
 Yes 38 (49) 10 (29) 28 (67) <.01e 
 No 39 (51) 25 (71) 14 (33) 
Imputation methods used in 38 RCTs with sensitivity analysisa 
 Single imputationb 15 (40) 2 (20) 13 (46) .26d 
 Multiple imputation 22 (58) 4 (40) 18 (64) .27d 
 Inverse probability weighting method 3 (8) 2 (20) 1 (4) .16d 
 Model-based imputationc 2 (5) 2 (7) — 
 Unclear 3 (8) 2 (20) 1 (4) .16d 
Sensitivity analysis yields different results from primary approach 
 Yes 1 (3) 1 (4) .74d 
 No 31 (81) 8 (80) 23 (82) 
 Unclear 6 (16) 2 (20) 4 (14) 

JAMA Pediatrics, Pediatrics, and Lancet Child and Adolescent Health; Lancet, New England Journal of Medicine, Journal of the American Medical Association, and British Medical Journal. —, no comparison was made.

a

RCTs do not add up to 38 because some trials used >1 imputation method.

b

The following single imputation methods were used: Worst/best-case scenario (11 RCTs; 73%), mean/mode imputation (3 RCTs; 20%), and last observation carried forward (1 RCT; 7%).

c

The following model-based imputation methods were used: Linear mixed-effects model and pattern-mixture model (1 RCT or 50% for each method).

d

Calculated from Fisher’s exact test.

e

Calculated from χ2 test.

Statisticians distinguish between data that are missing completely at random, data that are missing at random, and data that are missing not at random.10 An example of data that are missing completely at random may be an infant’s blood sample that is subsequently lost in a usually well-run laboratory. In this scenario, the blood test results are missing completely independently of the infant’s characteristics, the usual performance of the study setting, or the study team’s adherence to the study protocol.

Missing at random data can sometimes be deduced from other observed information about the participant or setting and statistical adjustments can be made accordingly. In this scenario, the missing data should likely be independent of any unobserved information for that participant. For example, an infant’s ventilation data may be missing because of known sloppiness at the study site.

In contrast, if data are missing not at random, bias can still arise from unobserved characteristics of the participant or setting, even after adjustments have been made for all observed variables. For example, outcome data may be missing in a delivery room resuscitation trial for a certain day of the week, possibly related to a shift schedule. These data are likely missing not at random, because their completeness may depend on the unobserved and likely variable skills or number of personnel in the delivery room.

Complete-case analysis was frequently used by reports in our survey as the only approach to the analysis of the primary outcome. Authors who restricted their analysis to participants with complete outcomes may have implicitly, but often incorrectly, assumed that outcome data in the excluded participants are missing completely at random. Understanding and exploring the reasons for missing data allow reassurance of the comparability between groups and inferences about treatment effects. Only data that are missing completely at random allow researchers to assume participants with complete outcome data remain a representative subset of the total study sample. However, it is generally unrealistic to assume that data are missing completely at random.3 Besides, unfortunately, there is no easy way to verify post hoc whether the data may be in truth missing at random or even missing not at random.11,12 Therefore, sensitivity analyses are required to evaluate the impact and robustness of various assumptions of why data are missing.13 

Missing outcome data increases the risk of bias, decreases statistical power and precision, and reduces the representativeness of the sample population, thereby weakening the applicability, reliability, validity, and interpretability of the trial results. Missing data that could affect trial conclusions may occur in trials of treatments for various diseases.14,17 Up to 33% of 235 RCTs with statistically significant results for their dichotomous primary outcomes were reported to lose statistical significance under the following plausible assumption: “The incidence of events in those lost to follow-up relative to those followed up is higher in the intervention than control group.”18 All these 235 RCTs had been published in leading medical journals.18 

Imputation is defined as “the practice of filling in missing data with plausible values.”19 Methods to perform this include single imputation, multiple imputation, inverse probability weighting, and model-based imputation. These 4 types of imputations are described and discussed in Table 2. In our survey, among the 38 trial reports with sensitivity analyses, a single imputation method was used in 15 (40%), a multiple imputation technique in 22 (58%), an inverse probability weighting method in 3 (8%), and a model-based method in 2 (5%) trials. The authors of 3 trials (8%) did not specify the type of imputation methods they used. Some trials used >1 imputation method.

TABLE 2

Description and Discussion of Imputation Methods

Type of MethodsDescriptionDiscussion
Single imputation methods 
 Mean imputation An infant’s missing data for a given variable is substituted by the mean value of that variable from other infants whose data are available. • This method artificially reduces variation of the variable, and ignores correlations with other observed variables.10  
 Last observation carried forward Data are imputed from the participant’s previously observed value, which assumes that this last value is “frozen in time.” • The assumption that the last observed value has remained unchanged is frequently questionable. For example, an infant who is discharged alive may die before the study follow-up date.20  
• Guidelines caution against this as primary approach.3  
 Worst/best-case scenarios This method replaces the missing data with the worst (or best) observed value for this variable in the available data set. • Both assumptions are frequently questionable.21  
• If used, further sensitivity analyses should be done to test the robustness of the assumption. 
 Fitted regression model This method uses a fitted regression model to “predict” missing data for imputation. • This method artificially exaggerates the multivariable associations in the data, and unduly enlarges the certainty of the imputed value as estimated from other participants.3  
Multiple imputation method This method involves 3 stages of imputation, analysis, and synthesis. The stages are conducted to account for uncertainty associated with the imputation technique itself. • This method accounts for uncertainty and provides unbiased results in the case of data that are missing at random.5,22  
1. Imputation stage: Multiple imputation method generates various copies of original data sets. In each data set, missing data are imputed with values generated by random sampling from the predicted distribution of observed data. That allows for the uncertainty of imputing by creating variation in the imputed data. • The multiple imputation method is becoming increasingly popular. 
2. Analysis stage: A model is fitted to estimate the intervention effect in each imputed data set. The degree of variation of the imputed data reflects the uncertainty of the technique. 
3. Synthesis stage: All effect estimates are pooled to calculate the average effect estimate from all imputed data sets as the point estimate; the SE incorporates both within and between variation of the imputed data sets. 
Inverse probability weighting method This method involves 3 steps. • This method inflates information for participants with complete data to account for similar participants with missing data. Hence, infants less likely to have complete data (ie, with low predicted probability of having complete data) but included in the final analysis will be given a large weighting.23,24  
1. First, a logistic regression model calculates each individual’s predicted probability of “nonmissingness,” where the response variable is the nonmissingness and the covariates are the available possible predictors of nonmissingness. • This method helps to reduce selection bias from complete-case analysis. 
2. The weighting for participants with no missing data is determined by the inverse of their predicted probability of nonmissingness. 
3. A weighted regression model is performed that includes only those participants with complete data for analysis. 
Model-based imputation methods These methods are based on statistical models, such as likelihood-based methods, mixed models, Monte Carlo Markov chains, and, more recently, expectation–maximization algorithms and artificial intelligence methods.25 27 Missing data are imputed on the basis of statistical models in 4 steps. • These methods are likely robust and efficient.25,28  
Step 1: The maximum likelihood function estimates parameters on the basis of the nonmissing data. • However, complex algorithms demand strong statistical background and create interpretation difficulty for clinicians. 
Step 2: Missing data are imputed using the maximum likelihood function and on the basis of the estimated parameters identified from Step 1. 
Step 3: Step 1 is reconducted on the basis of the imputed data from Step 2. 
Step 4: Iteration of Steps 1, 2 and 3; that is, repeating all previous steps until no more changes in parameter estimates are observed. 
Type of MethodsDescriptionDiscussion
Single imputation methods 
 Mean imputation An infant’s missing data for a given variable is substituted by the mean value of that variable from other infants whose data are available. • This method artificially reduces variation of the variable, and ignores correlations with other observed variables.10  
 Last observation carried forward Data are imputed from the participant’s previously observed value, which assumes that this last value is “frozen in time.” • The assumption that the last observed value has remained unchanged is frequently questionable. For example, an infant who is discharged alive may die before the study follow-up date.20  
• Guidelines caution against this as primary approach.3  
 Worst/best-case scenarios This method replaces the missing data with the worst (or best) observed value for this variable in the available data set. • Both assumptions are frequently questionable.21  
• If used, further sensitivity analyses should be done to test the robustness of the assumption. 
 Fitted regression model This method uses a fitted regression model to “predict” missing data for imputation. • This method artificially exaggerates the multivariable associations in the data, and unduly enlarges the certainty of the imputed value as estimated from other participants.3  
Multiple imputation method This method involves 3 stages of imputation, analysis, and synthesis. The stages are conducted to account for uncertainty associated with the imputation technique itself. • This method accounts for uncertainty and provides unbiased results in the case of data that are missing at random.5,22  
1. Imputation stage: Multiple imputation method generates various copies of original data sets. In each data set, missing data are imputed with values generated by random sampling from the predicted distribution of observed data. That allows for the uncertainty of imputing by creating variation in the imputed data. • The multiple imputation method is becoming increasingly popular. 
2. Analysis stage: A model is fitted to estimate the intervention effect in each imputed data set. The degree of variation of the imputed data reflects the uncertainty of the technique. 
3. Synthesis stage: All effect estimates are pooled to calculate the average effect estimate from all imputed data sets as the point estimate; the SE incorporates both within and between variation of the imputed data sets. 
Inverse probability weighting method This method involves 3 steps. • This method inflates information for participants with complete data to account for similar participants with missing data. Hence, infants less likely to have complete data (ie, with low predicted probability of having complete data) but included in the final analysis will be given a large weighting.23,24  
1. First, a logistic regression model calculates each individual’s predicted probability of “nonmissingness,” where the response variable is the nonmissingness and the covariates are the available possible predictors of nonmissingness. • This method helps to reduce selection bias from complete-case analysis. 
2. The weighting for participants with no missing data is determined by the inverse of their predicted probability of nonmissingness. 
3. A weighted regression model is performed that includes only those participants with complete data for analysis. 
Model-based imputation methods These methods are based on statistical models, such as likelihood-based methods, mixed models, Monte Carlo Markov chains, and, more recently, expectation–maximization algorithms and artificial intelligence methods.25 27 Missing data are imputed on the basis of statistical models in 4 steps. • These methods are likely robust and efficient.25,28  
Step 1: The maximum likelihood function estimates parameters on the basis of the nonmissing data. • However, complex algorithms demand strong statistical background and create interpretation difficulty for clinicians. 
Step 2: Missing data are imputed using the maximum likelihood function and on the basis of the estimated parameters identified from Step 1. 
Step 3: Step 1 is reconducted on the basis of the imputed data from Step 2. 
Step 4: Iteration of Steps 1, 2 and 3; that is, repeating all previous steps until no more changes in parameter estimates are observed. 

To illustrate potential differences resulting from a complete-case analysis and several single imputation methods (mean imputation, and worse/best case scenario method) in the presence of missing primary outcome data, we show a hypothetical RCT in Table 3. In this hypothetical example, when primary outcomes are known for all 100 infants, we obtain the true estimate of the treatment effect of a relative risk (RR) of 0.90, with 95% confidence interval (CI) of 0.54–1.49. However, if we assume that 10 of the 50 infants in each group of the trial have missing primary outcome data as described in Table 3, the different methods of analysis produce different results. For example, the best-case scenario method, which assumes that all patients with missing data in the placebo group experience the primary outcome event, yields the most biased and favorable estimate of the treatment effect (RR 0.63, 95% CI 0.37–1.04). In contrast, the worst-case scenario method, which assumes that all patients with missing data in the intervention group would experience the primary outcome event, results in the most biased and unfavorable estimate of the treatment effect (RR 1.79, 95% CI 1.06–3.02). This hypothetical example illustrates how sensitive the results are to the method of handling missing outcome data. The example is not intended to show which method works best. The best way to assess the performance of imputation methods is through simulation based on individual patient data, which could inform the selection of imputation methods in a specific context for handling missing data.3 

TABLE 3

Differences Between the Results of a Complete-Case Analysis and Single Imputation Methods in the Presence of Missing Outcome Data in a Hypothetical RCT

Method of Handling Missing Outcome DataNumber of Events Per Number of Patients With Imputed Outcome for Intervention GroupNumber of Events Per Number of Patients With Imputed Outcome for Placebo GroupTotal Number of Events Per Number of Patients Included for Intervention GroupTotal Number of Events Per Number of Patients Included for Placebo GroupRR (95% CI)
Complete-case analysis — — 15 per 40 14 per 40 1.07 (0.60–1.92) 
Mean imputation method 4 per 10a 4 per 10b 19 per 50 18 per 50 1.06 (0.63–1.76) 
Best-case scenario method 
 None had the event in both groups 0 per 10 0 per 10 15 per 50 14 per 50 1.07 (0.58–1.98) 
 All patients with missing data in placebo group had the event 0 per 10 10 per 10 15 per 50 24 per 50 0.63 (0.37–1.04) 
Worst-case scenario method 
 All patients with missing data in both groups had the event 10 per 10 10 per 10 25 per 50 24 per 50 1.04 (0.70–1.55) 
 All patients with missing data in intervention group had the event 10 per 10 0 per 10 25 per 50 14 per 50 1.79 (1.06–3.02) 
Method of Handling Missing Outcome DataNumber of Events Per Number of Patients With Imputed Outcome for Intervention GroupNumber of Events Per Number of Patients With Imputed Outcome for Placebo GroupTotal Number of Events Per Number of Patients Included for Intervention GroupTotal Number of Events Per Number of Patients Included for Placebo GroupRR (95% CI)
Complete-case analysis — — 15 per 40 14 per 40 1.07 (0.60–1.92) 
Mean imputation method 4 per 10a 4 per 10b 19 per 50 18 per 50 1.06 (0.63–1.76) 
Best-case scenario method 
 None had the event in both groups 0 per 10 0 per 10 15 per 50 14 per 50 1.07 (0.58–1.98) 
 All patients with missing data in placebo group had the event 0 per 10 10 per 10 15 per 50 24 per 50 0.63 (0.37–1.04) 
Worst-case scenario method 
 All patients with missing data in both groups had the event 10 per 10 10 per 10 25 per 50 24 per 50 1.04 (0.70–1.55) 
 All patients with missing data in intervention group had the event 10 per 10 0 per 10 25 per 50 14 per 50 1.79 (1.06–3.02) 

In this hypothetical RCT, 50 patients were randomly assigned to the intervention group, and 50 to the placebo group. The primary outcome event is known to be present or absent for all 100 patients: There were 18 events in the intervention group and 20 in the placebo group. Therefore, the true estimate of the treatment effect is an RR of 0.90 (95% CI 0.54–1.49). However, if we assume that 10 patients have missing outcome data in each group, and further, that among the remaining 80 patients, there are 15 patients with events documented in the intervention group, and 14 patients with events in the placebo group, the results would be as shown in this table. —, no imputation performed.

a

Event number for intervention group was estimated by using the incident rate (15 per 40) multiplied by the number of patients with missing data.

b

Event number for placebo group was estimated by using the incident rate (14 per 40) multiplied by the number of patients with missing data.

The appropriate choice of imputation method for an individual trial depends on the assumptions about the reasons for missing data. However, it is often difficult to decide whether the data are missing at random or missing not at random. Although the mean imputation method may be used to deal with missing baseline covariates in RCTs,29 the multiple imputation and mixed-model methods are preferred solutions to missing at random data because they have shown superiority to single imputation methods in overall performance and reduction of bias.30 Statisticians do not agree on the optimal choices of methods for all circumstances.12,24,28,30,31 Perinatal and neonatal trialists should consult their trial statistician about the best approach. More research is needed to better inform the selection of imputation methods in specific contexts. Indeed, it is strongly recommended to conduct sensitivity analyses using different imputation techniques to assess or support the robustness of the primary approach in clinical trials.3,4 Inevitably, selecting imputation methods involves guesswork. Therefore, we strongly recommend that trialists should devote substantial energy and resources during the design and conduct of the trial to prevent or at least minimize missing primary outcome data.

Among the 38 trials that performed sensitivity analyses in our survey, most did not find differences between the results of sensitivity analyses and those of their primary analysis (Table 1). However, in a trial of skin emollient applied from 2 weeks of age to prevent atopic dermatitis at age 1 year, the risk of atopic dermatitis was significantly increased in the emollient group when multiple imputation was used (risk difference 5.9%, 95% CI 2.0%–9.7%), whereas the primary approach using the best-case imputation method yielded a nonsignificant treatment effect (risk difference 3.1%, 95% CI −0.3% to 6.5%).32 This example supports the importance of performing sensitivity analysis to address missing outcome data.

Evaluating trial quality requires assessment of risk of bias because of missing data as part of the methodological considerations, as in the Cochrane Risk of Bias Tool 2.33 Moreover, reporting guidelines for study protocols and statistical analysis plans mandate that investigators detail how they will handle missing data.34,35 However, the CONsolidated Standards of Reporting Trials guideline does not currently and explicitly state the handling of missing data in reporting of trial results. This anomaly should be reconsidered and incorporated into future CONsolidated Standards of Reporting Trials iterations.

The majority of trials (70 of 77, 91%) with missing primary outcome data in our survey reported use of the intention-to-treat (ITT) principle for their primary analyses. A trial with a complete-case analysis in which missing data are ignored can still claim adherence to an ITT principle if participants are analyzed according to the group to which they are assigned. An ITT principle with a complete-case analysis does not contain any strategy to address potential problems of missing data, or protect against bias from missing primary outcome data. Hence, combining ITT analysis with different data imputation methods is recommended.36,37 

A common practice during trial design is to empirically increase the planned sample size to account for anticipated dropouts. This practice, however, is flawed because it cannot account for the bias that results from missing data, especially from data that are missing not at random. When the bias resulting from missing data is similar to or even larger than the expected intervention effect size, it is impossible to accurately estimate a true intervention effect, even with a large sample size.1 In short, increasing the sample size cannot correct for bias. Therefore, to calculate a sample size, it is suggested to estimate how the intervention effect could become attenuated because of dropout and nonadherence over time, or to generate a sophisticated statistical procedure that accounts for missing data and the associated uncertainty.3 

Study participants or their representatives may withdraw from the trial intervention, refuse further data collection, or even request that already collected data not be included in any analyses. Data withdrawal leads to missing data and little may be known about the reasons for withdrawal, although these are important for exploring the mechanisms of missing data and the potential impact on conclusions.

Although parents’ wishes to withdraw from a trial intervention must be respected, they should always be asked to consider allowing data collection to continue, in particular for the primary outcome. Those data could then be included in the ITT analysis. We have previously provided detailed suggestions about the handling of data withdrawal.38 

The foremost strategy to handle missing data is to prevent missing data, especially during the design and conduct of the trial. Many recommendations stress this.3,4,39,40 However, suggested strategies are rarely, if ever, specific to the unique situation of follow-up in perinatal and neonatal trials in which the participant is the parent–child dyad.

Both perinatal characteristics of the child (including gestational age, birth weight, and medical complications), as well as caregiver characteristics (including sociodemographic status, education, and parental perception of the child’s medical and developmental progress), are associated with likelihood of attrition.41 Parents report structural barriers to follow-up such as travel distance to appointments, problems related to finding child care for siblings, and anxiety surrounding developmental testing of the child.42 

To limit attrition in perinatal and neonatal studies, research teams should engage families in discussion about the importance of long-term follow-up beginning from the first encounter. Study team members should make efforts to establish strong relationships with the parent participants.43 In neonatal trials, trust is further enhanced when study team members understand both the research study and neonatal care practices and unit culture. Frequent interactions (including phone calls and birthday or holiday cards) provide opportunities to maintain these relationships while also updating family contact information. Additional strategies include flexible study visit scheduling (including evenings and weekends), free parking or transportation, availability of home study visits, child care for siblings, and provision of small age- and developmental status-appropriate gifts for the child, in addition to fair compensation for time spent attending the visit for the guardian. In study design phases, investigators can engage representative families in planning the protocol and selection of outcomes and measures, perform feasibility assessments to understand potential reasons for withdrawal, allow flexible treatment regimens, avoid outcome measures with likelihood of low response rates, and select assessments that provide relevant and helpful information for families about their children’s medical or developmental progress. In today’s era of telemedicine, investigators could also consider validated virtual assessments to evaluate outcomes.44,45 Finally, as study results become available, these must be shared with families at a sixth- to eighth-grade reading level. Such feedback provides opportunities for investigators to provide information, maintain relationships, and express gratitude to study participants.

These suggested strategies are summarized in a Population, Intervention, Control, Outcome, Time frame, Study design, and Analysis format in Table 4, aiming to enhance prompt understanding for readers.

TABLE 4

Strategies to Prevent or Minimize Missing Data in Perinatal and Neonatal Trials

Key ElementStrategies
Population • Educate families about the importance of collecting complete outcome data at the time of enrollment. 
Intervention/control • Attempt simple and easy access to interventions to minimize participants’ burden and inconvenience. 
Outcome • Engage families in selection of outcome measures. 
• Ensure parents see potential value of follow-up for their child; for example, information that may assist specific educational needs. 
• Choose assessments that provide relevant and helpful information for families about their children’s medical or developmental progress. 
• Avoid outcome measures with a high probability of low response rates, such as invasive tests. 
• Use “hard” outcomes already recorded in charts or administrative databases. 
• Consider validated virtual assessments for outcome measures. 
Time frame • Provide flexible study visit scheduling including evenings and weekends. 
Study design • Include parents and/or community members as advisors for study design and conduct. 
• Perform feasibility research before the main trial. 
• Randomize participants as immediately as possible before initiation of intervention. 
• Include stakeholder groups (such as local professional entities) where appropriate. 
Analysis • Set targets for missing data and solutions if not met. 
• Plan sensitivity analysis in trial protocol or statistical analysis plan. 
Others • Update contact information for participants or their family members in time. 
• Reduce inconvenience of participation or follow-up visit by providing parking service, child care, and fair compensation. 
• Share study results with families at an appropriate reading level. 
• Provide home visits or appointments at nonstandard times (weekends or evenings) for data collection. 
• Hire experienced study coordinators, including staff with clinical experience. 
• Follow relevant reporting guidelines. 
Key ElementStrategies
Population • Educate families about the importance of collecting complete outcome data at the time of enrollment. 
Intervention/control • Attempt simple and easy access to interventions to minimize participants’ burden and inconvenience. 
Outcome • Engage families in selection of outcome measures. 
• Ensure parents see potential value of follow-up for their child; for example, information that may assist specific educational needs. 
• Choose assessments that provide relevant and helpful information for families about their children’s medical or developmental progress. 
• Avoid outcome measures with a high probability of low response rates, such as invasive tests. 
• Use “hard” outcomes already recorded in charts or administrative databases. 
• Consider validated virtual assessments for outcome measures. 
Time frame • Provide flexible study visit scheduling including evenings and weekends. 
Study design • Include parents and/or community members as advisors for study design and conduct. 
• Perform feasibility research before the main trial. 
• Randomize participants as immediately as possible before initiation of intervention. 
• Include stakeholder groups (such as local professional entities) where appropriate. 
Analysis • Set targets for missing data and solutions if not met. 
• Plan sensitivity analysis in trial protocol or statistical analysis plan. 
Others • Update contact information for participants or their family members in time. 
• Reduce inconvenience of participation or follow-up visit by providing parking service, child care, and fair compensation. 
• Share study results with families at an appropriate reading level. 
• Provide home visits or appointments at nonstandard times (weekends or evenings) for data collection. 
• Hire experienced study coordinators, including staff with clinical experience. 
• Follow relevant reporting guidelines. 

The handling of missing primary outcome data has been largely inadequate in recent perinatal and neonatal trials. This is concerning because missing data may result in misleading inferences about treatment effects. Although carefully conducted imputations of missing data are a necessary component of the analysis of clinical trials, preventing missing data as much as possible remains the best approach to reduce bias from attrition.

Drs Li and Meng and Ms Liu and Ms Zhang conceived and designed this research, acquired data, ran statistical analyses, drafted the manuscript and shared first authorship equally; Drs Schmidt, Kirpalani, and Thabane conceived and designed the study, provided professional support, made multiple critical revisions to the manuscript and shared senior authorship equally; Drs DeMauro and Mbuagbaw provided statistical and professional support, and made multiple substantial revisions to the manuscript; and all authors approved the final manuscript as submitted and agree to be accountable for all aspects of the work.

COMPANION PAPER: A companion to this article can be found online at www.pediatrics.org/cgi/doi/10.1542/peds.2023-064938.

FUNDING: No external funding.

CONFLICT OF INTEREST DISCLOSURES: The authors have indicated they have no conflicts of interest to disclose.

CI

confidence interval

ITT

intention to treat

RCT

randomized controlled trial

RR

relative risk

1
Little
RJ
,
D’Agostino
R
,
Cohen
ML
, et al
.
The prevention and treatment of missing data in clinical trials
.
N Engl J Med
.
2012
;
367
(
14
):
1355
1360
2
McCann
ME
,
de Graaff
JC
,
Dorris
L
, et al
.
GAS Consortium
.
Neurodevelopmental outcome at 5 years of age after general anesthesia or awake–regional anesthesia in infancy (GAS): an international, multicenter, randomized, controlled equivalence trial
.
Lancet
.
2019
;
393
(
10172
):
664
677
3
National Research Council Panel on Handling Missing Data in Clinical Trials
.
The Prevention and Treatment of Missing Data in Clinical Trials
.
National Academies Press (US)
;
2010
4
European Medicines Agency
.
Guideline on missing data in confirmatory clinical trials
.
2011
:
1
12
5
Austin
PC
,
White
IR
,
Lee
DS
,
van Buuren
S
.
Missing data in clinical research: a tutorial on multiple imputation
.
Can J Cardiol
.
2021
;
37
(
9
):
1322
1331
6
Altman
DG
.
Missing outcomes in randomized trials: addressing the dilemma
.
Open Med
.
2009
;
3
(
2
):
e51
e53
7
Marino
M
,
Lucas
J
,
Latour
E
,
Heintzman
JD
.
Missing data in primary care research: importance, implications and approaches
.
Fam Pract
.
2021
;
38
(
2
):
200
203
8
Ren
Y
,
Jia
Y
,
Huang
Y
, et al
.
Missing data were poorly reported and handled in randomized controlled trials with repeatedly measured continuous outcomes: a cross-sectional survey
.
J Clin Epidemiol
.
2022
;
148
:
27
38
9
Bell
ML
,
Fiero
M
,
Horton
NJ
,
Hsu
CH
.
Handling missing data in RCTs; a review of the top medical journals
.
BMC Med Res Methodol
.
2014
;
14
:
118
10
Little
RJA
,
Rubin
DB
.
Statistical Analysis With Missing Data
, 2nd ed.
Wiley
;
2002
11
Sterne
JA
,
White
IR
,
Carlin
JB
, et al
.
Multiple imputation for missing data in epidemiological and clinical research: potential and pitfalls
.
BMJ
.
2009
;
338
:
b2393
12
Bell
ML
,
Fairclough
DL
.
Practical and statistical issues in missing data for longitudinal patient-reported outcomes
.
Stat Methods Med Res
.
2014
;
23
(
5
):
440
459
13
Cro
S
,
Morris
TP
,
Kenward
MG
,
Carpenter
JR
.
Sensitivity analysis for clinical trials with missing continuous outcome data using controlled multiple imputation: a practical guide
.
Stat Med
.
2020
;
39
(
21
):
2815
2842
14
Lurie
I
,
Levine
SZ
.
Meta-analysis of dropout rates in SSRIs versus placebo in randomized clinical trials of PTSD
.
J Nerv Ment Dis
.
2010
;
198
(
2
):
116
124
15
Rabinowitz
J
,
Levine
SZ
,
Barkai
O
,
Davidov
O
.
Dropout rates in randomized clinical trials of antipsychotics: a meta-analysis comparing first- and second-generation drugs and an examination of the role of trial design features
.
Schizophr Bull
.
2009
;
35
(
4
):
775
788
16
Lipinski
MJ
,
Cauthen
CA
,
Biondi-Zoccai
GG
, et al
.
Meta-analysis of randomized controlled trials of statins versus placebo in patients with heart failure
.
Am J Cardiol
.
2009
;
104
(
12
):
1708
1716
17
Raboud
JM
,
Montaner
JS
,
Thorne
A
,
Singer
J
,
Schechter
MT
.
Impact of missing data due to dropouts on estimates of the treatment effect in a randomized trial of antiretroviral therapy for HIV-infected individuals. Canadian HIV Trials Network A002 Study Group
.
J Acquir Immune Defic Syndr Hum Retrovirol
.
1996
;
12
(
1
):
46
55
18
Akl
EA
,
Briel
M
,
You
JJ
, et al
.
Potential impact on estimated treatment effects of information lost to follow-up in randomized controlled trials (LOST-IT): systematic review
.
BMJ
.
2012
;
344
:
e2809
19
Schafer
JL
.
Multiple imputation: a primer
.
Stat Methods Med Res
.
1999
;
8
(
1
):
3
15
20
Lachin
JM
.
Fallacies of last observation carried forward analyses
.
Clin Trials
.
2016
;
13
(
2
):
161
168
21
Papageorgiou
G
,
Grant
SW
,
Takkenberg
JJM
,
Mokhles
MM
.
Statistical primer: how to deal with missing data in scientific research?
Interact Cardiovasc Thorac Surg
.
2018
;
27
(
2
):
153
158
22
de Goeij
MC
,
van Diepen
M
,
Jager
KJ
,
Tripepi
G
,
Zoccali
C
,
Dekker
FW
.
Multiple imputation: dealing with missing data
.
Nephrol Dial Transplant
.
2013
;
28
(
10
):
2415
2420
23
Metten
MA
,
Costet
N
,
Multigner
L
,
Viel
JF
,
Chauvet
G
.
Inverse probability weighting to handle attrition in cohort studies: some guidance and a call for caution
.
BMC Med Res Methodol
.
2022
;
22
(
1
):
45
24
Seaman
SR
,
White
IR
.
Review of inverse probability weighting for dealing with missing data
.
Stat Methods Med Res
.
2013
;
22
(
3
):
278
295
25
Ratitch
B
,
O’Kelly
M
,
Tosiello
R
.
Missing data in clinical trials: from clinical assumptions to statistical analysis using pattern mixture models
.
Pharm Stat
.
2013
;
12
(
6
):
337
347
26
Enders
CK
,
Du
H
,
Keller
BT
.
A model-based imputation procedure for multilevel regression models with random coefficients, interaction effects, and nonlinear terms
.
Psychol Methods
.
2020
;
25
(
1
):
88
112
27
Kim
S
,
Sugar
CA
,
Belin
TR
.
Evaluating model-based imputation methods for missing covariates in regression models with interactions
.
Stat Med
.
2015
;
34
(
11
):
1876
1888
28
Tseng
CH
,
Chen
YH
.
Regularized approach for data missing not at random
.
Stat Methods Med Res
.
2019
;
28
(
1
):
134
150
29
White
IR
,
Thompson
SG
.
Adjusting for partially missing baseline measurements in randomized trials
.
Stat Med
.
2005
;
24
(
7
):
993
1007
30
Zhang
Y
,
Alyass
A
,
Vanniyasingam
T
, et al
.
A systematic survey of the methods literature on the reporting quality and optimal methods of handling participants with missing outcome data for continuous outcomes in randomized controlled trials
.
J Clin Epidemiol
.
2017
;
88
:
67
80
31
Sullivan
TR
,
White
IR
,
Salter
AB
,
Ryan
P
,
Lee
KJ
.
Should multiple imputation be the method of choice for handling missing data in randomized trials?
Stat Methods Med Res
.
2018
;
27
(
9
):
2610
2626
32
Skjerven
HO
,
Rehbinder
EM
,
Vettukattil
R
, et al
.
Skin emollient and early complementary feeding to prevent infant atopic dermatitis (PreventADALL): a factorial, multicenter, cluster-randomized trial
.
Lancet
.
2020
;
395
(
10228
):
951
961
33
Higgins
JP
,
Savović
J
,
Page
MJ
,
Elbers
RG
,
Sterne
JA
.
Cochrane Handbook for Systematic Reviews of Interventions Version 6.3. Chapter 8: assessing risk of bias in a randomized trial
. Available at: https://training.cochrane.org/handbook/current/chapter-08
34
Chan
AW
,
Tetzlaff
JM
,
Gøtzsche
PC
, et al
.
SPIRIT 2013 explanation and elaboration: guidance for protocols of clinical trials
.
BMJ
.
2013
;
346
:
e7586
35
Gamble
C
,
Krishan
A
,
Stocken
D
, et al
.
Guidelines for the content of statistical analysis plans in clinical trials
.
JAMA
.
2017
;
318
(
23
):
2337
2343
36
Alshurafa
M
,
Briel
M
,
Akl
EA
, et al
.
Inconsistent definitions for intention-to-treat in relation to missing outcome data: systematic review of the methods literature
.
PLoS One
.
2012
;
7
(
11
):
e49163
37
White
IR
,
Carpenter
J
,
Horton
NJ
.
Including all individuals is not enough: lessons for intention-to-treat analysis
.
Clin Trials
.
2012
;
9
(
4
):
396
407
38
Ye
C
,
Giangregorio
L
,
Holbrook
A
,
Pullenayegum
E
,
Goldsmith
CH
,
Thabane
L
.
Data withdrawal in randomized controlled trials: defining the problem and proposing solutions: a commentary
.
Contemp Clin Trials
.
2011
;
32
(
3
):
318
322
39
Dziura
JD
,
Post
LA
,
Zhao
Q
,
Fu
Z
,
Peduzzi
P
.
Strategies for dealing with missing data in clinical trials: from design to analysis
.
Yale J Biol Med
.
2013
;
86
(
3
):
343
358
40
Fleming
TR
.
Addressing missing data in clinical trials
.
Ann Intern Med
.
2011
;
154
(
2
):
113
117
41
DeMauro
SB
,
Bellamy
SL
,
Fernando
M
,
Hoffmann
J
,
Gratton
T
,
Schmidt
B
;
PROP Investigators
.
Patient, family, and center-based factors associated with attrition in neonatal clinical research: a prospective study
.
Neonatology
.
2019
;
115
(
4
):
328
334
42
Brady
JM
,
Pouppirt
N
,
Bernbaum
J
, et al
.
Why do children with severe bronchopulmonary dysplasia not attend neonatal follow-up care? Parental views of barriers
.
Acta Paediatr
.
2018
;
107
(
6
):
996
1002
43
DeMauro
SB
,
Cairnie
J
,
D’Ilario
J
,
Kirpalani
H
,
Schmidt
B
.
Honesty, trust, and respect during consent discussions in neonatal clinical trials
.
Pediatrics
.
2014
;
134
(
1
):
e1
e3
44
DeMauro
SB
,
Duncan
AF
,
Hurt
H
.
Telemedicine use in neonatal follow-up programs–what can we do and what we can’t–lessons learned from COVID-19
.
Semin Perinatol
.
2021
;
45
(
5
):
151430
45
Haffner
DN
,
Bauer Huang
SL
.
Using telemedicine to overcome barriers to neurodevelopmental care from the neonatal intensive care unit to school entry
.
Clin Perinatol
.
2023
;
50
(
1
):
253
268

Supplementary data