Docs/Blinded SSR

Blinded Sample Size Re-estimation (SSR)

Technical documentation for blinded interim sample size re-estimation in adaptive clinical trials. This page covers nuisance parameter re-estimation from pooled data, the Kieser–Friede framework, conditional power, sensitivity analysis, regulatory alignment, and Monte Carlo validation.

1. Overview & Motivation

A clinical trial's initial sample size depends on assumptions about nuisance parameters—the variance for continuous endpoints or the pooled event rate for binary endpoints. If these assumptions are wrong, the trial may be underpowered or wastefully overpowered.

Blinded sample size re-estimation (SSR) addresses this by re-estimating nuisance parameters from pooled (blinded) interim data and adjusting the target sample size accordingly. Because treatment assignments are never revealed, blinding is preserved and Type I error is controlled under mild regularity conditions.

Key advantage: The FDA classifies blinded SSR as a “well-understood” adaptation that does not require complex statistical methodology for Type I error control (FDA Guidance 2019, Section IV.B.2). This makes it the lowest-risk adaptive design option.

2. Theoretical Foundation

Continuous Endpoints

For a two-arm trial comparing means with common variance σ2\sigma^2 and treatment effect δ\delta, the required sample size per arm is:

n=2(zα+zβ)2σ2δ2n = \left\lceil \frac{2(z_\alpha + z_\beta)^2 \sigma^2}{\delta^2} \right\rceil

where zα=Φ1(1α)z_\alpha = \Phi^{-1}(1-\alpha) and zβ=Φ1(power)z_\beta = \Phi^{-1}(\text{power}). The treatment effect δ\delta is assumed known (the minimally clinically important difference), while σ2\sigma^2 is the nuisance parameter to be re-estimated.

At the interim analysis with information fraction tt, the pooled sample variance is computed from all ninterim=t2nn_{\text{interim}} = \lceil t \cdot 2n \rceil patients without unblinding:

σ^blinded2=1ninterim1i=1ninterim(XiXˉ)2\hat{\sigma}^2_{\text{blinded}} = \frac{1}{n_{\text{interim}}-1} \sum_{i=1}^{n_{\text{interim}}} (X_i - \bar{X})^2

This estimate is positively biased by δ2/4\delta^2/4 (Kieser & Friede 2003), but the bias is conservative—it leads to slight overestimation of the required sample size, which protects power.

Binary Endpoints

For binary endpoints with control rate pCp_C and treatment rate pTp_T, the sample size per arm uses the pooled-variance normal approximation (Fleiss, Levin & Paik):

n=(zα2pˉ(1pˉ)+zβpC(1pC)+pT(1pT)pTpC)2n = \left\lceil \left( \frac{z_\alpha \sqrt{2\bar{p}(1-\bar{p})} + z_\beta \sqrt{p_C(1-p_C) + p_T(1-p_T)}}{|p_T - p_C|} \right)^2 \right\rceil

where pˉ=(pC+pT)/2\bar{p} = (p_C + p_T)/2 is the pooled rate. At interim, the observed blinded pooled rate pˉ^\hat{\bar{p}} replaces the planned pˉ\bar{p} while the planned effect size pTpC|p_T - p_C| is maintained (Friede & Kieser 2004).

Survival (Time-to-Event) Endpoints

For event-driven trials comparing survival curves with hazard ratio HR\text{HR} and allocation ratio rr (treatment:control), the required number of events is given by the Schoenfeld (1981) formula:

d=(zα+zβ)2(1+r)2rlog2(HR)d = \left\lceil \frac{(z_\alpha + z_\beta)^2 (1+r)^2}{r \cdot \log^2(\text{HR})} \right\rceil

The number of events dd is the primary quantity. To convert to a sample size NN, compute the expected event probability under an exponential model with uniform accrual over TaT_a months and total study duration TT:

λC=ln2mC,Pˉ(event)=1exp ⁣(λC ⁣(TTa2))\lambda_C = \frac{\ln 2}{m_C}, \quad \bar{P}(\text{event}) = 1 - \exp\!\left(-\lambda_C \!\left(T - \tfrac{T_a}{2}\right)\right)
N=d/Pˉ(event)N = \left\lceil d \,/\, \bar{P}(\text{event}) \right\rceil

where mCm_C is the median survival in the control arm. At interim, the blinded pooled event rate P^\hat{P} is re-estimated from observed events across both arms. Because events dd are fixed, only NN changes.

Why events, not patients? In time-to-event designs, statistical power depends on the number of events, not the number of patients enrolled. The Schoenfeld formula fixes dd; the sample size NN is derived and can be updated at interim without affecting the event target.

Why only nuisance parameters? Blinded data cannot distinguish treatment from control, so only parameters that are estimable from pooled data can be updated. The treatment effect δ\delta (or HR\text{HR} for survival) remains fixed at the planned value.

3. Recalculation Algorithm

The blinded SSR procedure follows these steps:

1

Initial Sample Size

Compute N0=2nN_0 = 2n using the planned nuisance parameter (σ02\sigma^2_0 or pˉ0\bar{p}_0). For survival, compute dd events via Schoenfeld, then N0=d/Pˉ(event)N_0 = \lceil d / \bar{P}(\text{event}) \rceil.

2

Interim Look

At information fraction tt, ninterim=tN0n_{\text{interim}} = \lceil t \cdot N_0 \rceil patients have been enrolled.

3

Blinded Re-estimation

Compute σ^2\hat{\sigma}^2 (or pˉ^\hat{\bar{p}}) from pooled interim data without breaking the blind. For survival, re-estimate the blinded pooled event rate P^\hat{P} from observed events.

4

Recalculate

Recompute the required sample size N1N_1 using the re-estimated nuisance parameter while keeping the planned effect size fixed. For survival, N1=d/P^N_1 = \lceil d / \hat{P} \rceil — events dd stays fixed.

5

Constrain

Apply the interim floor (N1ninterimN_1 \geq n_{\text{interim}}, cannot un-enroll patients) and the protocol cap (N1N0fmaxN_1 \leq N_0 \cdot f_{\max}, where fmaxf_{\max} is the maximum inflation factor, typically 1.5–2.0). Enforce even parity for equal allocation.

6

Continue

Enroll remaining patients to the adjusted target and perform the final analysis using the standard z-test.

Constraint priority: Interim floor > Cap > Even parity. When the cap is binding, per-arm count is rounded down to respect the protocol limit. When the interim floor is binding, per-arm count is rounded up to accommodate already-enrolled patients.

4. Conditional Power

After re-estimation, the conditional power quantifies the probability of rejecting H0H_0 at the final analysis, given the interim data and the adjusted sample size:

CP=Φ(z1RzαR1)\text{CP} = \Phi\left( z_1 \sqrt{R} - z_\alpha \sqrt{R - 1} \right)

where R=Nfinal/ninterimR = N_{\text{final}} / n_{\text{interim}} is the ratio of final to interim sample size and z1z_1 is the expected interim z-statistic under the assumed treatment effect:

Continuous

z1=δ2σ^2/narm,interimz_1 = \frac{\delta}{\sqrt{2\hat{\sigma}^2 / n_{\text{arm,interim}}}}

Binary

z1=pTpC2pˉ(1pˉ)/narm,interimz_1 = \frac{|p_T - p_C|}{\sqrt{2\bar{p}(1-\bar{p}) / n_{\text{arm,interim}}}}

Survival

z1=log(HR)dinterimr(1+r)2z_1 = -\log(\text{HR}) \sqrt{\frac{d_{\text{interim}} \cdot r}{(1+r)^2}}

Survival note: For time-to-event endpoints, the information ratio RR uses events: R=dtotal/dinterimR = d_{\text{total}} / d_{\text{interim}} (not Ntotal/NinterimN_{\text{total}} / N_{\text{interim}}).

Edge cases: When Nfinal=ninterimN_{\text{final}} = n_{\text{interim}} (no additional recruitment, e.g., after the interim floor binds), CP reduces to Φ(z1zα)\Phi(z_1 - z_\alpha)—the probability of rejection based on the interim z alone.

5. Sensitivity Analysis

The calculator generates a sensitivity table showing how the recalculated sample size responds to different nuisance parameter values. This supports protocol planning and DMC communication.

Continuous Endpoints

The observed variance is varied as multiples of the planned variance: 50%, 75%, 100% (planned), 125%, 150%, and 200%. For each scenario, the recalculated N, inflation factor, and conditional power are reported.

Binary Endpoints

The pooled event rate is varied in offsets of −0.10, −0.05, 0.00 (planned), +0.05, and +0.10, clamped to [0.01, 0.99]. Each scenario reports the resulting sample size under the updated nuisance parameter.

Survival Endpoints

The observed event rate is varied as multiples of the planned event probability: 0.5×, 0.75×, 1.0× (planned), 1.25×, 1.5×, and 2.0×. Lower event rates require more patients to achieve the same number of events, so N increases.

Protocol tip: Include the sensitivity table in the SAP appendix to demonstrate that the pre-specified cap accommodates plausible variance inflation scenarios.

6. Statistical Assumptions

Allocation: 1:1 randomization by default for continuous and binary endpoints. Survival endpoints support unequal allocation via rr (treatment:control ratio).

Common variance: For continuous endpoints, the variance is assumed equal across both arms (homoscedasticity).

Fixed effect size: The treatment effect δ\delta, pTpC|p_T - p_C|, or HR\text{HR} remains at the planned value—only the nuisance parameter is re-estimated.

Blinding preserved: The re-estimation uses only pooled data. Treatment assignments are not accessed.

Normal approximation: The final test is a z-test (continuous/binary) or the logrank test (survival). For binary endpoints, adequate sample size per arm is needed for the normal approximation to hold (typically np5n \cdot p \geq 5 and n(1p)5n \cdot (1-p) \geq 5).

Exponential survival model: For survival endpoints, event times follow an exponential distribution (proportional hazards) with uniform accrual over the accrual period.

Single interim look: One pre-specified interim analysis for re-estimation. Multiple re-estimations require additional considerations.

7. Limitations & When Not to Use

Blinded variance bias: The pooled variance σ^blinded2\hat{\sigma}^2_{\text{blinded}} is biased upward by δ2/4\delta^2/4. This is conservative but may lead to unnecessary sample size increases when the true effect is large.

Early interim risk: If t<0.25t < 0.25, the variance estimate may be unstable due to small sample size. The calculator enforces t[0.1,0.9]t \in [0.1, 0.9].

Cannot detect effect size misspecification: Blinded SSR only re-estimates nuisance parameters. If the treatment effect was overestimated at planning, the trial may still be underpowered. Consider unblinded SSR if effect size uncertainty is the primary concern.

Exponential model assumption: Survival SSR assumes exponential event times and proportional hazards. Non-proportional hazards, cure-rate models, or complex censoring patterns may require external simulation.

Unequal allocation (continuous/binary): Unequal allocation ratios are supported for survival endpoints only. Continuous and binary endpoints assume 1:1 randomization.

8. Regulatory Considerations

Blinded SSR is one of the most regulatory-friendly adaptive designs. The FDA Guidance on Adaptive Designs (2019) explicitly acknowledges that blinded re-estimation preserves Type I error under standard conditions.

Documentation Checklist

Pre-specify the interim timing (information fraction) and the re-estimation procedure in the protocol and SAP.

Define the maximum sample size cap (fmaxf_{\max}) and justify it based on feasibility and budget constraints.

Specify that only nuisance parameters (variance or pooled rate) will be re-estimated; the treatment effect remains fixed.

Include the sensitivity analysis table showing recalculated N under various nuisance parameter scenarios.

Confirm that blinding is maintained during the re-estimation and that the independent statistician/DMC oversees the process.

Automated Warnings

The calculator generates context-specific regulatory notes:

Substantial increase (>50%): Flags impact on trial feasibility, budget, and timeline.

Cap binding: Notes that conditional power may fall below target; cap justification required in the protocol.

Deflated estimate: When the re-estimated size is notably smaller (<80% of planned), warns about potential early high-responder cohort bias.

9. Monte Carlo Validation

The calculator supports Tier 2 simulation validation through the Adaptive Core engine. Monte Carlo simulations independently verify the analytical results by:

  1. Generating interim data under the true parameters
  2. Computing the blinded variance (or pooled rate) estimate
  3. Recalculating the sample size with the same algorithm
  4. Generating remaining data to the adjusted target
  5. Performing the final z-test
  6. Repeating 1,000–100,000 times

Reported Metrics

Type I Error

Rejection rate under H0H_0 (true effect = 0). Should be α\leq \alpha.

Empirical Power

Rejection rate under H1H_1. Should approximate the analytical conditional power.

Final N Distribution

Mean, median, Q25, Q75, min, and max of the final sample size across simulations.

Discordance Check

If simulated power deviates from analytical by >3%, a warning is raised.

Reproducibility: Every simulation run is seeded (via numpy.random.Generator\texttt{numpy.random.Generator}) and the seed is stored alongside results. Re-running with the same seed produces identical output.

10. API Reference

POST /api/v1/calculators/ssr-blinded

Computes blinded sample size re-estimation with optional Monte Carlo simulation validation.

Request Parameters

ParameterTypeDefaultDescription
endpoint_typestring"continuous""continuous", "binary", or "survival"
alphafloat0.025One-sided significance level (0, 1)
powerfloat0.90Target power (0.5, 1)
mean_differencefloat0.3Treatment effect for continuous endpoints (>0)
initial_variancefloat1.0Planned variance for continuous endpoints (>0)
control_ratefloat?nullControl arm event rate for binary endpoints (0, 1)
treatment_ratefloat?nullTreatment arm event rate (0, 1); must be > control_rate
interim_fractionfloat0.5Information fraction at interim (0.1, 0.9)
n_max_factorfloat2.0Maximum inflation factor [1.0, 5.0]
observed_variancefloat?nullObserved blinded variance (continuous, >0)
observed_pooled_ratefloat?nullObserved blinded pooled rate (binary, (0,1))
hazard_ratiofloat?nullAssumed HR (survival, (0,2), ≠1)
median_controlfloat?nullMedian control survival in months (>0)
accrual_timefloat?nullAccrual period in months (>0)
follow_up_timefloat?nullFollow-up after accrual in months (≥0)
dropout_ratefloat0.0Annual dropout rate [0, 1)
allocation_ratiofloat1.0Randomization ratio treatment:control (>0, survival)
observed_event_ratefloat?nullBlinded pooled event rate at interim (survival, (0,1))
simulateboolfalseEnable Monte Carlo simulation tier
simulation_seedint?nullSeed for reproducibility; auto-generated if omitted
n_simulationsint10000Number of simulations [1000, 100000]

Example Request

{
  "endpoint_type": "continuous",
  "alpha": 0.025,
  "power": 0.90,
  "mean_difference": 0.3,
  "initial_variance": 1.0,
  "interim_fraction": 0.5,
  "n_max_factor": 2.0,
  "observed_variance": 1.4,
  "simulate": true,
  "n_simulations": 10000
}

Example Request (Survival Endpoint)

{
  "endpoint_type": "survival",
  "alpha": 0.025,
  "power": 0.90,
  "hazard_ratio": 0.7,
  "median_control": 12,
  "accrual_time": 24,
  "follow_up_time": 12,
  "allocation_ratio": 1.0,
  "interim_fraction": 0.5,
  "n_max_factor": 2.0,
  "observed_event_rate": 0.55,
  "simulate": true,
  "n_simulations": 10000
}

Response Fields

FieldDescription
initial_n_per_armSample size per arm before re-estimation
recalculated_n_per_armSample size per arm after re-estimation
inflation_factorRatio of recalculated to initial total N
conditional_powerConditional power at the adjusted sample size
n_cappedWhether the cap was binding
recalculation_scenariosSensitivity table with 5–6 scenarios
regulatory_notesContext-specific regulatory guidance
events_requiredRequired events d from Schoenfeld formula (survival only)
event_probabilityPlanned weighted event probability (survival only)
observed_event_probabilityRe-estimated event probability at interim (survival only)
initial_n_controlInitial control arm N (survival with allocation_ratio)
initial_n_treatmentInitial treatment arm N (survival with allocation_ratio)
recalculated_n_controlRecalculated control arm N (survival)
recalculated_n_treatmentRecalculated treatment arm N (survival)

11. Technical References

  1. Kieser M, Friede T. Simple procedures for blinded sample size adjustment that do not affect the type I error rate. Statistics in Medicine. 2003;22(23):3571–3581.
  2. Friede T, Kieser M. Sample size recalculation for binary data in internal pilot study designs. Pharmaceutical Statistics. 2004;3(4):269–279.
  3. Zucker DM, Wittes JT. The bias and efficiency of blinded variance estimates in clinical trials. Statistics in Medicine. 2004;23(4):565–574.
  4. FDA. Adaptive Designs for Clinical Trials of Drugs and Biologics: Guidance for Industry. 2019. Section IV.B.2.
  5. EMA. Reflection Paper on Methodological Issues in Confirmatory Clinical Trials Planned with an Adaptive Design. CHMP/EWP/2459/02. 2007.
  6. Cui L, Hung HMJ, Wang SJ. Modification of sample size in group sequential clinical trials. Biometrics. 1999;55(3):853–857.
  7. Fleiss JL, Levin B, Paik MC. Statistical Methods for Rates and Proportions. 3rd ed. Wiley; 2003.
  8. Schoenfeld D. The asymptotic properties of nonparametric tests for comparing survival distributions. Biometrika. 1981;68(1):316–319.
  9. Gould AL. Interim analyses for monitoring clinical trials that do not materially affect the type I error rate. Statistics in Medicine. 1992;11(1):55–66.
  10. Friede T, et al. Blinded sample size re-estimation in event-driven clinical trials. Pharmaceutical Statistics. 2019;18(5):578–588.