Feasibility is the single most consequential phase of study start-up. Get it right and your trial enrols on time, your sites are engaged, and your programme stays on budget. Get it wrong and you inherit months of delays, underperforming investigators, and costly protocol amendments that could have been avoided.

Yet feasibility is also the phase most sponsors rush. The pressure to activate sites quickly — driven by milestone commitments, investor expectations, or competitive enrolment — often leads to shortcuts that compound downstream. This article lays out how to do feasibility properly, what data actually matters for site selection in the UK, and where the common pitfalls catch even experienced sponsors.

Why Feasibility Matters More Than You Think

Clinical trial failure rates are staggering. Roughly 80% of trials fail to meet their original enrolment timelines. The root cause is rarely the protocol itself — it is almost always a mismatch between the study's requirements and the sites selected to execute it.

Feasibility is the process that prevents this mismatch. It is the structured assessment of whether a proposed study can realistically recruit its target population within the planned timeline and budget, at the specific sites under consideration. Done well, it identifies the optimal number and mix of sites, flags protocol elements that will slow enrolment, and gives sponsors evidence-based confidence in their activation plan.

Poor feasibility does not just delay enrolment. It forces protocol amendments, burns investigator goodwill, and can undermine the entire development programme's credibility with regulators and investors.

The Two Phases of Feasibility

Effective feasibility splits into two distinct phases, each with different objectives and methods.

Phase 1: Strategic Feasibility

Before you identify a single site, strategic feasibility asks the fundamental questions:

Strategic feasibility is a sponsor responsibility. It cannot be delegated entirely to a CRO because it requires alignment between clinical development strategy, commercial objectives, and operational reality. The best CROs — and the ones worth partnering with — will challenge your assumptions here rather than simply confirming what you want to hear.

Phase 2: Operational Feasibility and Site Selection

Once strategic feasibility confirms the study is viable, operational feasibility identifies which specific sites can deliver. This is where most of the practical work happens.

What Actually Predicts Site Performance

The single most reliable predictor of future site performance is past site performance. Not the PI's reputation, not the site's facilities, not the size of the institution — actual historical enrolment data for studies with similar patient populations and complexity.

Key data points that matter for UK site selection:

Red Flag

If a site's feasibility response promises enrolment numbers that significantly exceed their historical performance on similar protocols, treat it as a risk, not a reason to celebrate. Ask for the data behind the estimate.

The UK Site Landscape

UK clinical trials benefit from the NIHR Clinical Research Network (CRN), which provides infrastructure and support across England's 15 local CRNs. This network means sponsors have access to a structured site ecosystem that is unusual in global clinical research.

However, the UK landscape also has specific characteristics that feasibility must account for:

Common Feasibility Mistakes

After years of clinical programme delivery, these are the mistakes we see repeatedly:

1. Selecting Too Many Sites

The instinct when worried about enrolment is to add more sites. This usually makes things worse. More sites mean more startup costs, more monitoring visits, more TMF complexity, and more variation in data quality. The marginal return of each additional site diminishes rapidly after the first tranche of high-performers.

Better approach: Start with fewer, carefully selected sites based on data. Add sites only if the first tranche is demonstrably underperforming against realistic targets.

2. Ignoring Site-Level Operational Burden

Feasibility questionnaires focus on patient availability but rarely assess whether the site's operational team can handle the protocol's visit schedule, source data requirements, and safety reporting obligations. A site that can recruit patients but cannot manage the operational workload will generate deviations, missing data, and queries.

3. Over-Relying on PI Relationships

Key opinion leader PIs bring credibility and expertise, but they are also the busiest people in the hospital. A PI who is enthusiastic at the investigator meeting but unavailable for screening visits three months later is a well-known pattern. Feasibility must assess the entire site team, not just the PI.

4. Not Accounting for Competing Studies

In热门 therapeutic areas, a site may be running five or more competing protocols for the same patient population. Unless feasibility accounts for the site's complete trial portfolio, enrolment estimates will be unreliable.

5. Rushing the Process

Feasibility takes time. Proper site assessment, historical data review, and PI interviews cannot be compressed into a two-week window without sacrificing quality. The irony is that rushing feasibility to start faster almost always results in starting later.

A Data-Driven Approach to Site Selection

The best feasibility processes combine multiple data sources rather than relying on any single input:

Feasibility as a Continuous Process

Feasibility does not end when sites are selected. The most effective sponsors treat feasibility as an ongoing discipline throughout the trial:

DEOX Approach

We run feasibility as a structured, data-driven process — not a checkbox exercise. Our UK site network knowledge, combined with AI-assisted analytics and honest PI relationships, means our feasibility predictions are grounded in evidence rather than optimism. We would rather tell you a study needs 12 sites and be right than promise 6 and be wrong.

Key Takeaways

Getting feasibility right is not complicated, but it does require discipline, honest data, and the willingness to challenge assumptions. For sponsors running UK clinical programmes, the difference between a well-feasibilitied study and a poorly-feasibilitied one can be measured in months of timeline, hundreds of thousands in budget, and the credibility of the entire development programme.