Universal · 10 min read
How to Write a Winning Grant Proposal: A Universal Playbook
Funders are different. Reviewers across funders are not. After 50 years of grant-writing literature and tens of thousands of post-decision reviewer transcripts, the patterns in funded proposals are well-mapped. Most rejections aren’t about the science — they’re about the writing failing to give reviewers the right argument in the right order. This is the universal playbook.
1. Reviewers read fast, in a fixed order
Most reviewers read 8–30 proposals in a review round. They read in this order: title, abstract, aims, methodology, budget, CV. They form an opinion in the first two and look for confirmation or disconfirmation in the rest. Implications:
- The title and abstract carry 60% of the conversion. Spend disproportionate time on them.
- If the aims aren’t clear in the first sentence, you’ve lost the reviewer.
- The methodology is read for confirmation that you can deliver, not for novelty.
2. The single hypothesis principle
Every winning proposal can be summarised in one sentence: “We hypothesise X. To test it, we will do Y. If we’re right, Z changes.”
Three common failures of this principle:
- The proposal has three hypotheses. Reviewers can’t hold them simultaneously; they fall back on impressions.
- The hypothesis is stated as “we will study” rather than “we hypothesise that”.
- The hypothesis is so cautious it’s untestable (or so bold it’s untestable for different reasons).
3. State of the art: argument, not literature review
A weak proposal’s state-of-the-art reads as a survey of recent papers. A strong proposal’s state-of-the-art is an argument: “Field knows X. Field does not know Y. Y matters because Z. We are positioned to address Y because we have already shown W.”
Three structural moves:
- Open with what the field has agreed on, not with the gap.
- Name the specific gap. Not “more work is needed” — concrete unknowns.
- Position your group’s preliminary work as the bridge.
4. Methodology: by hypothesis, not by technique
Many proposals organise methodology around techniques (RNA-seq, mass spectrometry, fMRI). Reviewers want methodology organised around hypotheses or aims. The structural difference:
- Weak: “Section 4.1: scRNA-seq. We will use 10x Genomics platform with…”
- Strong: “Section 4.1: To test hypothesis 1, we will identify cell-type-specific transcriptional signatures using scRNA-seq…”
5. Risks and alternatives: real, not aim-restated
A weak risk section says “Risk: aim 1 fails. Mitigation: revise approach.” A strong risk section names a concrete technical risk that a methodologically-sophisticated reviewer would actually worry about — and shows you’ve thought about a real Plan B.
Examples of strong risk statements:
- “Mouse model X has 20% inter-individual variability in phenotype Y. We will increase n to 12 per group and validate findings in human iPSC-derived cells.”
- “Method M has not been published in our model species. We have generated preliminary data showing feasibility (Figure 3); if scale-up fails, we will switch to method M′ (cited references).”
6. Preliminary data: prove credibility
Reviewers funded research, not promises. Preliminary data answers the implicit question “can this team actually do this?”. The minimum bar:
- One figure showing your group has the methodology working.
- One figure showing the central hypothesis is at least plausible.
- (Ideally) one figure showing scale-up: your method scales from pilot to study size.
7. Budget: align with workplan
The single most-noted reason for partial funding (in funders that allow partial cuts) is misalignment between budget and workplan. Reviewers read the budget asking: does each line match a deliverable?If a postdoc is listed but no work package depends on her, the budget gets cut.
8. The CV: depth over breadth
A weak CV lists 80 publications. A strong CV lists 5–10 most-relevant outputs and explains, for each, what the contribution is. Reviewers care: can this person do this work? They don’t care: how many papers has this person co-authored.
9. The reviewer’s mental checklist
By the end of reading, reviewers ask themselves five questions. Make sure every one has a clear answer they can find in your proposal:
- What is the question? (Aims, abstract)
- Why does it matter? (State of the art, impact)
- How will they answer it? (Methodology)
- Can they actually do it? (Preliminary data, CV, budget)
- What if they’re wrong? (Risks, alternatives)
10. The 12-week schedule
- Week -12 to -10: lock the project idea. One-paragraph pitch tested with 3 senior colleagues.
- Week -9 to -7: full draft. Methodology, preliminary data, budget.
- Week -6 to -4: outside review. Two senior researchers, ideally one outside your direct field.
- Week -3 to -2: integrate feedback, finalise risks-and-alternatives section.
- Week -1: language polish, references checked, page budget enforced, budget reconciled.
- Week 0: submit at least 24h before deadline.
Track every relevant call automatically
Grant Radar tracks 24 funders worldwide and ranks every open call against your ORCID profile. Spend your time writing winning proposals, not chasing portals. Free during the public beta.
Start free trial