This article covers:
- Your submission data is diagnostic, not just descriptive
- Calculate whether your tools improve outcomes, not just feel easier
- Compare year over year, not programme to programme
- Look for patterns in judge behaviour across multiple programmes
- Test fee structures systematically, not reactively
- Refine judging criteria based on what differentiates quality
- Track demographic reach to measure access improvements
- Listen for signals in feedback
- Run deliberate experiments, not reactive changes
- Build institutional memory through systematic data review
You’ve run your programme. Entries came in, judging happened, winners got announced. Now what?
Most organisers file away their spreadsheets and start planning next year with the same assumptions they started with last year. They know roughly how many entries they received and whether judges complained, but they can’t explain why submissions dropped compared to two years ago or whether their fee increase suppressed applications.
Data exists. It’s sitting in your platform, your email inbox, your payment processor. Are you’re using it to improve your programme strategically rather than just repeating what worked (or didn’t) before?
Your submission data is diagnostic, not just descriptive
Numbers tell you what happened.
The right questions tell you why it happened and what to change.
Understanding your candidate journey shows you where people drop off during submission. That’s essential baseline information. But strategic programme improvement requires asking different questions:
- Did changing your deadline affect completion rates?
- Do certain categories consistently attract stronger work?
- Does judge satisfaction correlate with score distribution patterns?
Start with the decisions you need to make for next year’s programme. Fee structure, marketing timeline, category definitions, judging criteria, eligibility requirements. For each decision, identify what data would make the choice clearer rather than relying on instinct or copying what worked three years ago.
If you’re debating whether to add an early bird discount, historical data showing how many entries came in during your final week matters. If you’re wondering whether a category needs splitting, looking at entry volume alongside judge comments about “difficult to compare entries” matters. Data without a decision attached to it is a gut reaction.
Calculate whether your tools improve outcomes, not just feel easier
Everyone claims platforms save time. The real question is whether they improve your programme enough to justify the cost, or whether they just shift work around.
Track hours spent on administrative tasks across your programme lifecycle. Entry processing, judge coordination, payment reconciliation, applicant queries. For three programmes running email and spreadsheets, you might log 45 hours, 68 hours, 52 hours across similar entry volumes. For three programmes using automation, log the same metrics.
Beyond time savings
Total hours is the basic metric, but it’s also worth considering error rates (payment mismatches, lost files, miscommunicated deadlines), judge satisfaction (measured by returning judges year over year), and submission quality improvements from better applicant experiences.
Some programmes discover automation saves 40 hours but returns fewer judges because the process feels impersonal. Others find automation saves 20 hours while increasing entries by 15% because the submission experience improved dramatically. Both are real outcomes worth measuring.
If you can’t demonstrate programme improvements beyond “it feels smoother,” you might be paying for convenience rather than results. Sometimes convenience is worth it but knowing the difference matters for budget conversations.
Compare year over year, not programme to programme
Industry benchmarks are mostly useless. Comparing your photography award to someone else’s business innovation prize tells you nothing actionable because audience, budget, marketing reach, and reputation all differ.
Compare your programme against itself across cycles. Did entries increase or decrease? Did judge participation improve? Did completion rates shift after you simplified your form? Did extending your deadline generate more entries or just push procrastination later?
Build a simple tracking sheet with core metrics across programme cycles: total entries by category, average judge scores by category, judge participation rates, entry fee revenue versus processing costs, completion rates, entries per marketing channel where trackable.
Single programme data points can mislead you. Trends across multiple cycles reveal whether your changes work or backfire.
Look for patterns in judge behaviour across multiple programmes
Judge completion rates matter (and you should track them), but strategic improvement comes from understanding patterns across judges and programmes.
Some judges consistently score tight clusters (7-8-9 on every entry). Others use the full range (3-10). Neither is wrong necessarily, but if your programme consistently produces narrow scoring, your criteria might not help judges discriminate between excellent and adequate work.
Track score distributions by judge across programmes. If variance increases year over year, your criteria might be getting clearer. If it tightens, you might be losing the ability to identify truly exceptional work.
Comment frequency tells you whether judges feel confident in their assessments. Judges who leave detailed comments usually understand what they’re evaluating. Judges who score quickly without commentary might be confused by criteria or overwhelmed by volume. If comment frequency drops programme to programme, that’s worth investigating before you lose engaged judges.
Judges who return year after year signal satisfaction with the process. Track returning judges as a percentage of your panel. Declining return rates often precede judge recruitment becoming harder, because dissatisfied judges tell peers about their experience privately before telling you directly.
This can also be done approaching new judges. Make sure to include enough details of the platform/process you will be using in your approach and see how many of them accept or decline to take part.
Download launch checklist
Never miss a critical step: 300+ tasks across 8 phases, from securing judges to winner announcement.
Test fee structures systematically, not reactively
Entry fees generate revenue but suppress applications. Most programmes set fees by copying competitors or guessing, then wonder why entries declined.
Test fee changes systematically across programme cycles. A programme charging £40 for three years, then testing £30, then £50, collects evidence about price sensitivity in their specific audience. One cycle doesn’t prove anything (economic conditions, marketing timing, and category changes all influence entries), but three cycles with controlled fee tests reveal patterns.
Revenue versus reach trade-off
Track entries and revenue together. Dropping from 200 entries at £40 (£8,000) to 170 entries at £50 (£8,500) generated more revenue but cost you 30 submissions. Whether that’s acceptable depends on your programme goals. Are you trying to maximise community reach or revenue sustainability?
Early bird pricing reveals time sensitivity in your audience. If 60% of entries come in during early bird periods, your audience is price-conscious and deadline-driven. If only 15% claim early bird rates despite heavy promotion, price matters less than deadline proximity for your candidates.
Structuring fees that balance accessibility and sustainability requires understanding your specific audience behaviour, not industry averages. Test different approaches, measure outcomes, adjust based on evidence.
Modern platforms should support flexible pricing tests without requiring separate programme instances. If experimenting with fee structures means duplicating your entire setup, you’re working too hard to learn from your programme data.
Refine judging criteria based on what differentiates quality
Most programmes write judging criteria once during setup, then use them unchanged for years.
Review judge scores and comments annually. If scores cluster narrowly (everything between 7-9), your criteria aren’t helping judges distinguish excellent from adequate work. If scores scatter randomly with no pattern, judges interpret criteria differently or the criteria don’t match what matters in your field.
Look for mismatches between scores and winner selections. If your panel consistently overrides numerical rankings to select different winners, your scoring criteria don’t capture what experienced judges value. This isn’t necessarily wrong (judgment matters), but it suggests criteria refinement would make scoring more useful.
Track which criteria generate the most judge questions or comments. If judges frequently ask, “how do we evaluate X criterion?” across multiple programmes, that criterion is probably too vague or poorly defined. Clarify it or replace it with something judges can apply consistently.
Categories that consistently show different scoring patterns might need criteria adjustments. If your “emerging artist” category gets systematically lower scores than your “established artist” category despite similar entry quality, you might be using criteria that favour experience over potential. Adjust criteria to match what you want to reward in each category.
Track demographic reach to measure access improvements, not just report numbers
Collecting demographic data is easy. Using it to expand who your programme reaches takes sustained effort.
Compare applicant demographics against your target population year over year. If your field is 45% female but your applicants are 25% female for three consecutive years, your outreach isn’t reaching women effectively. If that percentage climbs from 25% to 32% to 38% across three programmes, your changes are working and you should accelerate them.
Geographic distribution reveals whether your programme is genuinely open or practically restricted. A “national” award that consistently receives 70% of entries from two cities isn’t serving a national community. Track this annually and adjust marketing to underserved regions deliberately.
Career stage distribution matters for programmes claiming to support emerging talent. If your “open to all” award consistently attracts 80% established professionals, emerging artists either don’t know about your programme or don’t believe they can compete. Look at whether winners reflect your applicant demographics or consistently come from the overrepresented group.
Don’t just collect demographic data for funder reports. Use it to identify which communities your programme isn’t reaching, then test specific outreach changes and measure whether demographics shift. Data without action is performance, not improvement.
Listen for signals in feedback
Post-programme surveys collect opinions. Strategic data use means identifying patterns that reveal problems or opportunities.
If three separate applicants mention confusion about category fit, you have a category definition problem. If multiple judges’ comment that they wanted more context about applicants’ circumstances, your submission form might be too simple. If returning entrants consistently mention a specific pain point, that’s worth prioritising over random complaints.
Recurring themes matter most
Track recurring feedback themes across programmes. A complaint mentioned once might be individual preference. The same issue raised by different people across multiple cycles signals something worth addressing (if these issues arise whilst people are submitting, be sure to address them immediately. This will ensure the rush submissions at the end of your programme are as seamless as possible).
Judge feedback often contradicts applicant feedback. Judges might want stricter eligibility while applicants want broader access. Competitors might want lower fees while your budget requires current pricing. Not all feedback deserves implementation but understanding where stakeholder interests conflict helps you make intentional trade-offs rather than trying to please everyone. Addressing these contradictions and reasons for your choices in your copy will go a long way in building trust with candidates.
Don’t survey after every interaction. Pick specific decision points where feedback influences next year’s programme. Post-submission surveys about the application experience. Post-judging surveys about criteria clarity. Winner surveys about prize value and programme impact. Each should be short (five questions maximum) and directly actionable.
Want more articles like this?
Give this one a like
Build your own baseline, not someone else’s benchmark
Comparing your programme to others is mostly pointless. Entry numbers, fee structures, marketing budgets, and target audiences vary too widely for meaningful cross-programme comparison.
Your programme three years ago is your best comparison point. Are entries trending up or down? Is judge satisfaction improving or declining? Are completion rates rising or falling? These trends reveal whether your changes work or backfire.
Track core metrics consistently: total entries, entries by category, average judge scores by category, judge participation rates, entry fee revenue, processing time investment. Plot these annually. Patterns emerge over three to five programme cycles, not from single-year data.
When you change something significant (adjust fees, revise guidelines, add new categories), track impact systematically. Compare metrics before and after the change across multiple cycles. One good year after a change doesn’t prove causation. Sustained improvement across subsequent programmes suggests your change worked.
Pro tip
When you make any changes, be sure to note down what you think the results of that change will be with tangible measurable data. This will guarantee you know if your changes reached your expectations. Without those you are likely to struggle with understanding if the change was the success you felt it was going to be later on.
Your own experiments generate more reliable data than copying what worked for programmes with different audiences, budgets, and goals.
Question your assumptions when data contradicts expectations
The most valuable data challenges what you thought you knew about your programme.
If your beautifully crafted guidelines that took weeks to write correlate with low submission starts, maybe they’re too long or detailed regardless of their clarity. If applicants consistently misinterpret a requirement despite explicit instructions, your clarity assumption is wrong. If judges consistently score one category lower than others, category definitions might need adjustment.
Good data reveals gaps between what organisers think works and what drives behaviour. If you’re never surprised by your metrics, you’re either not measuring the right things or not looking carefully enough at what the numbers reveal.
Question patterns that seem too predictable. Real programmes face unexpected challenges. Clean upward trends year after year might indicate measurement problems rather than continuous improvement. Perfectly consistent metrics across different programme cycles often mean you’re measuring after the fact rather than using data predictively.
Watch for conflicts between quantitative and qualitative data. If your survey says applicants loved the process but completion rates dropped, something’s wrong with either your survey design or your completion tracking (e.g. you survey was only sent to candidates who completed the process).
When metrics disagree, investigate why rather than picking the interpretation you prefer (personally I always favour data over surveys. I find it a better source of truth if measured correctly).
Recognise that ignoring data compounds problems across programmes
Programmes that don’t use data strategically repeat expensive mistakes instead of learning from them.
A programme that receives 30% fewer entries than three years ago without understanding why will keep losing ground. Maybe their fee increased while competitors stayed flat. Maybe their marketing shifted to channels their audience abandoned. Maybe their category structure became outdated. Without tracking causes, they’ll keep throwing marketing budget at a structural problem.
Or a programme that extends deadlines every year because judging always runs late without tracking which judges cause delays. The same three judges might create all the delays across multiple programmes, but without completion tracking, organisers treat late judging as inevitable rather than a coordination problem with specific solutions.
Data-informed programmes improve incrementally year over year. They identify bottlenecks, test solutions, measure results, and adjust. Programmes that rely on instinct often improve sporadically (when someone has a good idea) but regress unpredictably (when that person leaves or priorities shift).
Insights like these straight to your inbox
Receive weekly tips from our founder to grow your program’s impact; regardless of what tools you use.
Run deliberate experiments, not reactive changes
Once you’ve identified problems through data, test solutions systematically rather than implementing wholesale changes based on hunches.
If completion rates dropped last programme, you might blame your form length. But testing requires isolating variables. Shorten the form for next programme while keeping everything else constant. If completion rates improve, you’ve learned something. If they stay flat, form length wasn’t the issue.
Document what you change and why. Track results explicitly. Data-driven iteration beats inspired redesigns because you’re learning from real behaviour patterns rather than assumptions about what should work.
Not every change will succeed. Some will backfire (you’ll shorten categories and discover entries were already marginal, so fewer categories just reduced total submissions). That’s acceptable if you measure outcomes and reverse unsuccessful changes quickly.
Build a programme improvement log across cycles. “Programme 2023: Added early bird pricing, entries increased 12%. Programme 2024: Simplified form from 12 questions to 8, completion rate rose from 62% to 71%. Programme 2025: Split photography category into two specialisms, total photography entries dropped 18%.” This log becomes your programme knowledge base, especially valuable when staff transition.
Build institutional memory through systematic data review
The goal isn’t just collecting data or creating dashboards. The goal is organisational learning that survives staff transitions and memory gaps.
Set up quarterly review sessions where programme leads examine trends together. What changed compared to last programme? What didn’t work as expected? What surprised us? What should we test next? Document these discussions, because they contain context that raw numbers don’t capture.
Create simple tracking documents showing critical metrics across programmes. Total entries over time. Judges return rates. Category performance trends. Revenue versus processing costs. Keep it scannable – if your tracking system needs explaining, it’s too complex to use consistently.
Share relevant insights with stakeholders when appropriate. Judges want to understand how their panel’s scores compared to others (anonymously). Partners want to see reach and engagement trends. Funders want demographic impact data. Transparent information sharing builds trust and often generates useful suggestions from people who understand your programme context well.
The real lies in the conversations data enables about how to improve systematically rather than changing things randomly whenever someone gets frustrated with current approaches.
Data improves programmes through accumulated learning
Individual data points rarely change programmes. Three years of tracked improvements compound into meaningful programme evolution.
Your programme already generates useful information. Whether that information improves your programme year over year depends on whether you’re systematically learning from it or just filing it away.
Start with decisions you need to make for next programme. Fee structure, category definitions, judging criteria, marketing timing, eligibility requirements. For each decision, identify what historical data would make the choice clearer. Track those specific metrics across your next three programmes.
Understanding where candidates drop off in your submission journey gives you tactical information about immediate problems. Using multi-year data to refine your programme strategically gives you compounding advantages over programmes that repeat what worked three years ago regardless of changing conditions.
Good platforms make data collection automatic rather than burdensome. The right submission management tools track patterns naturally as your programme runs, so you can focus on interpreting trends rather than assembling spreadsheets.
Your programme improves through systematic learning, not inspiration. Use what you already know to make better decisions next year.
We can help!
Zealous makes running your programmes easier
But we’re not alone in the space – here are 8 others you may wish to consider (even if we would prefer you choose us!).
Want us to write more content like this? Give it a like
Share

Guy Armitage is the founder of Zealous and author of “Everyone is Creative“. He is on a mission to amplify the world’s creative potential.
Frequently Asked Questions
How do I know which data actually matters for improving my awards programme?
Start with the decisions you need to make for next year rather than tracking everything. If you’re debating fee structure, track entries and revenue across different pricing tests. If judge recruitment is getting harder, track which judges return year over year. Data matters when it informs specific decisions. Your programme improvement priorities determine which metrics are worth tracking.
Should I trust year-over-year comparisons when other factors change between programmes?
Trends across multiple programmes reveal more than single comparisons. If entries dropped 15% one year, that might reflect economic conditions or competitor programmes launching. If entries declined 12%, 8%, and 11% across three consecutive programmes despite different external conditions, you have a structural problem worth investigating. Track what you can control (fees, timelines, categories) and note what changed externally. Three programmes of data teach you more about what drives results than one programme in perfect conditions.
How can I measure whether automation actually improves my programme versus just feeling easier?
Track outcomes, not just efficiency. Administrative hours matter, but also measure judge satisfaction (return rates year over year), submission quality improvements (judge scores or winner calibre), entry volume changes, and error rates (payment mismatches, lost files, deadline confusion). Some programmes save 30 hours through automation but lose judges because the process feels impersonal. Others save 15 hours while increasing entries 20% because applicants enjoy the experience. Both are real outcomes. Calculate total programme improvement across multiple dimensions. Good awards management workflows improve results, not just reduce effort.
What should I do when judge scoring patterns suggest my criteria aren’t working?
Review score distributions across multiple programmes before changing anything. If scores cluster narrowly (7-9 on everything) for three years running, your criteria don’t help judges discriminate quality. If scores scatter randomly, your definitions are too vague. Look for mismatches between rankings and actual winner selections. If your panel consistently overrides numerical scores, your criteria don’t capture what experienced judges value. Revise one criterion at a time, test it, measure whether scoring improves. Wholesale changes make it impossible to identify what worked.
How do I collect demographic data that actually improves programme access rather than just reporting numbers?
Compare applicant demographics against your target population year over year, then adjust outreach based on gaps. If your field is 45% female but your applicants are 28% female for three years straight, your programme isn’t reaching women effectively. Test specific changes (partnering with women’s networks, featuring female winners prominently, adjusting marketing language) and track whether demographics shift. Geographic distribution matters too. A “national” programme that draws 75% of entries from two cities needs deliberate outreach to underserved regions. Track whether winners reflect applicant demographics or consistently come from overrepresented groups. Demographic data only improves access when you act on gaps systematically.
When should I experiment with major programme changes versus incremental adjustments?
Test incrementally when possible. Isolated changes teach you what actually works. If completion rates dropped, test one fix (shorter form or clearer instructions) while keeping everything else constant. Major simultaneous changes make it impossible to identify which change drove results. Sometimes programmes need substantial overhauls after years of decline, which is fine, but then you need to track everything carefully and accept you’ll learn less about causation. Document all changes in a programme log so future organisers understand what was tested and why.
How do I use programme data to make funding requests more convincing?
Show trends, not snapshots. Funders want evidence of sustained improvement. Three years of 15% entry growth or expanding geographic reach demonstrates momentum. Pair numbers with qualitative impact (winner testimonials, career advancement examples). Explain how previous funding enabled specific improvements and what you learned. Frame requests around evidence: “Demographic data shows we’re underserving Region X despite strong demand. Additional funding would support targeted outreach we’ve successfully tested elsewhere.”









