Measuring What Matters Booklet
Measuring What Matters Booklet
au
Measuring
What Matters
An introduction to project
evaluation for not-for-profits
www.ourcommunity.com.au
Community Funding Centre
Measuring What Matters
An introduction to project evaluation for not-for-profits
Measuring What Matters: An introduction to project evaluation for not-for-profits
This publication is copyright. Apart from any fair use as permitted under the Copyright Act 1968,
no part may be produced by any process without permission from the publisher
Or email [email protected]
Please note:
While all care has been taken in the preparation of this material, no responsibility is accepted by the author(s) or
Our Community, or its staff, or its partners, for any errors, omissions or inaccuracies. The material provided in this
guide has been prepared to provide general information only. It is not intended to be relied upon or be a substitute
for legal or other professional advice. No responsibility can be accepted by the author(s) or Our Community or our
partners for any known or unknown consequences that may result from reliance on any information provided in
this publication.
ISBN 978-1-876976-61-3
Terminology 7
Evaluation principles 8
Evaluation frameworks 12
Process evaluation 18
Output evaluation 19
Outcome evaluation 21
Impact evaluation 22
Further reading 39
6 www.ourcommunity.com.au
Terminology
In the world of evaluation, terminology can be confusing. Different
evaluation models have different names for each stage of the process.
In this book, we’ve adopted the most common definitions, as follows:
Inputs: The resources used to complete a project, such as money
and volunteer hours. How much money did we raise in the campaign
to repair the pavilion roof? How many trees were donated for the
tree-planting day? How many people volunteered to take part?
Mapping or counting the resources used across the operation – your
inputs – can be laborious, but it’s generally not conceptually taxing.
Activities: the things that you do with the inputs – run workshops,
rescue cats, counsel troubled teenagers, plant trees. Some project
evaluation models slot activities under “inputs”, and some regard
them as “outputs”, but we’ve put them in their own category.
Outputs: the immediately discernible products, goods, services or
achievements brought about by the inputs and activities – 20 cats rescued,
2000 trees planted, three workshops conducted, 65 teenagers counselled.
Outcomes: the medium-term effects that those products, goods,
services or achievements have when fed out to individuals and into
the social environment; for example, a 30% reduction in the incarceration
rate of teenagers in Troubledtown. Remember that outcomes might be
intended or unintended. An outcome of the rescue of 20 cats might be
a 500% increase in the rodent population.
Impacts: the long-term developments that can be attributed to what
you have done. How much has changed because of the work you put
in? In what way is the world a better (or different) place? For example,
the impact of the counselling of 65 teenagers might be increased social
cohesion in Troubledtown 12 months later.
You’ll have to pick your time horizon; the week after, the year after, five
years on, the next generation? The further you go, the harder it becomes
to see your faded signature on the benefits that arise.
Avoiding confusion
Some evaluation materials conflate output and outcome, and some blend
outcome and impact. Some, indeed, use the term “impact” for medium-
term results and “outcome” to describe the final population-wide effects
(the reverse of our terminology).
To avoid confusion, you should define how you use these terms in any
discussion or publication of your evaluation.
Mission-centred evaluation
You know that “aims” clause in your constitution, setting out all the
things you want to do? Have you done all those things? How do
you know?
Your entire purpose as an organisation is to achieve your mission.
Everything you do – everything – should be directed to that end.
You must, at some point, be able to say whether you’re doing what you
said you’d do. And you will need to present some evidence in support
of your claims. Evaluation is the process of collecting this evidence.
In part, you will need evidence of your achievements to satisfy external
stakeholders – grantmakers, government, donors, volunteers, business
partners, and the general public. All of these groups demand evidence-
based decision-making. Their expectations are high, and are often
embodied in funding contracts.
More importantly, though, you will want to evaluate your success to
assure yourself that you are in fact doing as much good as you think
you are. If you really want something to be true, it is easy to deceive
yourself into believing that it is true, and only unbiased evaluation
can ensure that you are not wasting your time and the community’s
resources on vanity projects.
If you don’t build evaluation in from the beginning, you’ll have a hell of a
time trying to bolt it on later.
Evaluation must be featured in your project design process from the
very earliest stages. Budgeting and timetabling implications must be
assessed and included in the design.
You should try to establish the parameters of your evaluation precisely
from the beginning. Having said that, it is not always possible to
establish in advance what the central features of a project will turn out
to be, and no evaluation can cover every possible set of events. Include
some flexible and open-ended opportunities to allow unforeseen
developments to be caught up by your examination.
Once you’ve decided what you want to measure, the first question
is what readings you need to take to establish a baseline. The figures
8 www.ourcommunity.com.au
that you come up with afterwards are meaningless unless they can be
compared with the pre-existing situation.
Evaluation principles 9
“Unnecessary” in this context means “we don’t care about this enough
to modify our policy”. Unless the data you are collecting could in some
circumstances lead you to change your practices, your procedures, or
your policies, then there is no point in collecting it.
Asking this question (“Do we care enough about this that we might
modify our policy?”) can be useful in cutting back on your information
demands. If you find that a survey form, for example, goes directly into
the filing cabinet to gather cobwebs, then something is badly wrong,
and unless you are prepared to put some time into analysing it then
you should strike it from your repertoire and save everybody down
the line valuable time.
10 www.ourcommunity.com.au
A commitment to following where evaluation leads will in many cases
involve arranging to do better next time. This involves documenting the
learnings from the evaluation, storing them until the opportunity recurs,
and having filing and indexing systems that enable you to recover and
apply them.
Every organisation has rules governing how practices, procedures,
and policies are altered – some flexible, some not. These governance
procedures must be held up against your evaluation outline to ensure
that the results of any evaluation will be fed into the decision-making
matrix and that no there are no serious obstacles to its impartial
consideration.
“The results were unfortunately not what we were hoping for,” lead
investigator Dr Sally Brinkman told the ABC in 2016. “The aim of the
program was to prevent teenage pregnancy. We can definitely say that
it didn’t do that.”
Researchers also found 53.8 per cent of the pregnancies among girls
who participated in the robot program were terminated, compared
to 60.1 per cent of those who did not.
Dr Brinkman said the study did not explore why the program had failed,
but anecdotally it seemed the girls enjoyed the process of caring for a baby.
She recommended schools stop using such programs.
Outputs, outcomes and impacts: evaluate them all
If we evaluate the success of the program in terms of outputs – e.g. the
number of schoolgirls who undergo the VIP program – then we might regard
“1000 girls educated” as a more successful program than “500 girls educated”.
However, if we evaluate the success of the program in terms of outcomes –
its effect on the rate of teen pregnancy – then we get a very different result
(more teen pregnancies).
Now let’s look back to our earlier definition of impacts: “the longer-term
effect of your activity. How much has changed because of the work you
put in? In what way is the world a better (or different) place?”
The impact of our activity might consist of the lower education levels and
higher rates of socioeconomic disadvantage associated with teen parenting.
Measuring and evaluating impacts was beyond the scope of the WA study.
However, known educational and socioeconomic indicators could be used
as proxy measures.
Evaluation principles 11
Evaluation frameworks
When you think of evaluating a project, perhaps the first thing you think
of is gathering data. That’s great, but data on its own is about as useful as
ice-cubes in a warm bath. Data needs a reason for being, and it all comes
back to your mission and vision: what is the change that your organisation
wants to see? What are you going to do to bring about that change? And
what makes you think that your intervention will work?
You can survey farmers and count chickens until the cows come home,
but you need to be pretty sure that counting the chickens will be a useful
indicator of the eggs that will be hatched if you want to increase the
number of omelettes in the world.
This is where evaluation frameworks come in.
Logic models
One of the most widely used frameworks for evaluating social change
projects is the logic model. A logic model explains how particular activities
lead to particular outcomes – if we do this, it will (we believe) result in
that.
A logic model is a useful way for you (and your funders and stakeholders)
to test your assumptions about a project (or about anything, really).
It provides a way for you to plot a causal chain from what you are
proposing to do, all the way to your eventual goals.
If you don’t know exactly where you’re going, any road will get you there.
In some ways, your evaluation planning doubles as a reality test for your
project. If you can’t design a way to evaluate it, it could be that your goals
are simply too vague and unfocused – for example, “This project will
improve health.”
If you are having difficulties designing your evaluation, go back and ask
yourself whether you need to tighten the project itself. Let’s say your
organisation’s vision is a state free from tobacco-related diseases. Perhaps,
when you say, “This project will improve health”, what you really mean is,
“This anti-smoking campaign aimed at migrant schoolchildren will reduce
the rate of smoking-related disease among this particular population
when they grow up.”
Starting at the end point (you want a state free from tobacco-related
12 www.ourcommunity.com.au
disease), you work back to find the interventions that you hope will
lead to that outcome. That’s your logic model. Here’s how it might look:
You set up an anti-smoking campaign.
And as a result:
Campaign materials reach migrant schoolchildren.
And as a result:
Knowledge of smoking-related conditions among migrant
schoolchildren increases.
And as a result:
Rates of smoking among migrant schoolchildren decrease.
And as a result:
Rates of smoking-related disease among these children when
they grow up decrease.
And as a result:
Rates of smoking-related disease in the state decrease.
Before you make a decision on whether to run such a campaign, you
need to be satisfied that all the links in the chain are sound.
The link between less smoking and less disease has been exhaustively
documented; the link between greater knowledge and lower smoking
rates in this population may require some literature citations or
references to earlier campaign data; the link between getting the
material and reading the material might need testing in advance; and
the link between the campaign and the material reaching the target
group would need to be spelled out in the plan.
Each link in the chain should be capable of being evaluated, and this is
where data comes in. At the end of the anti-smoking project you want
to be able to prove that
• The material did reach the target group (through survey data)
• The material did increase health knowledge (through in-depth focus
group data)
• Smoking rates did diminish (through survey and sales data).
Evaluation frameworks 13
and lower disease incidence cannot in this instance be demonstrated until
several decades later.
Few projects can provide meaningful evaluation over such timeframes.
In this instance, the weight of the scientific evidence that smoking causes
disease makes smoking rates a very reliable proxy measure, but such
robust correlations are not always available.
14 www.ourcommunity.com.au
What do you need to know?
A thorough evaluation will relate closely to your logic model. It will
enable you to check at each stage that your expectations have been
met. You can also work backwards from your goals, seeking measures
that would indicate that they had been met.
How will you get the answers?
Because the range of not-for-profit goals is effectively infinite, it is
impossible to list every method of evaluation you might find useful
in your particular situation. You may need to employ social surveys,
statistical overviews, or satellite photography. Your choice will be
guided by cost, by relevance, by estimates of effectiveness, and by
the constraints placed on your enterprise by considerations of ethics,
privacy, and public relations. For more on the most common evaluation
methods, see page 29.
Evaluation frameworks 15
What element of the project are we
evaluating, exactly?
Many things feed into the success of a project, not least the capacity
and performance of the organisation undertaking the work. It can be
conceptually useful to consider all these different elements of a project
separately.
16 www.ourcommunity.com.au
No examination of a project’s success is complete without looking at the
performance of the organisation that is running the project. The results
of such an examination might not be offered externally, but the appraisal
is an important exercise for a not-for-profit organisation to undertake
nonetheless.
Aspects of organisational performance such as management and process
evaluation are just as important as those that focus exclusively on the
project at hand – maybe even more so.
organisation from the board (“What are the risks associated with
borrowing $200,000 to fix the drainage problems on the footy oval?”)
to the roster of volunteers who clean the toilets after each match (“I’d
best not leave the bleach where the under-five players might find it –
common sense tells me it’d be dangerous, and besides, it’s covered by our
hazardous materials policy).
Taking a systems approach to evaluation – evaluating candidates, on-the-
job performance, risk, the budget, the finances, evaluating whatever you
care to evaluate, and doing it carefully, logically and thoughtfully – will help
you to:
• progress your mission
• look after your volunteers and paid staff
• know what’s working and what’s not, and why
• be aware of unintended outcomes
• adapt as needed.
If after all that you’re thinking, “But I just want to get on with the work”,
then be assured that you’re probably already practising evaluation. Every
time you reflect on why your client was frustrated at today’s meeting, or
think about how you might make sure that the annual fundraiser doesn’t
go under in the rain like it did last year, or survey your volunteers to see
what they like about working for your organisation, you’re evaluating.
Taking a systems approach just means understanding things as part of a
bigger picture.
18 www.ourcommunity.com.au
You may also find that some or all of these changes will work against
you, cutting into your budget bottom line rather than improving it. You
may also find that your timetable slips or that some elements on which
you have relied become unavailable. You must have in place systems
that allow you to work with your funders and stakeholders to resolve
these issues.
Feedback loop
Output evaluation
Did we keep our promises?
At the end of the project you will need to be able to say whether
you have done what you set out to do. In the contract with your
funder you promised to produce 100 widgets. Your plans set out how
you would reach your target; you monitored the production process
over the required time; and now, at the end of the process, you have
The most desirable sort of project is the one that, once started,
can continue successfully on its own when funding and support are
withdrawn. Ideally, the community becomes involved, individuals are
enthusiastic, other resource providers emerge, and institutions are
convinced of the importance of it all.
This is, however, a big ask. Many projects reach the stage where the
enterprise can continue if the funding continues, but far fewer cross the
bridge to full independence.
20 www.ourcommunity.com.au
Scalability: can we roll it out?
System improvement
Outcome evaluation
How has the situation changed?
Outcome evaluation describes the situation at the end of the project,
compares it to the situation at the beginning of the project (the
baseline) and maps the changes (or lack of changes).
Changes might be observed
• in the participants
• in the available resources
• in organisations and structures.
Changing attitudes
Building capacity
You may have increased the level of knowledge of the relevant facts in
the participants, or the region, or society, and given people the means to
change their lives in response. Their participation may have taught them
some lessons about collective social activity to bring about change, and
they may have developed technical skills, social skills, or bureaucratic skills
that will assist them on similar projects in the future – and will allow them
to mount projects of their own.
Because this is a process skill, it is more difficult to measure than changes
in attitude. The real test is whether people behave differently in the future.
Changing structures
It is also possible that your project will leave behind mechanisms that will
facilitate the next step – continuing associations with a common goal,
procedures for consultation, links between the community and local, state
or federal governments, precedents that will guide people working in the
field.
Any changes of this nature should be recorded centrally and taken into
account in the final balance.
Impact evaluation
How is the world different?
Chinese statesman Zhou Enlai was once asked what he thought the
effects of the 1789 French Revolution had been. “Too early to say,”
he replied.
Impact evaluation is the highest level of the evaluation hierarchy, and
measures what a project has left behind it that is sustainable in the long
term. It is by a long way the hardest thing to measure, and the most
confusing and ambiguous.
The simplest problem with long-term evaluation, though by no means the
least, is the length of its term. It is generally very difficult to find resources
sufficient to track a project’s effects over any extended time. Funders
are unwilling to commit funds years in advance for later evaluation
procedures. Boards and stakeholders change their priorities or move on
22 www.ourcommunity.com.au
and are replaced by others.
Most organisations have time horizons that extend to the end of their
current strategic plan, if that. Keeping a focus on the consequences
of what was done last year or last decade requires a wholehearted
commitment by everybody concerned, and is thus rare and valuable.
What’s even more difficult is proving that any particular project or
intervention created the change.
To find out what effects your work has had, you must begin by seeing
what, if anything, has altered permanently since you began in regard to
the issues you sought to address. Now that the funding has gone away,
what remains?
Is the population you have been working with happier? Or healthier?
Richer? Better informed? Better behaved? Better organised? Better
equipped? Better housed? More skilled? More generous? More tolerant?
What did you originally predict, or project, the situation would be at
this time? How does the vision compare to the reality?
24 www.ourcommunity.com.au
unplanned alterations. We decide in advance what it is we want to
know, what we regard as important, and what is going to be recorded.
Most evaluations don’t deal with what actually happens during a
project. They deal with a very much smaller subset: the things that
happen in the areas we thought were important before we had the
benefit of actually doing the project.
Some forms of evaluation are better at eliciting tangential
developments than others. Canadian health promotion researcher
Ron Labonte, for example, favours a methodology that is specifically
interpretive, where the research findings arise in the course of the
process of inquiry and rely on “iteration, analysis, critique, reiteration,
reanalysis, and synthesis” – or, to put it another way, talking it through
with the players. Labonte describes this as the story-dialogue method,
operating through a reflective dialogue between participants.
The project evaluation process might in this format begin with both
the practitioners and the community being asked to write up their
stories about what happened during the intervention, stories that
come from their own experience and their own understanding. The
next step is a workshop where all the parties collect to discuss these
case studies, breaking into small groups around each storyteller for
intense reflection. One person acts as recorder, and the group feels
its way to insights into the situation. The aim is to generate a deeper
understanding of the underlying themes and to identify the tensions
that exist, and the dialogue is structured to move the discussion from
a simple description of what happened to some sort of explanation, or
explanations, of how it happened.
At a more base level, some funders now are asking not-for-profit
organisations to include in their final project report details of any
unintended or unexpected outcomes. This is certainly better
than nothing.
Counterfactuals
What would have happened if we hadn’t been there?
If we wanted to establish scientifically how much, for example, The
Wilderness Society has contributed to the cause of the environment
in Australia, it would be necessary to run the last three decades of
Australian history over and over about 50,000 times with variant
hypothetical inputs and see what difference it would have made if the
Society had never existed, or had behaved in other ways. Would other
people have done the same work? Would the Franklin Dam have been
built, or was the culture moving against ecological disruption with or
26 www.ourcommunity.com.au
Gathering the data
The data collection method you choose will generally be dictated by
what you wish to measure, and how you wish to measure it.
Try not to over-complicate your data collection – but don’t over-
simplify it either. Work out what you need to know, then find the
methods that will deliver that knowledge.
Ideally, you want your measures to be SMART. SMART is an acronym
for five criteria:
Specific
The more specific you can make your aims, the easier it will be to
measure them. This approach favours goals that can be expressed in
numbers – dollars, percentages, client numbers. The objective is clearly
defined.
Measurable
Essentially, this calls for information that can be quantified. If your
objective genuinely cannot be reduced to quantities, this raises many
questions about whether it is a valid aim. Almost all activities are
measurable at some level.
Achievable
Your expectation of what can be accomplished must be realistic
given the time, the resources and the obstacles you are dealing with.
Relevant
In their quest for measurability and quantification, people will
sometimes settle on project measures that are hard-edged and definite
but not closely related to what they actually want to achieve. This
should be avoided.
Time-bound
Any objective that doesn’t have a due date is effectively worthless.
Quantitative data
Evaluators tend to like to measure items that can be counted
rather than estimated, and observed rather than deduced.
Quantitative research gathers data in numerical form which can
be put into categories, or in rank order, or measured in units of
measurement. This type of data can be used to construct graphs
and tables of raw data.
Qualitative data
Working with qualitative data involves the exercise of judgement. It
requires understanding and describe differences rather than converting
them into numbers. Subjectivity may be unavoidable. Empowerment,
for example, is not an object that exists separately from the internal
states of the people who experience it, so it can be difficult to measure.
Milestones
Baseline measurement
The first stage in any evaluation is to measure the baseline so that it is
possible later to know what has changed. Without a baseline, it is very
difficult to establish when any development occurred, and thus even more
difficult to establish what caused things to change.
Input measurement
In the course of your project, you must record all relevant inputs – the
number of hours spent on the project, the employees or volunteers
responsible, the costs incurred. The process for recording these items
28 www.ourcommunity.com.au
should be widely known and the person or persons responsible for
carrying it out clearly identified. You should be able to draw on this
database to analyse any relevant correlations.
Output measurement
As your project runs its course, you should also record all the
measureable outputs you have delivered: the number of seminars
you’ve run, the number of bales of hay you’ve given to rescued
donkeys, the number of games your team has played. Again, the process
for recording these items should be widely known and the people to
carry it out clearly identified.
Endpoint measurement
At the close of your project you will be called upon to take stock of
your completed efforts, to decide whether they are satisfactory, and
to extract from your conclusions those factors that will influence your
future planning.
30 www.ourcommunity.com.au
Listening to the data
Whatever your chosen data collection method, problems will almost
inevitably stand in the way of entirely clear answers to your questions.
You may have to make allowances for inadequate data (you may not,
for example, be able to survey everybody involved), missing data
(not everybody you invite to participate will respond to your survey),
and misleading data (not everybody who responds will answer your
questions honestly).
Nonetheless, some data is nearly always better than no data at all.
32 www.ourcommunity.com.au
Wasting resources
Sledgehammer, meet nut. Evaluation can be
• disproportionate, where small, cheap projects are burdened with
extensive reporting requirements that consume resources essential
for the program. While no project should be allowed to end without
any evaluation at all, that evaluation should be appropriate to the
project scale, and might at one end of the spectrum consist of no
more than half a page of written impressions.
• unused, where measurements called for simply by habit are taken
and filed away without being analysed. Any evaluation that will not
be incorporated into the decision-making process should not have
been carried out in the first place.
Plan the
next
round
Improve
the
offering
Do the job
Assess the
outcomes Measure the
variables
34 www.ourcommunity.com.au
Every successive iteration should do better.
The term “better” is, of course, subject to interpretation. It can mean
cheaper, or larger, or prettier – improved financially, or quantitatively, or
qualitatively.
And, because anything that changes the world has to interact with
everything else in the world, there are issues of time and chance.
You’ll have to find resources to fund the evaluation. Then you’ll have to
decide whether the evaluation was worth what you paid for it – you’ll
have to evaluate the evaluation. And then you can start all over again.
36 www.ourcommunity.com.au
Sample evaluation reports
What form should your evaluation report take? Should it be a single
page on your website, a 120-page commercially printed full-colour report
distributed to all your funders, staff, volunteers, members and project
participants, or something in between?
There’s no right or wrong answer to this question, but you probably
won’t be able to cover all the ground you need to cover in a single
page on your website, and there’s nothing intrinsically better about
a lengthy report.
Our advice is to consider your intended readership, consider the most
compelling, succinct way to convey the information you need to convey,
and write your report in plain English – no jargon, no weasel words, and
no more acronyms or abbreviations than absolutely necessary.
If you’ve planned and executed your evaluation thoughtfully, and if you
stick to plain English when it comes to writing the report, you’ll be on
a winner.
Below we’ve listed some examples of published evaluation reports.
Not all of them meet all those criteria (there’s not much value,
for example, in publishing a pie chart for each one of your data
sets without analysing the data and explaining why it’s significant),
but examining flawed evaluations can be as instructive as looking
at truly insightful ones.
38 www.ourcommunity.com.au
Further reading
There is a sea of information about project evaluation available online.
Here are our picks.
Website: Better Evaluation
http://betterevaluation.org/
An excellent website by an international coalition of evaluation
professionals and academics, including Australians. The section headed
“Approaches” highlights the fact that for every evaluation challenge,
there is an approach to suit. For example, if you’re finding it difficult
to prove causality when evaluating your program, try a contribution
analysis.
PDF booklet: Measuring Outcomes
http://strengtheningnonprofits.org/resources/guidebooks/
MeasuringOutcomes.pdf
A 55-page PDF booklet – a little dry, but well structured and
comprehensive. Written for the US market but easily transferable to
the Australian context.
Web page: Basic guide to program evaluation (including
outcomes evaluation)
http://managementhelp.org/evaluation/program-evaluation-guide.htm
A low-tech but highly readable guide from the US-based Free
Management Library.
The Complete Community Fundraising Handbook: How to Make the Most Money Ever
for Your Community Organisation
The Complete Schools Fundraising Handbook: How to Make the Most Money Ever for
Your School, Pre-school or Kindergarten
How to Manage Your Grant after Winning It! (includes templates on bonus CD)
Great Fetes: Fundraising and Fun – Without the Fuss (includes templates on bonus CD)
More than Money: How to Get Money, Resources and Skills from Australian Businesses
...............................................................
Making Meetings Work: Conquering the Challenges and Getting Great Results
Surviving and Thriving as a Safe, Effective Board Member: Facts You Need to Know Before,
During and After Joining a Community Board
Get on a Board (Even Better – Become the Chair): Advancing Diversity & Women in
Australia
...............................................................
Effective Letters: 50 of the Best Model Letters to Help Community Organisations Fundraise,
Connect, Lobby, Organise & Influence