Anna Downie (a.downie@ids.ac.uk) Monitoring and Evaluation Coordinator

Post on 06-Jan-2016

33 Views

Category:

Documents

0 Downloads

Preview:

Click to see full reader

DESCRIPTION

Valuing evaluation beyond programme boundaries: Communicating evaluations to enhance development effectiveness globally. Anna Downie (a.downie@ids.ac.uk) Monitoring and Evaluation Coordinator Strategic Learning Initiative Institute of Development Studies, UK. Evaluation as a public good. - PowerPoint PPT Presentation

Transcript

Valuing evaluation beyond programme boundaries:

Communicating evaluations to enhance development effectiveness

globallyAnna Downie (a.downie@ids.ac.uk)Monitoring and Evaluation Coordinator

Strategic Learning InitiativeInstitute of Development Studies, UK

Evaluation as a public good

• Communicating evaluations for accountability and communicating to share learning

• Access to lessons learnt for practitioners and policy makers

• Benchmarking and learning between organisations

Evaluations, context and the politics of knowledge

• Evaluations are politically sensitive• Results are context specific and often complex– Especially from participatory monitoring and evaluation

processes– Lessons learnt either to general or too specific to be useful

• Thematic/sector/country wide evaluations try to make lessons learnt more relevant

• Need to find ways to communicate different types of evaluation, at different stages during the process

Challenges• Few incentives to communicate beyond the

programme• Experience of the IDS Knowledge Services:– Large and bulky reports, hard to summarise– Written for a specific audience (often the donor)– Rigour and quality often unclear– Often reports are hidden away in remote places on the web

(if they make it onto the internet)

• Need to learn from pilots in rapidly changing areas such as climate change- but rarely shared

Examples of good practice• DAC Evaluation

Resource Centre• World Bank Evaluation

Department• Danida• ALNAP• 3ie• IFAD

• But…– Focused on large scale impact

evaluations– How much do smaller

evaluations, or those using different methodologies, get shared?

– How do policy-makers or practitioners access the information?

– How much synthesis is being done and is that shared beyond organisational boundaries?

Understanding how evaluation influences

• Can learn from research influence and uptake• What works (experiences from IDS):– ‘Sticky messages’ / Rallying ideas– ‘Knit working’- building coalitions of connectors

and champions– Strategic opportunism – identifying windows of

opportunity for impact/influence• Challenge of evaluating the influence of

evaluations on policy and practice

Increasing the use of evaluations in policy and practice

• Availability on websites important, but doesn’t necessarily mean it will be used

• Understand how target groups search for, access and use information

• Information literacy• Incentives to look for and use

evaluations; incentives for organisational learning

• Need multiple communication strategies

What can we learn from research communications?

• Timeliness and relevance

• Editing, summarising

• Brevity and clear messages

• Credibility and quality

• Synthesis important

• Marketing • Networking and

multi-way communication

• Being systematic and opportunistic

• Requires variety of different skills

Target groups• Identify different target

groups and tailor communication strategies

• Involve networks and communities of practice throughout the evaluation process

Multiple communication approaches

• Different tools– Print– Seminars– Toolkits– Email updates– Online discussions– Visual– Blogs– Pod casts– CD Roms/USB sticks– Policy briefings

• Different channels– Traditional academic:

e.g. journals, conferences, research networks

– Direct stakeholder involvement

– Practitioner and advocacy networks

– Information and Knowledge intermediaries

Conclusions• Building in incentives to

communicate evaluations

• Learn from experience of research communications– Tailored approaches for

different audiences– Build communications in

from the start

• Role of information and knowledge intermediaries

• Horizontal learning and accountability:– Involving and sharing

learning with wider range of stakeholders, including networks and communities of practice, throughout the evaluation process

Questions for the future• How can context specific and potentially sensitive evaluations be

shared, adapted and applied beyond the programme context? • How do we assess the influence of evaluations on policy/practice?• How can ‘decision-makers’ be encouraged and supported to use

evaluations from other contexts/programmes for evaluation informed decision-making?

• What strategies/channels/methods are effective in communicating evaluations beyond the specific programme context?

• What kinds of networks and communities could both benefit from, and add insight to, the final conclusions of an evaluation itself?

• Share your views: www.alineplanning.org

top related