Posts filed under ‘Trainings, Seminars & Conferences’
I’m happy to announce that next week – Friday April 26th – I’ll be participating in a webinar presented by SEA Change community of practice. Further details and link to registration:
This webinar titled “Presenting evaluation findings and evidence-based organizational learning” has been developed based on requests by SEA Change CoP members to learn more about presenting evaluation findings and the pre-requisites for their organization to better learn from evaluation findings.
Mr. Glenn O’Neil, evaluation expert and communications professional, will look at how to present evaluation finding with a focus on different messages for different stakeholders in order to improve the use of evaluation findings. After that he will discuss what it takes to become an evidence-based learning organization and give some guidance on how to improve evidence-based learning in your organization.
Join us for a Webinar on April 26 (14:00 ICT / 9:00 CET).
- Define What Is To Be Evaluated
- Understand Causes of Outcomes and Impacts
- Synthesise Data from One or More Evaluations
- Report and Support Use of Findings
“Unlocking business performance- The role of communications research and analytics” is the theme of the 5th European Summit on Measurement of PR and communications – to be held in Madrid from 5 -7 June, 2013.
The Summit is presented by AMEC, the International Association for the Measurement and Evaluation of Communication. The Summit will include workshop sessions, discussion groups and plenary speaking spots by international senior communications professionals.
A two day workshop on “realist evaluation” is planned for 3-4 May 2012 in Berne, Switzerland. some more information:
Realist evaluation (Pawson and Tilley, 1997) is a form of theory driven evaluation. It starts by assuming that nothing works for everyone ; that how participants respond matters just as much as what programs do ; and that context does indeed make a difference. It doesn’t ask “Does this work ?”, but “In what contexts does this work, for whom, to what extent, and how ?”. Realist evaluation can be used at any level from individual casework to international development, and with simple, complicated and complex programs.
This two day workshop will introduce the key concepts of realist evaluation and their implications for evaluation design, methods and the role of the evaluator ; introduce the key ideas from complexity theory that are useful in realist evaluation ; explain the structure of realist program theory and provide practice in developing realist program theories for simple, complicated and complex programs… The program will involve a combination of presentations, discussion, and exercises.
The course is being organised by the University of Fribourg, Sociology, Social Policy and Social Work Department in collaboration with Marlène Läubli Loud of LAUCO Evaluation and Training and the Swiss Evaluation Society (SEVAL).
Further information and registration:
Together with Raj Rana, I will be running a workshop on communications and evaluation this coming November in Bern, Switzerland, further information:
Integrating communications in evaluation
Date and place : 10-11 November 2011, Bern
Organisers: University of Fribourg & Swiss Evaluation Society
An often-overlooked step of evaluation is ensuring that findings are communicated, understood and acted upon. Communicating throughout the evaluation process equally poses many challenges. Communicating effectively implies using different means, messages and methods to reach different stakeholder groups, with very different needs and expectations.
A mix of presentations, case studies and practical exercises will be used to promote new approaches for communicating results including social media, interactive presentations and data visualization. The workshop delivery will include a mix of facilitation techniques to introduce effective means of engaging stakeholders in the evaluation process (World Café methodology, buzz groups, visualization techniques, developing consensus, etc.) Participants are encouraged to bring examples of evaluations they have commissioned/implemented, to be used as case studies during the workshop.
The annual North American Summit on communication/PR measurement is coming up in September 2011:
Since it began in 2003, the North American Summit on Public Relations Measurement has enjoyed an international reputation for being one of the world’s leading annual conferences about research, measurement and evaluation in communications and public relations.
Each year this event features a number of unique, hands-on pre-conference workshops along with a day and a half of superb program sessions focusing on how measurement is being used effectively throughout the communications industry. This measurement summit is also noted for having several superb networking events where attendees have opportunities to exchange insights with international experts.
Through lectures, case studies and interactive discussions led by some of the world’s most noted measurement experts, the North American Summit on Public Relations Measurement annually exposes conference delegates to innovations, methodologies and best practices from some of the world’s most successful public relations measurement programs.
On Tuesday, 15 March 2011 our friends and colleagues Professor Tom Watson & Professor Dr Ansgar Zerfass will be giving a lecture at Bournemouth University “The evolution of evaluation – Public Relations’ erratic path to the measurement of effectiveness” and “Corporate communications revisited – How communication drives corporate value and legitimisation”
The lecture is part of Bournemouth University’s free lecture series, open to everyone, giving you a unique insight into the knowledge and research that they are renowned for. Previous lectures attracted a varied audience, including fellow academics, businesses and members of the local community who learned how BU’s research is helping to change the world for the better.
You are welcome to attend as many lectures as you wish. Register online >>
Lectures will be held in the Executive Business Centre, 89 Holdenhurst Road, Bournemouth BH8 8EB, UK (Lansdowne Campus) with registration and refreshments available from 5pm.
Here is an interesting workshop on “Realist evaluation”, taking place in Wageningen, the Netherlands on 29 March 2011. More information from the organisers:
Realist evaluation – understanding how programs work in their context.
‘Realist evaluation (Pawson and Tilley, 1997) is one type of theory based evaluation. It aims to explore “what works, for whom, in what contexts, to what extent and how”. It adopts a particular understanding of how programs work, and uses a particular format for program theories to help guide evaluation design, data collection and analysis.
Realist evaluation has a particular focus on understanding the interactions between programs and their contexts and the ways that these influence how programs work. Evaluation expert Dr. Gill Westhorp will discuss the concepts and assumptions that underpin this theory based evaluation approach. What is it that realist evaluation brings to the table of evaluating development programs? How is it different from existing approaches in evaluation in development? How does it understand, and deal with, complexity? What new insights can help strengthen the utility of evaluation for development?
During the morning, Gill will introduce the basic assumptions and key concepts in realist evaluation. She will also briefly demonstrate how these ideas can be built into other evaluation models using two examples. These models – realist action research and realist program logic – are participatory models which were designed for use in settings where limited resources, lack of capacity to collect outcomes data, complex programs, and (sometimes) small participant numbers make evaluation difficult. In the afternoon, the practical implications for evaluation design, data collection and analysis will be discussed. Examples and practical exercises will be included throughout the day.
Seminar with Dr. Patton : ‘Developmental evaluation – new kid on the evaluation block’: 29th March 2011, the Netherlands
Here is information on a seminar coming up with Dr Michael Quinn Patton – I’ve just bought his new book which is a very good read:
Date: 29th March 2011 in the Netherlands: ‘Developmental evaluation – new kid on the evaluation block’.
Description: Developmental evaluation is based on insights from complex dynamic systems, uncertainty, nonlinearity, and emergence. World renowned, award-winning evaluation expert Dr. Michael Quinn Patton will discuss the developmental evaluation framework as detailed in his book `Developmental Evaluation: Applying Complexity Concepts to Enhance Innovation and Use’. Patton will illustrate how developmental evaluation can be used for a range of purposes: ongoing program development; adapting effective principles of practice to local contexts; generating innovations and taking them to scale; and facilitating rapid response in crisis situations.
During the morning, Dr. Patton will explain developmental evaluation and illustrate it with many examples from his own experience. In the afternoon, participants will debate the practical application of developmental evaluation in development, based on participants’ existing evaluation questions.
For more info and registration: http://www.cdi.wur.nl/UK/newsagenda/agenda/DevelopmentalEvaluation_MichaelPatton
Here is an interesting online course on Participatory Monitoring and Evaluation from International Institute for Development and Colorado State University:
Description: This course stresses participatory methods in monitoring and evaluation for community development, where multiple stakeholders are involved in the process of planning, collecting, interpreting, communicating, and using information. This approach emphasizes a regular monitoring process that leads to continuous improvements. The course uses a case study and team discussions to illustrate the participatory monitoring and evaluation process.
The course starts next on 18 March 2011. More information >>