SlideShare una empresa de Scribd logo
1 de 6
Descargar para leer sin conexión
299Fifth RUFORUM Biennial Regional Conference 17 - 21 October 2016, Cape Town, South Africa
A paradigm shift from monitoring and evaluation to interdisciplinary evaluative
systems thinking: The status of evaluative processes and use in
African Universities
Owuor, C., Waswa, M., Osiru, M. &Adipala, E.
Regional Universities Forum for Capacity Building inAgriculture (RUFORUM), P. O. Box 1681,
Wandegeya- Kampala, Uganda
Corresponding author: c.owuor@ruforum.org
Abstract
This paper is an initial attempt to assess the status and use of evaluative processes in
African universities to inform their strategic decisions in face of growing demands for
universities to transform and be more relevant in development and innovation processes.
The desired transformation is part of the broad organisational change process that requires
universities to adapt to meet the ever increasing complex development challenges of national
and regional scope. The study draws from evaluation studies and specific reports over a 10
year time horizon in over 20 African universities in 10 countries. The finding reveals that
universities have fully established Planning Units with clear operational strategies and
monitoring frameworks. However in practice, monitoring is largely activity based and used
to assess compliance with work plans and budgets and few universities have well defined
monitoring, evaluation, and learning strategies and where they exist, implementation is limited
or at most, ad hoc. The paper seeks to build on the body of growing literature on evaluative
thinking and the need for entities to put in place adaptive systems for learning and improving.
Key words: African universities, evaluation process, organisation change process
Résumé
Ce document est une première tentative pour évaluer l’état et l’utilisation des processus
d’évaluation dans les universités africaines afin d’éclairer les décisions stratégiques de ces
dernières face à la demande croissante aux Universités de se transformer et d’augmenter
leur visibilité dans les processus de développement d’innovations. La transformation souhaitée
fait partie du vaste processus de changement organisationnel qui demande que les universités
changent pour répondre aux défis croissants et complexes de développement à l’étendue
nationale et régionale. L’étude tire des examens d’évaluation et des rapports spécifiques de
20 Universités de 10 pays et ceci sur une période de 10 ans. Les résultats révèlent que les
universités ont pleinement mis en place des unités de planification avec des stratégies
opérationnelles claires et des cadres de suivi. Cependant, dans la pratique, le suivi est l’activité
appliquée à une grande base et utilisée pour évaluer la conformité des plans de travail et
RUFORUM Working Document Series (ISSN 1607-9345) No. 14 (1): 299-304.
Available from http://repository.ruforum.org
Research Application Summary
300 Owuor, C. et al.
plans budgétaires. Quelques universités ont des stratégies de suivi-évaluation et
d’apprentissage bien définies et là où elles existent, leur mise en œuvre est limitée ou plus
appropriée. Le document cherche à enrichir la littérature croissante sur la réflexion évaluative
et la nécessité pour les entités à mettre en place des systèmes adaptatifs pour l’apprentissage
et le perfectionnement.
Mots clés: Universités africaines, processus d’évaluation, processus de changement
organisationnel
Introduction
Organizations are continuously grappling with the use of findings from evaluative processes
such as performance evaluation, development evaluations, formative evaluations, cost
effective studies and cases studies to inform their internal systems, improve operational
efficiency and strengthen organizational learning (Suarez-Balcazar and Taylor-Ritzer, 2014).
A learning organization is conditioned to adapt and improve on its performance and contribute
to innovation for evaluation (Eilertsen and London, 2005). This necessitates putting in place
systems and structures that foster generation of lessons for learning and knowledge creation.
The learning processes lead to creation of organization knowledge through “learning the
how”, a vital component for organizational effectiveness (Edmondson and Moingeon, 1996).
According to Schein (2004), organizations do the same things all the time until the “tried”
and “true” way no longer holds. This is due to the learning anxiety that is produced by the
need and shift to learn something new. The organizational “learning how” and “learning
why” therefore, determines its capability to compete in open market space and to deal with
changing market environment (Brown and Duguid, 1996; Grant, 1996). This learning emanate
from multiple sources but more critically, through evaluative processes. Substantive
organisational readiness for evaluation is favourably influenced through direct evaluation
capacitybuilding,stimulatingcriticalthinking,supportreflectivepracticeandfacilitatinginternal
evaluation culture to promote organisational learning also referred to as “evaluative thinking”
(Cousins, 2004; McCoy et al., 2013; Befani et al., 2015). The concept of evaluative thinking
continue to receive significant attention in both public and not-for-profit sectors as it seeks
to contribute to new learning by providing evidence to chronicle, map and monitor progress,
successes, failures and roadblocks in the innovation as it unfolds (Buckley and Archibald,
2013; Earl and Timperley, 2015). There are quite striking differences in organisational
capacities to use evaluation in government agencies and voluntary organisations (Cousins et
al., 2014). Though, the underlying impetus is the ever increasing need for organisations to
build internal evaluative capacities to inform sustainable organisational and cultural change
(McCoy et al., 2013).
In this respect, it is now a standard project management practice to embed evaluative actions
during programme design. Blending the programme design consideration with the concept
of evaluative thinking, leads to better understanding of organisational operating context,
setting clear organizational learning questions, improve the theory of change through
questioning the underlying assumptions and intervention logic, and enriching stakeholder
engagement in the entire implementation cycle (Bonbright, 2013). The other dimension that
301Fifth RUFORUM Biennial Regional Conference 17 - 21 October 2016, Cape Town, South Africa
sets ground for evaluative processes in organizations, is the ever increasing incentives and
rewards systems that is based on the growing climate of accountability that demand for
evidence of results and transparent based implementation approaches. Major multilateral
and institutional agencies have instituted robust accountability frameworks that use rigorous
methods to prove impact and to the least, the effect of interventions and their investments.
In 2012, the US Office of Management and Budget issued new standards that promote
programmes demonstrating rigorous studies of effectiveness setting stringent impact
measurement standards (USGOV, 2013). Similarly other agencies soon followed suit.
In respect to these global trends in demands for increased accountability, the Regional
Universities Forum for Capacity Building in Agriculture (RUFORUM) institutionalized a
project management policy of conducting evaluative activities of each initiative and assess
its contribution to the organisation mission and vision of “High performing African
Universities that produce skilled, proactive graduates, demand driven research outputs
and innovations in response to local, regional and national agricultural development
priorities”. The purpose for this is threefold; (a) to provide opportunities of better
understanding of the realities and therefore institute more realistic plans, (b) ensure timely
identification of bottlenecks to inform adaptive actions, and (c) whether the initiatives have
achieved their intended impact. Over the last 10 years, RUFORUM successfully implemented
a number of projects that have left an indelible mark of impact and transformation of agricultural
higher education landscape in Africa.
Methodology
In this paper, we attempt to draw lessons from RUFORUM engagement with universities
over the last 10 years to strengthen the monitoring and evaluation functions. The analysis is
not based on the programmatic achievements of each initiative, but rather on the cross
cutting elements of the need to build a body of credible evidence for policy engagement and
partnerships. The analysis is informed by the growing demand for development actors to
institute internal evaluation adaptive systems, influence policy decisions and build multi-level
collaborations.
Current evaluative processes in Africa in Universities
The cross-cutting thread in all RUFORUM initiatives is the capacity building element at
both individual and institutional levels. At individual level, the results are clear cut, with
performance measurement metrics taking into consideration parameters like the number of
graduates trained, scholarships awarded, academic mobility and alumni placement. These
metrics are standard regardless of the training institution. Therefore the individual capacity
development outcomes are clearly understood and definition of success level well specified,
and easy to ascertain the chain of cause and effect. However, at institutional level (University),
the assessment and measurement of institutional capacity development is a complex affair.
This comes at the backdrop of growing global trend in demands for increased accountability
that have not spared the Africa universities either. The demand towards performance
management, award of service contracts, ISO certification, and impact of university training
302 Owuor, C. et al.
and research on technology and innovations development and the need for broad
accountability is leading to growing demand for evaluative systems to support decision making.
On the outlook most Africa universities have fully established Planning Units with clear
operational strategies and monitoring frameworks. However in practice, monitoring is largely
activity based and used to assess compliance with work plans and budgets and few universities
have well defined monitoring, evaluation, and learning strategies and where they exist,
implementation is limited or at most, ad hoc. Most universities have clear procedures for
QualityAssurance (QA) and this is seen as a fundamental proposition that ground their core
training and research functions. We observe that universities are setting higher benchmarks
above the basic requirements of their respective National Councils of Higher Education
regulations and other regional quality assurance framework. This is a right ambition for
African universities as it has potential to cause their transformation into world class
universities. We observe that universities are increasingly investing in tracer studies, a widely
accepted approach for tracking impact of universities thought benchmark indices such as
proportion of past students who find employment, sectors in which they work, their job
performance, and employer satisfaction with them. The tracer study is a powerful strategic
evaluative tool for universities to support curricula reviews, designing new market relevant
courses and training students ready for the market.
Whereas universities hold a pool of knowledge, evaluations of research projects are mostly
externally driven and or restricted to projects undertaken in collaboration with governments,
Non-Governmental Organizations (NGOs) or CGIAR centres. This implies that mostAfrica
Universities have inadequately constituted internal evaluation systems to meet external
accountability and satisfy their organisational learning requirements. Most universities have
instituted information management systems to effectively improve students’ record
management. This is seen as a bold step. However, we observe that there is little cataloguing
of research findings and how they are being used. In general terms, there is limited capacity
to implement monitoring and evaluative actions in most universities. In the next section, we
provide an example of how RUFORUM uses evaluative processes to engage in policy
dialogue and the associated challenges in doing so.
RUFORUM made a strategic shift to engage at multiple levels in Policy advocacy as a
transitional step to become a champion and voice for agricultural higher education inAfrica.
However, from the monitoring and evaluation perspective, the challenges lies in estimating
the attribution and contribution effects in face of actions by other actors of the resulting
regional and continental policy shifts. The increased organizational engagement in advocacy
calls for reflective and continuous assessment progress markers to elicit flexibility and
responsiveness to capture incremental qualitative output for purposes of estimating impact.
In this case, the need for numbers to qualify results is not a necessary requirement and most
likely the quantitative indices turn out to be meaningless as the basis for assessing progress
is premised on an arbitrary scale that is not universally understood. The remedy for this
challenge is to have measured engagement with teams to reflect on emerging trends,
harmonise processes and systems. This necessitates continuous acquisition of “political
intelligence” and having adaptive monitoring and evaluation systems to capture the engagement
processes. The presence of promoters and like-minded actors in positions of influence may
303Fifth RUFORUM Biennial Regional Conference 17 - 21 October 2016, Cape Town, South Africa
considerably hasten the achievement of the end result and in the process, render the
implementation strategies obsolete.
Conclusion
Evaluative thinking which is the on-going systematic practice of asking questions, questioning
assumption, collecting and validating and analysing evidence and using evidence to inform
decision is gaining credibility among development actors. This new paradigm will necessitate
rethinking of University organisational learning and development systems and instituting
adaptive frameworks to support their core roles of training, research and outreach.
References
Befani, B., Ramalingam, B. and Stern, E. 2015. Introduction – Towards systemic approaches
to evaluation and impact. IDS Bulletin 46: 1–6. doi:10.1111/1759-5436.12116
Bonbright, D. 2012. Using impact evaluations results. Impact evaluation Notes. No.4
November 2012. Key Stone Accountability.
Bradley, J., Swee, C., Goh, C., Lark, S. and Lee, L.E. 2004. Integrating evaluative inquiry
into the organizational culture: a review and synthesis of the knowledge base. The
Canadian Journal of Program Evaluation 19 (2): 99–141. ISSN 0834-1516. Canadian
Evaluation Society.
Brown, J.S. and Duguid, P. 1996. Organizational learning and communities-of-practice:
Toward a unified view of working, learning and innovation. In: Cohen, M.D. and Sproull,
L.S. (Eds), Organizational Learning. Thousand Oaks, CA: Sage.
Buckley, J. andArchibald, T. 2013. EvaluativeThinking: Principles and Practices to Enhance
Evaluation Capacity and Quality. Professional Development Workshop offered at the
Annual Conference of the American Evaluation Association on October 20, 2013,
Washington, D.C.
Buckley, J., Archibald, T., Hargraves, M. and Trochim, W. M. 2015. Defining and teaching
evaluative thinking. Insights from research on critical thinking. American Journal of
Evaluation 36 (3): 375-388. doi: 10.1177/1098214015581706
Cousinsa, J.B., Gohb, S.C., Elliott, C., Aubryc, T. and Gilbert, N. 2015. Government and
voluntary sector differences in organizational capacity to do and use evaluation.
Evaluation and Program Planning 4: 1–13.
Earl, L. and Timperley, H. 2015. Evaluative thinking for successful educational innovation.
ECD Education Working. Papers. No. 122, OECD Publishing, Paris. http://dx.doi.org/
10.1787/5jrxtk1jtdwf-en.
Eilerstsen, S. and London, K.M.A. 2005. Modes of Organisational Learning. The Kollner
Group, Inc.
Grant, R. 1996. Prospering in dynamically competitive environments: Organizationalcapability
as knowledge integration. Organization Science 7 (4): 375-387.
McCoy, A., Rose, D. and Connolly, M. 2013. Developing evaluation cultures in human
service organisations [online]. Evaluation Journal of Australasia 13 (1): 15-20.
Availability: http://search.informit.com.au/documentSummary;dn=59920164573607;
res=IELNZC> ISSN: 1035-719X. [cited 21 Sep 16].
304 Owuor, C. et al.
Schein, E.H. 2004. A case of Organisational (Cultural?) Change. Organisational Culture
and Leadership. Third Edition. Jossu-Bass a Willy Imprint. 365pp.
Suarez-Balcazar Yolanda and Taylor-Ritzler, T. 2014. Moving from science to practice in
evaluation capacity building. American Journal of Evaluation 35: 95 - 99. August 15,
2013, doi: 10.1177/1098214013499440
Teshome, A. and Tully, R.C. 2012. Higher education partnerships: Experiences and impact
in Sub-Saharan Africa. RUFORUM Third Biennial Conference Proceedings, Entebbe,
Uganda, 24-28 September 2012. pp.1163-1177.
USAGov 2013. 78590 Federal Register Vol. 78, No. 248 / Thursday, December 26, 2013 /
Rules and Regulations OFFICE OF MANAGEMENTAND BUDGET 2 CFR Chapter
I, and Chapter II, Parts 200, 215, 220, 225, and 230 UniformAdministrative Requirements,
Cost Principles, andAudit Requirements for FederalAwards Vol. 78, No. 248 / Thursday,
December 26, 2013 / Rules and Regulations.

Más contenido relacionado

La actualidad más candente

A Performance Appraisal of the Pedagogic in Service Training Programme ISTP M...
A Performance Appraisal of the Pedagogic in Service Training Programme ISTP M...A Performance Appraisal of the Pedagogic in Service Training Programme ISTP M...
A Performance Appraisal of the Pedagogic in Service Training Programme ISTP M...
ijtsrd
 
Quality Assurance and institutional accreditation performance indicators and ...
Quality Assurance and institutional accreditation performance indicators and ...Quality Assurance and institutional accreditation performance indicators and ...
Quality Assurance and institutional accreditation performance indicators and ...
Ganesh Shukla
 

La actualidad más candente (20)

Building a Quality Management System in Higher Education
Building a Quality Management System in Higher EducationBuilding a Quality Management System in Higher Education
Building a Quality Management System in Higher Education
 
Dep ed order no 44 s 2010
Dep ed order no 44 s 2010Dep ed order no 44 s 2010
Dep ed order no 44 s 2010
 
Role and dynamics of corporate R&D
Role and dynamics of corporate R&DRole and dynamics of corporate R&D
Role and dynamics of corporate R&D
 
PRG Poster_AEA2015
PRG Poster_AEA2015PRG Poster_AEA2015
PRG Poster_AEA2015
 
Cqaf vet larissa ali presentation
Cqaf vet larissa ali presentationCqaf vet larissa ali presentation
Cqaf vet larissa ali presentation
 
Academic & administrative audit
Academic & administrative audit Academic & administrative audit
Academic & administrative audit
 
LEADERSHIP Ed. 33 - CPC
LEADERSHIP Ed. 33 - CPCLEADERSHIP Ed. 33 - CPC
LEADERSHIP Ed. 33 - CPC
 
A Performance Appraisal of the Pedagogic in Service Training Programme ISTP M...
A Performance Appraisal of the Pedagogic in Service Training Programme ISTP M...A Performance Appraisal of the Pedagogic in Service Training Programme ISTP M...
A Performance Appraisal of the Pedagogic in Service Training Programme ISTP M...
 
An Exploration of the Imperatives for Successful Strategy Execution in ODL In...
An Exploration of the Imperatives for Successful Strategy Execution in ODL In...An Exploration of the Imperatives for Successful Strategy Execution in ODL In...
An Exploration of the Imperatives for Successful Strategy Execution in ODL In...
 
A Brief expression on National Assessment and Accreditation Council (NAAC) Ma...
A Brief expression on National Assessment and Accreditation Council (NAAC) Ma...A Brief expression on National Assessment and Accreditation Council (NAAC) Ma...
A Brief expression on National Assessment and Accreditation Council (NAAC) Ma...
 
Evidence On Trial: weighing the value of evidence in academic enquiry, policy...
Evidence On Trial: weighing the value of evidence in academic enquiry, policy...Evidence On Trial: weighing the value of evidence in academic enquiry, policy...
Evidence On Trial: weighing the value of evidence in academic enquiry, policy...
 
The IAMU-PAESP Project Implementation Benefits : The MAAP Experience
The IAMU-PAESP Project Implementation Benefits : The MAAP Experience The IAMU-PAESP Project Implementation Benefits : The MAAP Experience
The IAMU-PAESP Project Implementation Benefits : The MAAP Experience
 
Using Data to Improve Minority-Serving Institution Success
Using Data to Improve Minority-Serving Institution SuccessUsing Data to Improve Minority-Serving Institution Success
Using Data to Improve Minority-Serving Institution Success
 
Quality Assurance and institutional accreditation performance indicators and ...
Quality Assurance and institutional accreditation performance indicators and ...Quality Assurance and institutional accreditation performance indicators and ...
Quality Assurance and institutional accreditation performance indicators and ...
 
Naac criterion 6 Governance, leadership and Management
Naac criterion 6 Governance, leadership and ManagementNaac criterion 6 Governance, leadership and Management
Naac criterion 6 Governance, leadership and Management
 
Evaluating Student Success Initiatives
Evaluating Student Success InitiativesEvaluating Student Success Initiatives
Evaluating Student Success Initiatives
 
Usfq&puce research
Usfq&puce researchUsfq&puce research
Usfq&puce research
 
International Journal of Business and Management Invention (IJBMI)
International Journal of Business and Management Invention (IJBMI)International Journal of Business and Management Invention (IJBMI)
International Journal of Business and Management Invention (IJBMI)
 
Framework for ranking higher institutions in nigeria
Framework for ranking higher institutions in nigeriaFramework for ranking higher institutions in nigeria
Framework for ranking higher institutions in nigeria
 
International Journal of Business and Management Invention (IJBMI)
International Journal of Business and Management Invention (IJBMI)International Journal of Business and Management Invention (IJBMI)
International Journal of Business and Management Invention (IJBMI)
 

Destacado

Instructor powerpoint
Instructor powerpointInstructor powerpoint
Instructor powerpoint
tanglin
 
ecdl_windows_8_office_2013_biblia_minta
ecdl_windows_8_office_2013_biblia_mintaecdl_windows_8_office_2013_biblia_minta
ecdl_windows_8_office_2013_biblia_minta
Krist P
 
Wonder Woman Feminist Icon 1031 (3)
Wonder Woman Feminist Icon 1031 (3)Wonder Woman Feminist Icon 1031 (3)
Wonder Woman Feminist Icon 1031 (3)
Simmons Jessie
 
Yousef Aburub - Resume - Final(1)
Yousef Aburub - Resume - Final(1)Yousef Aburub - Resume - Final(1)
Yousef Aburub - Resume - Final(1)
Yousef Aburub
 
Etude Prêt à porter féminin : Qu'est ce qui incite les femmes à se rendre en ...
Etude Prêt à porter féminin : Qu'est ce qui incite les femmes à se rendre en ...Etude Prêt à porter féminin : Qu'est ce qui incite les femmes à se rendre en ...
Etude Prêt à porter féminin : Qu'est ce qui incite les femmes à se rendre en ...
Clotilde Chenevoy
 

Destacado (18)

Instructor powerpoint
Instructor powerpointInstructor powerpoint
Instructor powerpoint
 
ecdl_windows_8_office_2013_biblia_minta
ecdl_windows_8_office_2013_biblia_mintaecdl_windows_8_office_2013_biblia_minta
ecdl_windows_8_office_2013_biblia_minta
 
Trend 2025 tanyer sonmezer finansbank
Trend 2025 tanyer sonmezer finansbankTrend 2025 tanyer sonmezer finansbank
Trend 2025 tanyer sonmezer finansbank
 
Wonder Woman Feminist Icon 1031 (3)
Wonder Woman Feminist Icon 1031 (3)Wonder Woman Feminist Icon 1031 (3)
Wonder Woman Feminist Icon 1031 (3)
 
Tvigle & Media3 - NOAH13 London
Tvigle & Media3 - NOAH13 LondonTvigle & Media3 - NOAH13 London
Tvigle & Media3 - NOAH13 London
 
Yousef Aburub - Resume - Final(1)
Yousef Aburub - Resume - Final(1)Yousef Aburub - Resume - Final(1)
Yousef Aburub - Resume - Final(1)
 
Conceptboek Grill & Chill
Conceptboek Grill & ChillConceptboek Grill & Chill
Conceptboek Grill & Chill
 
Taxonomies in Search
Taxonomies in SearchTaxonomies in Search
Taxonomies in Search
 
CCNAS Ch01
CCNAS Ch01 CCNAS Ch01
CCNAS Ch01
 
5 Signs You Are In A Waterfall Agile Transformation
5 Signs You Are In A Waterfall Agile Transformation5 Signs You Are In A Waterfall Agile Transformation
5 Signs You Are In A Waterfall Agile Transformation
 
CIF16: Knock, Knock: Unikernels Calling! (Richard Mortier, Cambridge University)
CIF16: Knock, Knock: Unikernels Calling! (Richard Mortier, Cambridge University)CIF16: Knock, Knock: Unikernels Calling! (Richard Mortier, Cambridge University)
CIF16: Knock, Knock: Unikernels Calling! (Richard Mortier, Cambridge University)
 
It’s A Trap! Don't fly into budget airline traps!
It’s A Trap! Don't fly into budget airline traps!It’s A Trap! Don't fly into budget airline traps!
It’s A Trap! Don't fly into budget airline traps!
 
Etude Prêt à porter féminin : Qu'est ce qui incite les femmes à se rendre en ...
Etude Prêt à porter féminin : Qu'est ce qui incite les femmes à se rendre en ...Etude Prêt à porter féminin : Qu'est ce qui incite les femmes à se rendre en ...
Etude Prêt à porter féminin : Qu'est ce qui incite les femmes à se rendre en ...
 
CIF16: Unikernels, Meet Docker! Containing Unikernels (Richard Mortier, Anil ...
CIF16: Unikernels, Meet Docker! Containing Unikernels (Richard Mortier, Anil ...CIF16: Unikernels, Meet Docker! Containing Unikernels (Richard Mortier, Anil ...
CIF16: Unikernels, Meet Docker! Containing Unikernels (Richard Mortier, Anil ...
 
الفجوة الرقمية
الفجوة الرقميةالفجوة الرقمية
الفجوة الرقمية
 
XPDS16: CPUID handling for guests - Andrew Cooper, Citrix
XPDS16:  CPUID handling for guests - Andrew Cooper, CitrixXPDS16:  CPUID handling for guests - Andrew Cooper, Citrix
XPDS16: CPUID handling for guests - Andrew Cooper, Citrix
 
AWS Lambda from the Trenches
AWS Lambda from the TrenchesAWS Lambda from the Trenches
AWS Lambda from the Trenches
 
plan educacion_artistica
plan educacion_artisticaplan educacion_artistica
plan educacion_artistica
 

Similar a Owuor paradigm

Organizational Effectiveness of Naval State University: Proposed Institutiona...
Organizational Effectiveness of Naval State University: Proposed Institutiona...Organizational Effectiveness of Naval State University: Proposed Institutiona...
Organizational Effectiveness of Naval State University: Proposed Institutiona...
Dr. Amarjeet Singh
 
Institutional and Program Self-Evaluation (IPSE): Towards Institutional Susta...
Institutional and Program Self-Evaluation (IPSE): Towards Institutional Susta...Institutional and Program Self-Evaluation (IPSE): Towards Institutional Susta...
Institutional and Program Self-Evaluation (IPSE): Towards Institutional Susta...
IJAEMSJORNAL
 
CHAPTER 3. CulTivATiNg COllAbORATivE CulTuREsCollaborative inqu.docx
CHAPTER 3. CulTivATiNg COllAbORATivE CulTuREsCollaborative inqu.docxCHAPTER 3. CulTivATiNg COllAbORATivE CulTuREsCollaborative inqu.docx
CHAPTER 3. CulTivATiNg COllAbORATivE CulTuREsCollaborative inqu.docx
keturahhazelhurst
 
To Study the Role of Quality Management System to Improve Effectiveness of Qu...
To Study the Role of Quality Management System to Improve Effectiveness of Qu...To Study the Role of Quality Management System to Improve Effectiveness of Qu...
To Study the Role of Quality Management System to Improve Effectiveness of Qu...
lalikjan
 
3A - Effective evaluation of the impact of CPD - What does research tell us
3A - Effective evaluation of the impact of CPD - What does research tell us3A - Effective evaluation of the impact of CPD - What does research tell us
3A - Effective evaluation of the impact of CPD - What does research tell us
Sandra Wharton
 
Wasc Evaluator Training Webinar - July 13, 2011
Wasc Evaluator Training Webinar - July 13, 2011Wasc Evaluator Training Webinar - July 13, 2011
Wasc Evaluator Training Webinar - July 13, 2011
WASC Senior
 

Similar a Owuor paradigm (20)

Owuor African
Owuor AfricanOwuor African
Owuor African
 
An Organizational Development Framework For Assessing Readiness And Capacity ...
An Organizational Development Framework For Assessing Readiness And Capacity ...An Organizational Development Framework For Assessing Readiness And Capacity ...
An Organizational Development Framework For Assessing Readiness And Capacity ...
 
Strategic innovation and performance of public universities in kenya
Strategic innovation and performance of public universities in kenyaStrategic innovation and performance of public universities in kenya
Strategic innovation and performance of public universities in kenya
 
Organizational Effectiveness of Naval State University: Proposed Institutiona...
Organizational Effectiveness of Naval State University: Proposed Institutiona...Organizational Effectiveness of Naval State University: Proposed Institutiona...
Organizational Effectiveness of Naval State University: Proposed Institutiona...
 
Institutional and Program Self-Evaluation (IPSE): Towards Institutional Susta...
Institutional and Program Self-Evaluation (IPSE): Towards Institutional Susta...Institutional and Program Self-Evaluation (IPSE): Towards Institutional Susta...
Institutional and Program Self-Evaluation (IPSE): Towards Institutional Susta...
 
AUTONOMY IN HIGHER EDUCATION
AUTONOMY IN HIGHER EDUCATIONAUTONOMY IN HIGHER EDUCATION
AUTONOMY IN HIGHER EDUCATION
 
Exploring scholarship and scholarly activity in college-based Higher Education
Exploring scholarship and scholarly activity in college-based Higher EducationExploring scholarship and scholarly activity in college-based Higher Education
Exploring scholarship and scholarly activity in college-based Higher Education
 
A370111.pdf
A370111.pdfA370111.pdf
A370111.pdf
 
Quality is Never an Accident: A Survey on the Total Quality Management Practices
Quality is Never an Accident: A Survey on the Total Quality Management PracticesQuality is Never an Accident: A Survey on the Total Quality Management Practices
Quality is Never an Accident: A Survey on the Total Quality Management Practices
 
Basheer 2013 calidad total y dllo organizativo
Basheer 2013 calidad total y dllo organizativoBasheer 2013 calidad total y dllo organizativo
Basheer 2013 calidad total y dllo organizativo
 
CHAPTER 3. CulTivATiNg COllAbORATivE CulTuREsCollaborative inqu.docx
CHAPTER 3. CulTivATiNg COllAbORATivE CulTuREsCollaborative inqu.docxCHAPTER 3. CulTivATiNg COllAbORATivE CulTuREsCollaborative inqu.docx
CHAPTER 3. CulTivATiNg COllAbORATivE CulTuREsCollaborative inqu.docx
 
3 ms edu_quality_assurance
3 ms edu_quality_assurance3 ms edu_quality_assurance
3 ms edu_quality_assurance
 
To Study the Role of Quality Management System to Improve Effectiveness of Qu...
To Study the Role of Quality Management System to Improve Effectiveness of Qu...To Study the Role of Quality Management System to Improve Effectiveness of Qu...
To Study the Role of Quality Management System to Improve Effectiveness of Qu...
 
Cemca newsletter-vol1-no3-september-2015
Cemca newsletter-vol1-no3-september-2015Cemca newsletter-vol1-no3-september-2015
Cemca newsletter-vol1-no3-september-2015
 
3A - Effective evaluation of the impact of CPD - What does research tell us
3A - Effective evaluation of the impact of CPD - What does research tell us3A - Effective evaluation of the impact of CPD - What does research tell us
3A - Effective evaluation of the impact of CPD - What does research tell us
 
Project Management And The Sustainability Of Organizations From The Knowledge...
Project Management And The Sustainability Of Organizations From The Knowledge...Project Management And The Sustainability Of Organizations From The Knowledge...
Project Management And The Sustainability Of Organizations From The Knowledge...
 
Assessment in higher education final draft 05-11-16 (1) final
Assessment in higher education   final draft 05-11-16 (1) finalAssessment in higher education   final draft 05-11-16 (1) final
Assessment in higher education final draft 05-11-16 (1) final
 
Wasc Evaluator Training Webinar - July 13, 2011
Wasc Evaluator Training Webinar - July 13, 2011Wasc Evaluator Training Webinar - July 13, 2011
Wasc Evaluator Training Webinar - July 13, 2011
 
Application of the value chain analysis framework
Application of the value chain analysis frameworkApplication of the value chain analysis framework
Application of the value chain analysis framework
 
Performance Assessment of Agricultural Research Organisation Priority Setting...
Performance Assessment of Agricultural Research Organisation Priority Setting...Performance Assessment of Agricultural Research Organisation Priority Setting...
Performance Assessment of Agricultural Research Organisation Priority Setting...
 

Owuor paradigm

  • 1. 299Fifth RUFORUM Biennial Regional Conference 17 - 21 October 2016, Cape Town, South Africa A paradigm shift from monitoring and evaluation to interdisciplinary evaluative systems thinking: The status of evaluative processes and use in African Universities Owuor, C., Waswa, M., Osiru, M. &Adipala, E. Regional Universities Forum for Capacity Building inAgriculture (RUFORUM), P. O. Box 1681, Wandegeya- Kampala, Uganda Corresponding author: c.owuor@ruforum.org Abstract This paper is an initial attempt to assess the status and use of evaluative processes in African universities to inform their strategic decisions in face of growing demands for universities to transform and be more relevant in development and innovation processes. The desired transformation is part of the broad organisational change process that requires universities to adapt to meet the ever increasing complex development challenges of national and regional scope. The study draws from evaluation studies and specific reports over a 10 year time horizon in over 20 African universities in 10 countries. The finding reveals that universities have fully established Planning Units with clear operational strategies and monitoring frameworks. However in practice, monitoring is largely activity based and used to assess compliance with work plans and budgets and few universities have well defined monitoring, evaluation, and learning strategies and where they exist, implementation is limited or at most, ad hoc. The paper seeks to build on the body of growing literature on evaluative thinking and the need for entities to put in place adaptive systems for learning and improving. Key words: African universities, evaluation process, organisation change process Résumé Ce document est une première tentative pour évaluer l’état et l’utilisation des processus d’évaluation dans les universités africaines afin d’éclairer les décisions stratégiques de ces dernières face à la demande croissante aux Universités de se transformer et d’augmenter leur visibilité dans les processus de développement d’innovations. La transformation souhaitée fait partie du vaste processus de changement organisationnel qui demande que les universités changent pour répondre aux défis croissants et complexes de développement à l’étendue nationale et régionale. L’étude tire des examens d’évaluation et des rapports spécifiques de 20 Universités de 10 pays et ceci sur une période de 10 ans. Les résultats révèlent que les universités ont pleinement mis en place des unités de planification avec des stratégies opérationnelles claires et des cadres de suivi. Cependant, dans la pratique, le suivi est l’activité appliquée à une grande base et utilisée pour évaluer la conformité des plans de travail et RUFORUM Working Document Series (ISSN 1607-9345) No. 14 (1): 299-304. Available from http://repository.ruforum.org Research Application Summary
  • 2. 300 Owuor, C. et al. plans budgétaires. Quelques universités ont des stratégies de suivi-évaluation et d’apprentissage bien définies et là où elles existent, leur mise en œuvre est limitée ou plus appropriée. Le document cherche à enrichir la littérature croissante sur la réflexion évaluative et la nécessité pour les entités à mettre en place des systèmes adaptatifs pour l’apprentissage et le perfectionnement. Mots clés: Universités africaines, processus d’évaluation, processus de changement organisationnel Introduction Organizations are continuously grappling with the use of findings from evaluative processes such as performance evaluation, development evaluations, formative evaluations, cost effective studies and cases studies to inform their internal systems, improve operational efficiency and strengthen organizational learning (Suarez-Balcazar and Taylor-Ritzer, 2014). A learning organization is conditioned to adapt and improve on its performance and contribute to innovation for evaluation (Eilertsen and London, 2005). This necessitates putting in place systems and structures that foster generation of lessons for learning and knowledge creation. The learning processes lead to creation of organization knowledge through “learning the how”, a vital component for organizational effectiveness (Edmondson and Moingeon, 1996). According to Schein (2004), organizations do the same things all the time until the “tried” and “true” way no longer holds. This is due to the learning anxiety that is produced by the need and shift to learn something new. The organizational “learning how” and “learning why” therefore, determines its capability to compete in open market space and to deal with changing market environment (Brown and Duguid, 1996; Grant, 1996). This learning emanate from multiple sources but more critically, through evaluative processes. Substantive organisational readiness for evaluation is favourably influenced through direct evaluation capacitybuilding,stimulatingcriticalthinking,supportreflectivepracticeandfacilitatinginternal evaluation culture to promote organisational learning also referred to as “evaluative thinking” (Cousins, 2004; McCoy et al., 2013; Befani et al., 2015). The concept of evaluative thinking continue to receive significant attention in both public and not-for-profit sectors as it seeks to contribute to new learning by providing evidence to chronicle, map and monitor progress, successes, failures and roadblocks in the innovation as it unfolds (Buckley and Archibald, 2013; Earl and Timperley, 2015). There are quite striking differences in organisational capacities to use evaluation in government agencies and voluntary organisations (Cousins et al., 2014). Though, the underlying impetus is the ever increasing need for organisations to build internal evaluative capacities to inform sustainable organisational and cultural change (McCoy et al., 2013). In this respect, it is now a standard project management practice to embed evaluative actions during programme design. Blending the programme design consideration with the concept of evaluative thinking, leads to better understanding of organisational operating context, setting clear organizational learning questions, improve the theory of change through questioning the underlying assumptions and intervention logic, and enriching stakeholder engagement in the entire implementation cycle (Bonbright, 2013). The other dimension that
  • 3. 301Fifth RUFORUM Biennial Regional Conference 17 - 21 October 2016, Cape Town, South Africa sets ground for evaluative processes in organizations, is the ever increasing incentives and rewards systems that is based on the growing climate of accountability that demand for evidence of results and transparent based implementation approaches. Major multilateral and institutional agencies have instituted robust accountability frameworks that use rigorous methods to prove impact and to the least, the effect of interventions and their investments. In 2012, the US Office of Management and Budget issued new standards that promote programmes demonstrating rigorous studies of effectiveness setting stringent impact measurement standards (USGOV, 2013). Similarly other agencies soon followed suit. In respect to these global trends in demands for increased accountability, the Regional Universities Forum for Capacity Building in Agriculture (RUFORUM) institutionalized a project management policy of conducting evaluative activities of each initiative and assess its contribution to the organisation mission and vision of “High performing African Universities that produce skilled, proactive graduates, demand driven research outputs and innovations in response to local, regional and national agricultural development priorities”. The purpose for this is threefold; (a) to provide opportunities of better understanding of the realities and therefore institute more realistic plans, (b) ensure timely identification of bottlenecks to inform adaptive actions, and (c) whether the initiatives have achieved their intended impact. Over the last 10 years, RUFORUM successfully implemented a number of projects that have left an indelible mark of impact and transformation of agricultural higher education landscape in Africa. Methodology In this paper, we attempt to draw lessons from RUFORUM engagement with universities over the last 10 years to strengthen the monitoring and evaluation functions. The analysis is not based on the programmatic achievements of each initiative, but rather on the cross cutting elements of the need to build a body of credible evidence for policy engagement and partnerships. The analysis is informed by the growing demand for development actors to institute internal evaluation adaptive systems, influence policy decisions and build multi-level collaborations. Current evaluative processes in Africa in Universities The cross-cutting thread in all RUFORUM initiatives is the capacity building element at both individual and institutional levels. At individual level, the results are clear cut, with performance measurement metrics taking into consideration parameters like the number of graduates trained, scholarships awarded, academic mobility and alumni placement. These metrics are standard regardless of the training institution. Therefore the individual capacity development outcomes are clearly understood and definition of success level well specified, and easy to ascertain the chain of cause and effect. However, at institutional level (University), the assessment and measurement of institutional capacity development is a complex affair. This comes at the backdrop of growing global trend in demands for increased accountability that have not spared the Africa universities either. The demand towards performance management, award of service contracts, ISO certification, and impact of university training
  • 4. 302 Owuor, C. et al. and research on technology and innovations development and the need for broad accountability is leading to growing demand for evaluative systems to support decision making. On the outlook most Africa universities have fully established Planning Units with clear operational strategies and monitoring frameworks. However in practice, monitoring is largely activity based and used to assess compliance with work plans and budgets and few universities have well defined monitoring, evaluation, and learning strategies and where they exist, implementation is limited or at most, ad hoc. Most universities have clear procedures for QualityAssurance (QA) and this is seen as a fundamental proposition that ground their core training and research functions. We observe that universities are setting higher benchmarks above the basic requirements of their respective National Councils of Higher Education regulations and other regional quality assurance framework. This is a right ambition for African universities as it has potential to cause their transformation into world class universities. We observe that universities are increasingly investing in tracer studies, a widely accepted approach for tracking impact of universities thought benchmark indices such as proportion of past students who find employment, sectors in which they work, their job performance, and employer satisfaction with them. The tracer study is a powerful strategic evaluative tool for universities to support curricula reviews, designing new market relevant courses and training students ready for the market. Whereas universities hold a pool of knowledge, evaluations of research projects are mostly externally driven and or restricted to projects undertaken in collaboration with governments, Non-Governmental Organizations (NGOs) or CGIAR centres. This implies that mostAfrica Universities have inadequately constituted internal evaluation systems to meet external accountability and satisfy their organisational learning requirements. Most universities have instituted information management systems to effectively improve students’ record management. This is seen as a bold step. However, we observe that there is little cataloguing of research findings and how they are being used. In general terms, there is limited capacity to implement monitoring and evaluative actions in most universities. In the next section, we provide an example of how RUFORUM uses evaluative processes to engage in policy dialogue and the associated challenges in doing so. RUFORUM made a strategic shift to engage at multiple levels in Policy advocacy as a transitional step to become a champion and voice for agricultural higher education inAfrica. However, from the monitoring and evaluation perspective, the challenges lies in estimating the attribution and contribution effects in face of actions by other actors of the resulting regional and continental policy shifts. The increased organizational engagement in advocacy calls for reflective and continuous assessment progress markers to elicit flexibility and responsiveness to capture incremental qualitative output for purposes of estimating impact. In this case, the need for numbers to qualify results is not a necessary requirement and most likely the quantitative indices turn out to be meaningless as the basis for assessing progress is premised on an arbitrary scale that is not universally understood. The remedy for this challenge is to have measured engagement with teams to reflect on emerging trends, harmonise processes and systems. This necessitates continuous acquisition of “political intelligence” and having adaptive monitoring and evaluation systems to capture the engagement processes. The presence of promoters and like-minded actors in positions of influence may
  • 5. 303Fifth RUFORUM Biennial Regional Conference 17 - 21 October 2016, Cape Town, South Africa considerably hasten the achievement of the end result and in the process, render the implementation strategies obsolete. Conclusion Evaluative thinking which is the on-going systematic practice of asking questions, questioning assumption, collecting and validating and analysing evidence and using evidence to inform decision is gaining credibility among development actors. This new paradigm will necessitate rethinking of University organisational learning and development systems and instituting adaptive frameworks to support their core roles of training, research and outreach. References Befani, B., Ramalingam, B. and Stern, E. 2015. Introduction – Towards systemic approaches to evaluation and impact. IDS Bulletin 46: 1–6. doi:10.1111/1759-5436.12116 Bonbright, D. 2012. Using impact evaluations results. Impact evaluation Notes. No.4 November 2012. Key Stone Accountability. Bradley, J., Swee, C., Goh, C., Lark, S. and Lee, L.E. 2004. Integrating evaluative inquiry into the organizational culture: a review and synthesis of the knowledge base. The Canadian Journal of Program Evaluation 19 (2): 99–141. ISSN 0834-1516. Canadian Evaluation Society. Brown, J.S. and Duguid, P. 1996. Organizational learning and communities-of-practice: Toward a unified view of working, learning and innovation. In: Cohen, M.D. and Sproull, L.S. (Eds), Organizational Learning. Thousand Oaks, CA: Sage. Buckley, J. andArchibald, T. 2013. EvaluativeThinking: Principles and Practices to Enhance Evaluation Capacity and Quality. Professional Development Workshop offered at the Annual Conference of the American Evaluation Association on October 20, 2013, Washington, D.C. Buckley, J., Archibald, T., Hargraves, M. and Trochim, W. M. 2015. Defining and teaching evaluative thinking. Insights from research on critical thinking. American Journal of Evaluation 36 (3): 375-388. doi: 10.1177/1098214015581706 Cousinsa, J.B., Gohb, S.C., Elliott, C., Aubryc, T. and Gilbert, N. 2015. Government and voluntary sector differences in organizational capacity to do and use evaluation. Evaluation and Program Planning 4: 1–13. Earl, L. and Timperley, H. 2015. Evaluative thinking for successful educational innovation. ECD Education Working. Papers. No. 122, OECD Publishing, Paris. http://dx.doi.org/ 10.1787/5jrxtk1jtdwf-en. Eilerstsen, S. and London, K.M.A. 2005. Modes of Organisational Learning. The Kollner Group, Inc. Grant, R. 1996. Prospering in dynamically competitive environments: Organizationalcapability as knowledge integration. Organization Science 7 (4): 375-387. McCoy, A., Rose, D. and Connolly, M. 2013. Developing evaluation cultures in human service organisations [online]. Evaluation Journal of Australasia 13 (1): 15-20. Availability: http://search.informit.com.au/documentSummary;dn=59920164573607; res=IELNZC> ISSN: 1035-719X. [cited 21 Sep 16].
  • 6. 304 Owuor, C. et al. Schein, E.H. 2004. A case of Organisational (Cultural?) Change. Organisational Culture and Leadership. Third Edition. Jossu-Bass a Willy Imprint. 365pp. Suarez-Balcazar Yolanda and Taylor-Ritzler, T. 2014. Moving from science to practice in evaluation capacity building. American Journal of Evaluation 35: 95 - 99. August 15, 2013, doi: 10.1177/1098214013499440 Teshome, A. and Tully, R.C. 2012. Higher education partnerships: Experiences and impact in Sub-Saharan Africa. RUFORUM Third Biennial Conference Proceedings, Entebbe, Uganda, 24-28 September 2012. pp.1163-1177. USAGov 2013. 78590 Federal Register Vol. 78, No. 248 / Thursday, December 26, 2013 / Rules and Regulations OFFICE OF MANAGEMENTAND BUDGET 2 CFR Chapter I, and Chapter II, Parts 200, 215, 220, 225, and 230 UniformAdministrative Requirements, Cost Principles, andAudit Requirements for FederalAwards Vol. 78, No. 248 / Thursday, December 26, 2013 / Rules and Regulations.