Reflections on Required Competencies For Health Systems Monitoring, Evaluation, and Learning In India

Authors

DOI:

https://doi.org/10.56645/jmde.v19i46.873

Abstract

The movement towards systems thinking and complexity-informed monitoring and evaluation has been ongoing for some time. Covid-19 has accelerated this shift and increased the salience of contextually aware, adaptive forms of evaluation. Drawing from our experience over three years of providing monitoring, evaluation, research and learning (MERL) to a large international philanthropic organization’s health systems portfolio of work in India, we identify three types of evaluator capacities and associated set of competencies that derive from complexity-driven practice in a developing country.

The first type of capacity relates to deriving evidence in contexts where there are very few traditional forms of data. The MERL portfolio related to health systems outcomes of access, financial protection, and equity even as the health programs themselves were evolving in a period of rapid transformation and dynamism of policy priorities and implementation mechanisms. This required an integration of traditional performance indicators drawn from government datasets with qualitative ‘signals’ focused on drivers of change within the system. Assessment of signals in turn required synthesizing different sources of information, including informal channels of obtaining evidence such as partner meetings or government-sponsored events. The evaluating team thus consciously empowered different kinds of researchers with differential expertise with the goal of building a much more pixelated picture. The goal was not to identify a single source of truth but rather a patchwork of validated information where the relevance of different pieces of data were dependent on evolving outcomes of interest.  

The second set of competencies related to the skills required to play a convening role for donors and implementing partners, supporting better understanding of the changing operating context and help inform decision-making by program officers and partners. This involved building and sustaining relationships across different stakeholders at different stages of the project – from proposal development to review. Competencies relating to effective dialogue and developing an understanding of the core interests of international, national and sub-national partners as well as international donors and experts could only be developed iteratively and over time, but this was crucial in a distributed health decision-making ecosystem like India.

The third and final set of competencies relate to operational adaptiveness, while evaluating an ecosystem with few constants. This can be the hardest competency to acquire because it is the farthest from the traditional notions embedded in the training of evaluation scientists and MERL practitioners. We found that the degree of buy-in and commitment to previously agreed upon frames of reference for evaluation can be shifted by changes in personnel or internal organizational structures. These shifts can lead to chain reactions of mismatched expectations that needed to be understood and managed in real time by MERL partners. The pandemic further created a natural experiment that on the one hand required a reexamination of program priorities and on the other depended on reliability of donor support.

Each of these three types of capacities – synthesizing nuanced evidence for adaptive action, relationship building and communication, and managing operational discontinuities are in fact inter-dependent. Building evaluator competencies isn’t simply about capacity-building but rather a recognition of the diversity of skills and worldviews that need to be encompassed within our monitoring and evaluation functions for today’s complex, discontinuous health systems.

Downloads

Download data is not yet available.

References

Barnes, M., Matka, E., & Sullivan, H. (2003). Evidence, Understanding and Complexity: Evaluation in Non-Linear Systems. Evaluation, 9(3), 265–284. https://doi.org/10.1177/13563890030093003 DOI: https://doi.org/10.1177/13563890030093003

Best practices and innovations—PMJAY Implementation. (2019). National Health Authority.

Brown, E. (2004). Effective training of program evaluators: A mixture of art and science. New Directions for Program Evaluation, 1980, 79–87. https://doi.org/10.1002/ev.1263 DOI: https://doi.org/10.1002/ev.1263

Davies, R., & MacKay, K. (2014). Evaluator Training: Content and Topic Valuation in University Evaluation Courses. American Journal of Evaluation, 35(3), 419–429. https://doi.org/10.1177/1098214013520066 DOI: https://doi.org/10.1177/1098214013520066

Dewey, J. D., Montrosse, B. E., Schröter, D. C., Sullins, C. D., & Mattox, J. R. (2008). Evaluator Competencies: What’s Taught Versus What’s Sought. American Journal of Evaluation, 29(3), 268–287. https://doi.org/10.1177/1098214008321152 DOI: https://doi.org/10.1177/1098214008321152

Diaz, J., Chaudhary, A. K., Jayaratne, K. S. U., & Assan, E. (2020). Expanding evaluator competency research: Exploring competencies for program evaluation using the context of non-formal education. Evaluation and Program Planning, 79. https://doi.org/10.1016/j.evalprogplan.2020.101790 DOI: https://doi.org/10.1016/j.evalprogplan.2020.101790

Fadlallah, R., El-Jardali, F., Hemadi, N., Morsi, R. Z., Abou Samra, C. A., Ahmad, A., Arif, K., Hishi, L., Honein-Abouhaidar, G., & Akl, E. A. (2018). Barriers and facilitators to implementation, uptake and sustainability of community-based health insurance schemes in low- and middle-income countries: A systematic review. International Journal for Equity in Health, 17(1). https://doi.org/10.1186/S12939-018-0721-4 DOI: https://doi.org/10.1186/s12939-018-0721-4

Gates, E. F. (2017). Learning from seasoned evaluators: Implications of systems approaches for evaluation practice. Evaluation, 23(2), 152–171. https://doi.org/10.1177/1356389017697613 DOI: https://doi.org/10.1177/1356389017697613

Khurana, N. (2021). Issue Analysis: A Use-Driven Approach to Data Governance Can Promote the Quality of Routine Health Data in India. In Global Health: Science and Practice (Vol. 9, Issue 2). www.ghspjournal.org DOI: https://doi.org/10.9745/GHSP-D-20-00347

King, J. A., & Ayoo, S. (2020). What do we know about evaluator education? A review of peer-reviewed publications (1978–2018). Evaluation and Program Planning, 79, 101785. https://doi.org/10.1016/j.evalprogplan.2020.101785 DOI: https://doi.org/10.1016/j.evalprogplan.2020.101785

LaVelle, J. M. (2020). Educating Evaluators 1976–2017: An Expanded Analysis of University-Based Evaluation Education Programs. American Journal of Evaluation, 41(4), 494–509. https://doi.org/10.1177/1098214019860914 DOI: https://doi.org/10.1177/1098214019860914

Pandey, A., Roy, N., Bhawsar, R., & Mishra, R. (2010). Health Information System in India: Issues of Data Availability and Quality 1. Demography India, 39, 111–128.

Patton, Michael Quinn. (2006). Evaluation for the Way We Work. Nonprofit Quarterly, 13(1), 28–33.

Pawson, R., & Tilley, N. (1997). An introduction to scientific realist evaluation. In Evaluation for the 21st century: A handbook (pp. 405–418). Sage Publications, Inc. https://doi.org/10.4135/9781483348896.n29 DOI: https://doi.org/10.4135/9781483348896.n29

Rao, N. V., Prashanth, N. S., & Hebbar, P. B. (2021). Beyond numbers, coverage and cost: Adaptive governance for post-COVID-19 reforms in India. BMJ Global Health, 6(2), e004392. https://doi.org/10.1136/bmjgh-2020-004392 DOI: https://doi.org/10.1136/bmjgh-2020-004392

Saxena, N., Singh, P., & Mishra, A. (2019). A qualitative comparative analysis of healthcare Supply-Demand side barriers under the publicly funded health insurance scheme in India. Indian Journal of Public Health, 63(4), 318–323. https://doi.org/10.4103/IJPH.IJPH_409_18 DOI: https://doi.org/10.4103/ijph.IJPH_409_18

Sridharan, S., & Nakaima, A. (2020). Valuing and embracing complexity: How an understanding of complex interventions needs to shape our evaluation capacities building initiatives. Evaluation and Program Planning, 80, 101440. https://doi.org/10.1016/J.EVALPROGPLAN.2017.05.001 DOI: https://doi.org/10.1016/j.evalprogplan.2017.05.001

Thakur, H. (2016). Study of Awareness, Enrollment, and Utilization of Rashtriya Swasthya Bima Yojana (National Health Insurance Scheme) in Maharashtra, India. Frontiers in Public Health, 3. https://doi.org/10.3389/FPUBH.2015.00282 DOI: https://doi.org/10.3389/fpubh.2015.00282

Vang, K. Y., Moore, M., & Nicklin, C. (2021). Applying a systems and complexity lens to building evaluation capacity: Learning from a multicountry donor’s experience. New Directions for Evaluation, 2021(170), 101–111. https://doi.org/10.1002/EV.20463 DOI: https://doi.org/10.1002/ev.20463

Witter, S., Palmer, N., Balabanova, D., Mounier-Jack, S., Martineau, T., Klicpera, A., Jensen, C., Garcia, M. P., & Gilson, L. (2019). Evidence review of what works for health systems strengthening, where and when? 99.

World Health Organization, European Observatory on Health Systems and Policies, Papanicolas, I., Rajan, D., Karanikolos, M., Soucat, A., & Figueras, J. (2022). Health system performance assessment: A framework for policy analysis. World Health Organization. https://apps.who.int/iris/handle/10665/352686

cover page

Downloads

Published

2023-12-11

How to Cite

Rao, N., & Nambiar, D. (2023). Reflections on Required Competencies For Health Systems Monitoring, Evaluation, and Learning In India. Journal of MultiDisciplinary Evaluation, 19(46), 65–75. https://doi.org/10.56645/jmde.v19i46.873

Issue

Section

International Evaluator Competence Perspectives