← All insights

Advising a critical thinking RCT in government schools

India Institute is evaluating a critical thinking curriculum across Tamil Nadu government schools through a randomised controlled trial commissioned by the state government. We came in as research advisors — not the principal investigators on the trial itself, but working with the team on evaluation design, instrument review, and the measurement choices that will determine what the RCT can and cannot conclude.

The question we kept returning to was the one the team was somewhat reluctant to ask: what exactly are we measuring when we measure critical thinking?

This is not a philosophical objection to the study. It is a prior methodological question that the study cannot avoid. RCTs produce precise estimates of the effect of an intervention on a measured outcome. If the measured outcome does not map cleanly to the construct the intervention is trying to shift, the trial can be rigorously conducted and produce a finding that is technically correct and practically misleading.

Critical thinking is a construct with a relatively settled philosophical definition — it involves the capacity for structured analysis of claims, recognition of logical fallacies, evaluation of evidence, and calibrated epistemic humility about one's own reasoning. What it does not have is a universally accepted measurement instrument, particularly for children in non-Anglophone, state-school contexts.

The instruments that exist carry assumptions. The Cornell Critical Thinking Test and the Collegiate Learning Assessment were developed for US undergraduate populations. The PISA framework for reasoning comes closer to a cross-cultural instrument, but its validation in Tamil-medium government school settings is limited. Locally developed instruments solve the cultural fit problem but introduce reliability and comparability problems. There is no clean option.

The advice we gave the team centred on triangulation. A single-instrument approach to measuring critical thinking will produce a headline number whose interpretation depends entirely on whether the reader trusts the instrument. A multi-instrument approach — test scores alongside structured classroom observation protocols and teacher practice measures — produces a picture that is harder to summarise in a press release but much harder to dismiss.

The classroom observation piece matters more than it might initially appear. Critical thinking curricula change not just what students know but how classrooms operate: the ratio of teacher talk to student talk, whether students are invited to challenge assertions, whether error is treated as information or failure. These practice changes are often the proximate mechanism through which student outcomes shift. Measuring only the student outcome without measuring the practice change makes it impossible to distinguish a well-implemented curriculum from a poorly-implemented one that happens to have produced a positive test score through unrelated factors.

Tamil Nadu is a useful site for this kind of study. The state has a relatively high-capacity government school system by Indian standards — UDISE data consistently places Tamil Nadu among the top states on infrastructure, teacher availability, and enrolment ratios. The ASER 2023 report shows Tamil Nadu's foundational literacy and numeracy outcomes are strong relative to national averages, which means the study is working above the floor. The gains from a critical thinking curriculum are likely to be more visible in a system where basic literacy is not the binding constraint.

The RCT design assigns schools, not students, to treatment and control — a cluster-randomised design that is appropriate given the curriculum's classroom-level implementation. The power calculations needed to account for intra-cluster correlation in Tamil Nadu government schools; we advised the team to use intraclass correlation estimates from ASER and NAS data rather than generic assumptions.

Whether the trial will detect effects depends significantly on implementation fidelity. A curriculum that teachers do not understand well enough to teach is a different treatment than a curriculum that is delivered as designed. Teacher training quality and ongoing coaching support are, in a meaningful sense, part of the intervention — and they need to be documented, not assumed.

UDISE+ state report cards provide school-level baseline data. Tamil Nadu EMIS carries more granular district data on enrolment, attendance, and teacher deployment.