Just add rigour: Three do’s and don’ts when using AI for text analysis.
A lot of evaluation work is a kind of text analysis: processing reports, interview transcripts, etc. A bit like qualitative social science research. So this little piece is for evaluators in particular and (qualitative) social scientists in general.
How do we get from texts to evaluative judgements?
Recently many evaluators and researchers have been turning to AI to help.
BUT if you didn’t have a clear workflow from data to judgements before AI, don’t lean on the black box of the AI to cover that up. Here is my first set of Do’s and Don’ts. More soon.
1) DO Break up big, vague tasks into multiple smaller, clearer steps
Do
Don't
DO Break up complex, vague tasks into smaller steps which can be intersubjectively verified.
DON’T Ask AI to make broad evaluative judgments (like "Is this good?")
DODocument your methodology so that you can explain step by step how you reached your conclusions in a way which anyone can check. No black boxes. Use the AI to speed up many simple tasks which you could have done yourself if you had the time.
DON’T Trust the AI's explanations of how it reached its conclusions.
AIs often create plausible-sounding but unreliable explanations after the fact. Normal AIs have very limited information about their inner processes.
DO Break up the data into pieces for AI analysis. Ideally run each piece as a separate prompt. Failing that, number each section and ask for a numbered, section-by-section answer, for example in a table.
DON’T Give an AI large pieces of text and expect it will pay due attention to all of it. It will claim to have done, and may provide references to relevant passages, but attention is expensive and it is always trying to reduce that expense. If you let it, it will always try to skim read and jump to conclusions.
DOUse explicit, manual methods (Excel?!) to synthesise the results of the multiple separate tasks you gave the AI.
DON’T Ask an AI to do maths for you, like adding up the number of positive or negative findings on a rubric. AIs are still terrible at maths.
Even worse, DON’T ask an AI to do implicit counting and comparison like “are there more positive or negative mentions of X in this report?”
“Intersubjectively verifiable” just means that most people will more or less agree on the answer most of the time.
It creates transparency and allows others to verify your work.
Clear instructions lead to more reliable results.
If you can’t check it, you can’t trust it.
Example of an intersubjectively verifiable task:
✅ Does this paragraph mention water and sanitation?
✅ If so, are any recent changes mentioned?
✅ If so, do these sound like positive changes according to the interviewee?
Notice that here we’ve broken down a larger task into three smaller and simpler steps.
Examples of tasks which are not intersubjectively verifiable:
❌ Is the intervention described in this report efficient and effective?
Text needs breaking up into sections, judgements on efficiency and effectiveness need breaking down into pieces, e.g. using rubrics.
❌ What are the main themes in this document?
This is a very common question in qualitative research, but it’s a terrible task to give to an AI without further details. What do we mean by a theme? Are we interested in economic aspects? Interpersonal aspects? How are the themes to be identified and refined? Here, a whole world of qualitative social science experience, skills and workflows (grounded theory, thematic analysis) have been bypassed in a single sentence.
❌ Summarise this document!
Yes, everyone does it. Evaluators do it. Schoolchildren do it. Pets will be doing it soon. As a quick time-saver for low-stakes tasks, it’s very useful. But it’s the vaguest, highest-level instruction, not a systematic analysis.
How do you break down a high-level judgement into a workflow of smaller tasks? Well isn’t that what evaluation methods and qualitative research methods are for? Go read a book!
We’re not saying you have to specify in advance exactly what methods you will use. That’s a bit too positivistic. But you should at least document them as you go along and be prepared to defend them when your analysis is done. That’s the untranslatable Nachvollziehbarkeit.
In terms of how to implement your workflow technically, see this great contribution from Christopher Robert. At Causal Map, we’re also working on ways to make workflows accessible. See how we currently use AI in Causal Map here.
This post is based on my recent contribution to the NLP-CoP Ethics & Governance Working Group, along with colleagues Niamh Barry, Elizabeth Long and Grace Lyn Higdon. In the next couple of weeks we’ll look at two more do’s and don’ts.
This post was originally published by Steve Powell on LinkedIn and has been republished here. See the original article here
Over the last three years we have developed a whole range of algorithms for causal mapping. The algorithms are published open-source in the form of functions for the programming language R.
I've got a load of texts to analyse, can I use Causal Map just for standard Qualitative Data Analysis, like ordinary thematic analysis, to identify important themes?