As AI transforms social science, development practitioners must navigate the tension between scaling qualitative data and falling into a ‘hivemind’ of Western-centric consensus. In this blog Arul Chib collaborates with AI to discuss how in shifting from data collectors to synthesis architects, researchers can utilize local models to protect privacy while ensuring the ‘soul’ of the narrative remains human-led.

History suggests that the panic surrounding new technologies often follows a predictable pattern of resistance, moving from initial bans and surveillance toward eventual integration. In the 1970s, the introduction of the calculator sparked fears that it would lead to mental atrophy and destroy the fundamental understanding of mathematics.
However, this shift did not kill maths; rather, it ended the era of the human arithmetic clerk, allowing students to focus on higher-order problem-solving rather than the mechanics of long division. Similarly, the rise of Wikipedia as an online resource in the 2000s was initially seen as a threat to academic rigor, but the Internet eventually became a baseline for information literacy where the skill shifted from finding facts to verifying their sources. Today, the emergence of AI in research represents a new phase where the value of academic work is shifting from merely examining the final output (e.g., essay) to understanding the documentation of the thought process through mandatory disclosure of GenAI prompts.
In the field of social science, traditional qualitative research has long been hampered by being slow, small-scale and expensive. This creates a persistent bottleneck where researchers are often limited to a handful of interviews due to the labour-intensive nature of manual transcription and coding. AI offers the promise of ‘deep data at scale,’ enabling practitioners to move from ten interviews to analysing hundreds or even thousands of open-ended responses with increased efficiency. By offloading the clerical work of tagging sentences, the researcher can reclaim time to focus on policy implications and the logic of the synthesis. Within this framework, AI serves as an augmented research assistant rather than a lead investigator, effectively ending the era of the human transcription clerk.
Despite these efficiency gains, AI models present significant ethical risks for development practitioners because they are predominantly trained on Western, Educated, Industrialized, Rich and Democratic (WEIRD) data. This inherent bias can skew the interpretation of indigenous knowledge and non-Western development contexts. Furthermore, uploading sensitive field data to cloud-based Large Language Models (LLMs) represents a major ethical breach and a potential violation of privacy regulations like GDPR. To protect data sovereignty, researchers are increasingly turning to local LLMs that run directly on their own hardware, ensuring that the narratives of marginalized communities never hit the corporate cloud.
A core tension exists between the deterministic nature of AI and the expansive reality of qualitative research. While larger models are often expected to be more creative and inclusive, they frequently converge toward a single deterministic consensus, or a global average, due to alignment training. This phenomenon of inter-model homogeneity results in ‘hivemind’ outputs that strip away the originality and understanding (e.g., cultural silences) essential for breakthrough work. Because AI tends to average out outliers, the most critical data points – such as the lone dissenting voice in a village who disagrees with the chief – are often erased or ignored. If the cost of verifying an AI’s summary to ensure it has not hallucinated a consensus is higher than the cost of reading the raw transcripts, then AI becomes a net-negative for research productivity.
To avoid these pitfalls, researchers should utilize AI for deductive tasks, such as applying established frameworks like the Sustainable Development Goals (SDGs), while reserving inductive discovery for human intuition. Specialized tools like ATLAS.ti allow for thematic interrogation where the researcher maintains a direct link to the raw text, preventing the ‘black box’ problem found in general chatbots. AI can also serve as an ‘unbiased mirror’, identifying themes that contradict a researcher’s own subconscious biases or finding hidden gaps in existing literature through semantic mapping. Ultimately, the value of the development practitioner using qualitative techniques has shifted from a data collector to a synthesis architect. While AI can automate the clerical tagging of data, the final interpretive leap must remain human to ensure cultural nuance and reflexive awareness. By embracing a ‘human-in-the-loop’ mandate, researchers move beyond being simple data sorters to become data critics who provide the creativity that a hivemind cannot replicate.
About the Synthesis Architect:

Arul Chib specializes in the intersection of technology and development, focusing on how digital tools can be ethically integrated into qualitative social science research. This piece was generated by Google Gemini based on his talk recorded at The Centre for Social and Economic Progress, Delhi on 30 Mar 2026. The output was edited for clarity while retaining the authentic voice (e.g., inverted commas for emphasis) of the LLM.
All opinions expressed in this blog are the author’s own, and are not necessarily representative of BLISS, the International Institute of Social Studies, or Erasmus University Rotterdam. Please use generative AI tools with care.
Discover more from Bliss
Subscribe to get the latest posts sent to your email.
What do you think?