Zing Forum

Reading

Using LLMs to Analyze 'Sacred Cow Urine' Health Misinformation on Indian YouTube: How Cultural Confusion Deceives Algorithms and Humans

A research team from the University of Michigan developed a discourse analysis framework based on large language models (LLMs) specifically for identifying and analyzing the hybrid rhetorical strategies in health promotion content about 'gomutra' (cow urine) on Indian YouTube. This project reveals how traditional cultural metaphors and pseudoscientific discourse intertwine to form a complex discourse system that poses challenges to LLMs trained primarily on Western corpora.

健康谣言大语言模型话语分析文化混淆YouTube内容审核多语言处理计算社会科学
Published 2026-04-22 01:15Recent activity 2026-04-22 01:21Estimated read 7 min
Using LLMs to Analyze 'Sacred Cow Urine' Health Misinformation on Indian YouTube: How Cultural Confusion Deceives Algorithms and Humans
1

Section 01

[Introduction] Using LLMs to Analyze 'Sacred Cow Urine' Health Misinformation on Indian YouTube: How Cultural Confusion Challenges Algorithms and Humans

A research team from the University of Michigan developed a discourse analysis framework based on large language models (LLMs) to study the hybrid rhetorical strategies in health promotion content about 'gomutra' (cow urine) on Indian YouTube. This project reveals the complex system formed by the interweaving of traditional cultural metaphors and pseudoscientific discourse, which poses challenges to LLMs trained primarily on Western corpora and traditional content moderation mechanisms, providing a new perspective for understanding health misinformation rooted in cultural confusion.

2

Section 02

Research Background: Intertwining of Traditional Culture and Health Misinformation, and Moderation Challenges

In India, cow urine (gomutra) is regarded by some groups as a traditional substance with sacred healing properties. In recent years, when spread through platforms like YouTube, religious discourse and modern health science terminology are conflated, forming a phenomenon of 'cultural confusion'. Traditional moderation methods based on keywords or simple semantic analysis struggle to identify such content that appears as 'cultural expression' but actually spreads unsubstantiated health claims; moreover, the content is often a mix of English, Hindi, and Urdu, further increasing the complexity of automated analysis.

3

Section 03

Research Design: Multi-Stage Discourse Analysis Framework Assisted by LLMs

The study constructed a post-hoc analysis framework to evaluate the limitations of mainstream LLMs in handling culturally confused content. The steps include: 1. Sample collection: 30 multilingual videos (both promotional and debunking content); 2. Audio transcription: Using the OpenAI Whisper large model, with manual proofreading of 16% of samples, achieving an average word error rate of 7.04%; 3. Term extraction: GPT-4o identifies traditional cultural metaphors (religious symbols, traditional medicine concepts) and scientific terms (chemical components, etc.); 4. Intensity word analysis: Gemini, GPT-4o-mini, and DeepSeek extract emphasis words under zero/few-shot, formal/friendly tone conditions, and Cohen's Kappa coefficient is calculated to evaluate consistency.

4

Section 04

Key Findings: Systemic Limitations of Mainstream LLMs in Handling Culturally Confused Content

  1. Western-centric training corpora lead to biased understanding of Indian traditional medicine (e.g., Ayurveda) and religious metaphors, making it difficult to judge the misleading nature of juxtaposed traditional and scientific terms; 2. Multilingual mixing (code-switching) increases analysis difficulty; 3. Models underestimate the correlation between emotional intensity and factual accuracy: promotional content uses strong words like 'miraculous', while debunking content is restrained in expression, but cultural expressions themselves carry emotions that easily lead to misjudgment.
5

Section 05

Methodological Innovations and Ethical Considerations: Balancing Transparency and Responsibility

Methodological innovations: Publicly releasing evaluation scripts (WER calculator, F1 evaluator, Kappa analyzer) and prompt templates for GPT-4o, Gemini 2.5 Pro, and DeepSeek to enhance replicability. Ethical considerations: Excluding personal information of viewers/commenters, limiting the dataset to non-commercial academic use, requiring email applications for controlled access, balancing research value and privacy protection.

6

Section 06

Practical Implications: Providing New Directions for Content Moderation and Fact-Checking

For platforms: It is recommended to introduce a multi-dimensional analysis framework that not only detects factual accuracy but also analyzes the rhetorical strategies of mixed traditional and scientific discourse, enhancing multilingual and cross-cultural understanding capabilities. For fact-checking organizations: LLMs can assist in identifying suspicious rhetorical patterns, but the final judgment requires the cultural sensitivity of human experts.

7

Section 07

Limitations, Future Directions, and Conclusion: Balancing Technical Neutrality and Humanistic Care

Limitations: Small sample size (N=30), single topic, post-hoc analysis not tracking propagation dynamics. Future directions: Expand topic coverage, introduce user behavior data, explore multimodal analysis (combining video visuals/audio intonation). Conclusion: Technical solutions need to be combined with humanistic care; avoid cultural stigmatization when combating misinformation, and build an effective and responsible information ecosystem.