Zing Forum

Reading

Art as an Algorithmic Virus: Unifying Generative Collapse and AI Value Alignment via Cognitive Affordances

Researchers propose art as a cognitive framework for understanding AI value alignment issues, using the concept of "generative collapse" to reveal value convergence phenomena in large model training, providing a new perspective for AI safety research.

AI价值对齐生成式崩溃认知可供性大语言模型RLHFAI安全艺术理论人机交互
Published 2026-04-27 19:21Recent activity 2026-04-27 19:24Estimated read 5 min
Art as an Algorithmic Virus: Unifying Generative Collapse and AI Value Alignment via Cognitive Affordances
1

Section 01

Art as an Algorithmic Virus: A New Perspective Linking AI Value Alignment and Generative Collapse

This article proposes the metaphor of art as an algorithmic virus, unifying the two phenomena of generative collapse and AI value convergence through the concept of cognitive affordances, providing an interdisciplinary new path for AI safety research. The core view is that art can serve as a key framework for understanding AI value alignment issues, revealing the mechanisms of value conflict and convergence in large model training.

2

Section 02

Core Challenges and Related Phenomena of AI Value Alignment

AI value alignment is a core issue in current AI safety research, which requires ensuring that model behaviors align with human values. Generative collapse refers to value-level conflicts caused by the tension between the model's immediate requests and safety constraints (e.g., fluctuating outputs when rejecting harmful requests); AI value convergence means that diverse values tend to converge to specific patterns during training, but methods like RLHF have limitations such as uncertain intent inference, scale constraints, and execution failures.

3

Section 03

Cognitive Affordances: From Human-Computer Interaction to AI Value Learning

Cognitive affordances borrow concepts from the field of human-computer interaction and are extended to the possibilities that artworks provide for human cognition, such as interpretation and emotional responses. Large models not only learn language rules during training but also grasp the cultural context, social norms, and values behind language. As a condensed expression of human culture, art provides rich value signals for models.

4

Section 04

Multi-layered Meanings of the Algorithmic Virus Metaphor

The metaphor of art as an algorithmic virus includes three layers: 1. Contagiousness—art and AI content can spread quickly and resonate; 2. Host: human mind (cognitive architecture, emotions, cultural background); 3. Selection pressure: the training mechanism based on human feedback is similar to the selection process in cultural evolution.

5

Section 05

Theoretical Framework Unifying Generative Collapse and Value Convergence

Generative collapse is the friction when there is insufficient cognitive affordance (the model cannot coordinate conflicting goals); value convergence is the result of capturing cognitive affordance (the model identifies and uses value patterns that meet multiple goals). Art provides an ideal training signal for this kind of learning, and the paper formalizes the framework through six hypotheses, providing a roadmap for empirical research.

6

Section 06

Key Implications for AI Safety Practice

Implications of the research for AI safety: 1. When designing value alignment solutions, focus on cognitive affordance design rather than just listing rules; 2. Use artistic narratives as a medium for value education (similar to how humans learn morality through stories); 3. Treat generative collapse as a signal of the model's internal value conflicts, not just a bug to be eliminated.

7

Section 07

Theoretical Limitations and Future Research Directions

Theoretical limitations include insufficient empirical basis and unproven cross-cultural applicability; future directions: developing cognitive affordance measurement methods, exploring the impact of different art types on model values, researching human-AI collaborative value negotiation, and extending to multimodal AI systems.