Zing Forum

Reading

How LLMs Understand Rhetorical Questions: A Multi-Dimensional Representation Mechanism Revealed by Linear Probing

Research using linear probing technology found that LLMs' representations of rhetorical questions exhibit early emergence characteristics; rhetorical signals can be encoded through multiple linear directions, and probes trained on different datasets capture different rhetorical phenomena.

LLM表征反问句线性探针可解释性修辞分析自然语言理解神经网络
Published 2026-04-16 01:50Recent activity 2026-04-16 11:50Estimated read 7 min
How LLMs Understand Rhetorical Questions: A Multi-Dimensional Representation Mechanism Revealed by Linear Probing
1

Section 01

Introduction: Core of the Study on Multi-Dimensional Representation Mechanism of Rhetorical Questions in LLMs

This study uses linear probing technology to explore the internal representation mechanism of rhetorical questions in LLMs. Key findings include: Rhetorical signals emerge in the early layers of the model, and the representation of the last token is the most stable; rhetorical questions are encoded along multiple linear directions in the representation space, and probes trained on different datasets capture different rhetorical phenomena; cross-dataset transfer is detectable but has differences, revealing LLMs' multi-dimensional understanding of rhetorical questions.

2

Section 02

Background: Complexity of Rhetorical Questions and Challenges in Automatic Recognition

Rhetorical questions are a special linguistic phenomenon whose core function is rhetorical expression rather than information acquisition (e.g., "Shouldn't we protect the environment?" emphasizes an opinion). The tension between their semantics and pragmatics makes automatic recognition complex, requiring reliance on context, tone, and intent rather than just syntactic structure. For LLMs to understand these subtle differences, they need to form internal representations that distinguish rhetorical intent.

3

Section 03

Research Methods: Linear Probing Technology and Dataset Selection

Linear probing technology is used to analyze the internal representations of LLMs: freeze the pre-trained model parameters, train a linear classifier on the hidden layer outputs. If it can distinguish rhetorical questions from ordinary questions, it indicates that the relevant features have been learned by the model. The study was conducted on two different social media datasets to test the generality of the findings.

4

Section 04

Key Findings: Early Emergence and Last Token Representation Characteristics

Rhetorical signals start to emerge in the early layers of the model, indicating that LLMs can recognize rhetorical features of rhetorical questions early when processing sentences; the rhetorical signal is most stable in the last token's representation, which is consistent with LLMs often using the last token for downstream prediction; rhetorical questions are linearly separable within a single dataset, and cross-dataset transfer AUROC reaches 0.7-0.8, indicating the existence of general rhetorical question-related representations.

5

Section 05

Multi-Dimensional Representation Findings: Non-Single Direction Encoding Mechanism

Cross-dataset transfer is feasible, but when probes from different datasets are applied to the same corpus, the ranking results differ significantly (overlap of top-ranked instances is less than 0.2), suggesting that rhetorical questions are encoded along multiple linear directions in the representation space, with each direction emphasizing different clues. Qualitative analysis shows: some probes capture rhetorical stance at the discourse level, while others emphasize locally syntactically driven questioning behavior.

6

Section 06

Diversity of Rhetorical Phenomena: Different Types of Rhetorical Questions and Representation Modes

Rhetorical questions include multiple rhetorical strategies: emphasis type (e.g., "Who doesn't want to succeed?"), questioning type (e.g., "Do you really believe this statement?"), and sarcastic type (e.g., "Isn't this great?" in a negative context). Different types of rhetorical questions activate different internal representation modes in LLMs, explaining why a single probe cannot capture all rhetorical phenomena.

7

Section 07

Implications for Interpretability: Reflections on LLM Concept Probing

Implications of the study for LLM interpretability: 1. A seemingly single concept (such as rhetorical questions) may be decomposed into multiple dimensions, and concept probing needs to consider the internal structure; 2. Early layers capture rhetorical signals, which is consistent with the characteristic of LLMs processing language information layer by layer; 3. The feasibility and differences of cross-dataset transfer indicate that LLMs have general rhetorical perception capabilities, but their manifestations vary depending on training data.

8

Section 08

Future Research Directions: Expansion from Mechanism to Application

Future research directions: 1. Develop fine-grained probing methods to capture multiple linear directions simultaneously to fully understand the representation structure of rhetorical questions; 2. Explore the relationship between the representations of rhetorical questions and other rhetorical phenomena (metaphors, irony) to see if a unified rhetorical framework can be formed; 3. Apply the findings to NLP tasks such as sentiment analysis and stance detection to improve performance.