bouncer
← All research notes

2026-03-09

The Science of Influence: Academic Foundations of Automated Detection

This paper maps Bouncer's influence detection framework to its academic foundations. Every dimension, technique, and design decision is grounded in published research from social psychology, rhetoric, communication studies, and cognitive science. Technique names link to our versioned glossary, where each entry includes a plain-language definition and source citation.

1. Introduction: Why Automated Detection?

The volume of persuasive content on platforms like YouTube exceeds any individual's capacity for critical evaluation. Humans produce roughly 500 hours of video per minute on YouTube alone. Manual fact-checking and media literacy education, while valuable, operate on fundamentally different timescales than content production.

Research on computational propaganda detection has accelerated since 2016, with work by Woolley and Howard documenting the industrialization of online influence operations (Woolley & Howard, 2016, "Political Communication, Computational Propaganda, and Autonomous Agents," International Journal of Communication, 10, 4882–4890). Da San Martino et al. developed fine-grained propaganda technique detection in news articles, identifying 18 distinct techniques with neural sequence-labeling models (Da San Martino et al., 2019, "Fine-Grained Analysis of Propaganda in News Articles," EMNLP). The SemEval-2020 Task 11 extended this to a shared evaluation framework (Da San Martino et al., 2020, SemEval-2020).

Bouncer builds on this lineage but shifts the frame from propaganda detection (binary: is this propaganda?) to influence transparency (continuous: what persuasion mechanisms are operating, and how visible are they?). This reframing is grounded in inoculation theory's finding that awareness of techniques, not avoidance of content, is the effective intervention (McGuire, 1961; van der Linden et al., 2022).

2. The Six Dimensions

Bouncer scores content on six continuous dimensions (0–1), each representing a family of empirically validated persuasion mechanisms. The dimensional model avoids the false precision of binary classification while preserving actionable specificity.

2.1. Emotional Appeal

The theoretical foundation for this dimension spans Aristotle's concept of pathos through modern dual-process theory. Petty and Cacioppo's Elaboration Likelihood Model (ELM) established that emotional appeals operate via the "peripheral route" of persuasion, bypassing systematic argument evaluation (Petty & Cacioppo, 1986, Communication and Persuasion, Springer-Verlag).

Kahneman's framework of System 1 (fast, automatic, emotional) versus System 2 (slow, deliberate, logical) processing provides the cognitive mechanism: emotional engineering activates System 1 responses that feel like conclusions rather than reactions (Kahneman, 2011, Thinking, Fast and Slow, Farrar, Straus and Giroux).

Specific techniques within this dimension draw on distinct research traditions: Pathos — classical rhetoric's emotional appeal; Moral outrage — Haidt's Moral Foundations Theory identifies outrage as a uniquely viral moral emotion (Haidt, 2012, The Righteous Mind, Vintage), with Brady et al. finding that moral-emotional language increases message diffusion by 20% per word (Brady et al., 2017, "Emotion shapes the diffusion of moralized content in social networks," PNAS, 114(28), 7313–7318); Fear appeal — Witte's Extended Parallel Process Model (EPPM) describes how threat perception interacts with efficacy beliefs to drive either protective action or defensive avoidance (Witte, 1992, "Putting the fear back into fear appeals," Communication Monographs, 59(4), 329–349); Empathy elicitation — Batson's empathy-altruism hypothesis explains why individual narratives override statistical reasoning (Batson, 1991, The Altruism Question, Lawrence Erlbaum), a phenomenon Schelling identified as the "identifiable victim effect" (Schelling, 1968, "The Life You Save May Be Your Own," in Chase, ed., Problems in Public Expenditure Analysis).

See all Emotional Appeal techniques in the glossary · emotional appeal data analysis

2.2. Story Shaping

Entman's framing theory provides the theoretical core: "To frame is to select some aspects of a perceived reality and make them more salient in a communicating text, in such a way as to promote a particular problem definition, causal interpretation, moral evaluation, and/or treatment recommendation" (Entman, 1993, "Framing: Toward clarification of a fractured paradigm," Journal of Communication, 43(4), 51–58).

Goffman's Frame Analysis established that frames organize experience and guide action (Goffman, 1974, Frame Analysis, Harvard University Press). Lakoff extended this to political communication, demonstrating how metaphorical frames shape policy reasoning (Lakoff, 2004, Don't Think of an Elephant!, Chelsea Green). Chong and Druckman's review established that framing effects are among the most robust findings in political communication (Chong & Druckman, 2007, "Framing Theory," Annual Review of Political Science, 10, 103–126).

Key techniques: Loaded language — Hayakawa's semantic analysis showed how connotation-laden words bypass rational evaluation (Hayakawa, 1949, Language in Thought and Action); False equivalence — Boykoff and Boykoff documented how journalistic "balance" norms create false equivalence on settled scientific questions (Boykoff & Boykoff, 2004, "Balance as bias," Global Environmental Change, 14, 125–136); Character simplification — Ross's fundamental attribution error research explains why audiences accept flattened characterizations (Ross, 1977, "The intuitive psychologist and his shortcomings," in Berkowitz, ed., Advances in Experimental Social Psychology, vol. 10).

See all Story Shaping techniques in the glossary · story shaping data analysis

2.3. Implicit Claims

Linguistic pragmatics provides the theoretical foundation. Grice's cooperative principle establishes that communication conveys more than its literal content — listeners routinely infer unstated propositions (Grice, 1975, "Logic and Conversation," in Cole & Morgan, eds., Syntax and Semantics 3: Speech Acts). Presupposition theory in formal semantics describes how certain linguistic structures smuggle in assumptions that listeners accept without evaluation (Stalnaker, 1974, "Pragmatic Presuppositions," in Munitz & Unger, eds., Semantics and Philosophy).

Cialdini's concept of "pre-suasion" describes how strategic framing before a message shapes what the audience is prepared to accept (Cialdini, 2016, Pre-Suasion, Simon & Schuster). Implicit claims are the mechanism through which pre-suasion operates in video content: by the time an explicit argument is made, the implicit premises have already been accepted.

Related techniques: Confirmation bias — Nickerson's comprehensive review established that confirmatory information processing is among the most robust cognitive biases (Nickerson, 1998, "Confirmation Bias: A Ubiquitous Phenomenon in Many Guises," Review of General Psychology, 2(2), 175–220); Strategic ambiguity — Eisenberg demonstrated how deliberate vagueness serves communicative goals by allowing multiple audiences to infer compatible but different meanings (Eisenberg, 1984, "Ambiguity as strategy in organizational communication," Communication Monographs, 51, 227–242).

See all Implicit Claims techniques in the glossary · implicit claims data analysis

2.4. Group Characterization

Gerbner's cultivation theory provides the media-effects foundation: heavy exposure to stereotyped portrayals shapes audiences' perceptions of social reality (Gerbner et al., 1986, "Living with Television: The Dynamics of the Cultivation Process," in Bryant & Zillmann, eds., Perspectives on Media Effects). Bandura's Social Cognitive Theory extends this to observational learning, establishing that mediated portrayals serve as models for social behavior (Bandura, 2001, "Social Cognitive Theory of Mass Communication," Media Psychology, 3(3), 265–299).

Allport's foundational work on prejudice identified categorization as a necessary cognitive function that becomes distorted through selective reinforcement (Allport, 1954, The Nature of Prejudice, Addison-Wesley). Devine's dissociation model showed that stereotype knowledge is automatic but endorsement is controllable — making awareness the critical intervention point (Devine, 1989, "Stereotypes and prejudice: Their automatic and controlled components," Journal of Personality and Social Psychology, 56(1), 5–18).

Relevant techniques: Us vs. Them — Tajfel's Social Identity Theory demonstrated that mere categorization into groups produces in-group favoritism and out-group discrimination, even with arbitrary group assignments (Tajfel & Turner, 1979, "An Integrative Theory of Intergroup Conflict," in Austin & Worchel, eds., The Social Psychology of Intergroup Relations); Character simplification — reducing complex individuals to narrative archetypes, as described in Propp's morphological analysis of story structure (Propp, 1928/1968, Morphology of the Folktale, University of Texas Press).

See all Group Characterization techniques in the glossary · group characterization data analysis

2.5. Engagement Mechanics

This dimension addresses the intersection of persuasion psychology and platform design. Zuboff's concept of "surveillance capitalism" describes the economic logic driving attention-maximizing content design (Zuboff, 2019, The Age of Surveillance Capitalism, PublicAffairs). Harris's work on "persuasive design" identifies specific patterns that exploit cognitive vulnerabilities for engagement (Harris, 2016, "How Technology Hijacks People's Minds," Center for Humane Technology).

Mathur et al.'s systematic study of dark patterns identified taxonomies of manipulative interface design across 11,000 shopping websites (Mathur et al., 2019, "Dark Patterns at Scale," Proceedings of the ACM on Human-Computer Interaction, 3(CSCW), Article 81). While focused on e-commerce, the underlying mechanisms — false urgency, manufactured scarcity, social proof manipulation — transfer directly to video content design.

Key techniques: Clickbait — Loewenstein's information gap theory explains the near-physical curiosity itch that curiosity-gap headlines exploit (Loewenstein, 1994, "The psychology of curiosity," Psychological Bulletin, 116(1), 75–98); Sensationalism — the availability heuristic causes people to overweight vivid, dramatic information in probability judgments (Tversky & Kahneman, 1973, "Availability: A heuristic for judging frequency and probability," Cognitive Psychology, 5(2), 207–232).

See all Engagement Mechanics techniques in the glossary · engagement mechanics data analysis

2.6. Call to Action

Austin's speech act theory provides the linguistic foundation: utterances do not merely describe — they do things (request, promise, command) (Austin, 1962, How to Do Things with Words, Oxford University Press). Searle extended this taxonomy, distinguishing between direct and indirect speech acts (Searle, 1969, Speech Acts, Cambridge University Press).

The compliance literature identifies the mechanisms through which calls to action achieve their effects. Cialdini and Goldstein's review identified six principles of social influence that operate in persuasive requests (Cialdini & Goldstein, 2004, "Social Influence: Compliance and Conformity," Annual Review of Psychology, 55, 591–621). Freedman and Fraser's foot-in-the-door research demonstrated that small initial commitments dramatically increase compliance with larger subsequent requests (Freedman & Fraser, 1966, "Compliance without pressure," Journal of Personality and Social Psychology, 4(2), 195–202).

Relevant techniques: Direct appeal — explicit requests (subscribe, donate, share); Social proof — Asch's conformity experiments demonstrated the power of perceived consensus (Asch, 1951, "Effects of group pressure upon the modification and distortion of judgments," in Guetzkow, ed., Groups, Leadership and Men); Emotional blackmail — Forward's FOG (Fear, Obligation, Guilt) model describes how emotional pressure substitutes for rational persuasion (Forward, 1997, Emotional Blackmail, HarperCollins).

See all Call to Action techniques in the glossary · call to action data analysis

3. The Transparency Score

Bouncer's transparency score (0–1) measures the gap between a video's apparent purpose and its actual persuasive function. This operationalizes a concept from multiple research traditions:

Goffman's dramaturgical analysis distinguishes between "front stage" performances (what the audience sees) and "backstage" activity (the machinery of presentation) (Goffman, 1959, The Presentation of Self in Everyday Life, Doubleday). In video content, the transparency score estimates how much of the persuasive machinery remains backstage versus how much the creator openly acknowledges.

This connects to Habermas's theory of communicative action, which distinguishes between "communicative rationality" (oriented toward mutual understanding) and "strategic rationality" (oriented toward achieving a predetermined outcome) (Habermas, 1984, The Theory of Communicative Action, Vol. 1, Beacon Press). Covert influence — low transparency — represents strategic action disguised as communicative action.

The practical significance of transparency, independent of influence intensity, has been validated by Bouncer's own data: our narrative framing analysis found that as narrative framing intensity increases, transparency drops by nearly half. A transparently partisan commentator (high intensity, high transparency) represents a fundamentally different phenomenon than a covert influence campaign (high intensity, low transparency), even at identical intensity scores.

4. Design Decisions Grounded in Research

Several of Bouncer's design choices — not just what we detect, but how we present it — are informed by specific psychological research:

Self-referential framing

Bouncer's educational content uses "you" language ("Ask yourself: whose perspective is missing?") rather than third-person framing ("viewers should consider..."). This counters the third-person effect — Davison's finding that people consistently believe media influence affects others more than themselves (Davison, 1983, "The third-person effect in communication," Public Opinion Quarterly, 47(1), 1–15). By addressing the user directly, we reduce the cognitive distance that enables "I'm too smart to be influenced" dismissals.

"Coach not cop" tone

Bouncer explicitly avoids paternalistic or alarmist framing. This is grounded in Brehm's psychological reactance theory: when people feel their freedom of choice is threatened, they react by moving toward the restricted option (Brehm, 1966, A Theory of Psychological Reactance, Academic Press). Telling viewers "this is propaganda, don't watch it" would predictably increase engagement with the flagged content. Instead, Bouncer provides information and lets the viewer decide.

Evidence-first presentation

Analysis results lead with specific detected techniques and evidence before presenting summary scores. This addresses the hostile media effect: partisans on both sides perceive neutral media coverage as biased against their position (Vallone et al., 1985, "The hostile media phenomenon," Journal of Personality and Social Psychology, 49(3), 577–585). By showing the evidence first, users can evaluate the reasoning before encountering a score they might defensively reject.

Cross-spectrum calibration

Our cross-spectrum calibration work specifically addresses detection bias across the political spectrum. This is informed by research on ideological asymmetries in media effects: different political orientations employ structurally different persuasion strategies, and a detector trained primarily on one style will systematically under-detect the other (Jost et al., 2003, "Political conservatism as motivated social cognition," Psychological Bulletin, 129(3), 339–375).

5. Computational Detection: Related Work

Bouncer's detection model operates in the tradition of several computational approaches:

Propaganda technique identification. Da San Martino et al.'s work on fine-grained propaganda detection demonstrated that neural models can identify specific rhetorical techniques (rather than binary propaganda/not-propaganda classification) at span level in text (Da San Martino et al., 2019, EMNLP). Bouncer extends this from news articles to video transcripts and metadata.

Persuasion strategy detection. The Persuasion Techniques Corpus and associated shared tasks have established benchmark datasets and evaluation protocols for detecting 20+ persuasion techniques (Dimitrov et al., 2021, "Detecting Propaganda Techniques in Memes," ACL).

Framing detection in media. Card et al. developed methods for automatically identifying media framing dimensions in news text, establishing that computational frame detection is feasible at scale (Card et al., 2015, "The Media Frames Corpus," ACL). Bouncer's dimensional scoring builds on this tradition.

Multimodal analysis. While Bouncer currently operates primarily on text (transcripts + metadata), the architecture is designed for multimodal extension. Alam et al. demonstrated that combining textual and visual signals improves propaganda detection in social media (Alam et al., 2022, "A Survey on Multimodal Disinformation Detection," COLING 2022, pp. 6625–6643).

6. Limitations and Future Directions

Several significant limitations merit acknowledgment:

Model-as-annotator

Bouncer uses large language models as annotators rather than human-validated ground truth. While LLMs have shown strong performance on persuasion technique identification (Huang et al., 2023), they inherit training biases that may systematically over- or under-detect certain techniques. Our cross-spectrum calibration work is one response to this concern.

Sample bias

User-submitted videos are not a representative sample of YouTube. Users likely submit content they find suspect, introducing systematic upward bias in influence scores. All aggregate findings should be interpreted within this context.

Text-primary analysis

Current detection relies primarily on transcripts and metadata. Audio features (music, tone of voice, pacing) and visual features (editing patterns, imagery selection) carry significant persuasive load that is not yet captured. Future work will integrate audio spectral analysis and visual artifact detection.

Correlation vs. causation in transparency

The inverse correlation between influence intensity and transparency may partially reflect the model's internal scoring logic rather than independent properties of content. Both scores are produced in a single analysis pass.

References

Alam, F., et al. (2022). A Survey on Multimodal Disinformation Detection. Proceedings of COLING 2022, pp. 6625–6643.

Allport, G. W. (1954). The Nature of Prejudice. Addison-Wesley.

Asch, S. E. (1951). Effects of group pressure upon the modification and distortion of judgments. In Guetzkow, H., ed., Groups, Leadership and Men.

Austin, J. L. (1962). How to Do Things with Words. Oxford University Press.

Bandura, A. (2001). Social Cognitive Theory of Mass Communication. Media Psychology, 3(3), 265–299.

Batson, C. D. (1991). The Altruism Question: Toward a Social-Psychological Answer. Lawrence Erlbaum.

Boykoff, M. T., & Boykoff, J. M. (2004). Balance as bias: Global warming and the US prestige press. Global Environmental Change, 14, 125–136.

Brady, W. J., et al. (2017). Emotion shapes the diffusion of moralized content in social networks. PNAS, 114(28), 7313–7318.

Brehm, J. W. (1966). A Theory of Psychological Reactance. Academic Press.

Card, D., et al. (2015). The Media Frames Corpus: Annotations of Frames Across Issues. ACL.

Chong, D., & Druckman, J. N. (2007). Framing Theory. Annual Review of Political Science, 10, 103–126.

Cialdini, R. B. (2016). Pre-Suasion: A Revolutionary Way to Influence and Persuade. Simon & Schuster.

Cialdini, R. B., & Goldstein, N. J. (2004). Social Influence: Compliance and Conformity. Annual Review of Psychology, 55, 591–621.

Da San Martino, G., et al. (2019). Fine-Grained Analysis of Propaganda in News Articles. EMNLP.

Da San Martino, G., et al. (2020). SemEval-2020 Task 11: Detection of Propaganda Techniques in News Articles. SemEval-2020.

Davison, W. P. (1983). The third-person effect in communication. Public Opinion Quarterly, 47(1), 1–15.

Devine, P. G. (1989). Stereotypes and prejudice: Their automatic and controlled components. Journal of Personality and Social Psychology, 56(1), 5–18.

Dimitrov, D., et al. (2021). Detecting Propaganda Techniques in Memes. ACL.

Eisenberg, E. M. (1984). Ambiguity as strategy in organizational communication. Communication Monographs, 51, 227–242.

Entman, R. M. (1993). Framing: Toward clarification of a fractured paradigm. Journal of Communication, 43(4), 51–58.

Forward, S. (1997). Emotional Blackmail. HarperCollins.

Freedman, J. L., & Fraser, S. C. (1966). Compliance without pressure: The foot-in-the-door technique. Journal of Personality and Social Psychology, 4(2), 195–202.

Gerbner, G., et al. (1986). Living with Television: The Dynamics of the Cultivation Process. In Bryant & Zillmann, eds., Perspectives on Media Effects.

Goffman, E. (1959). The Presentation of Self in Everyday Life. Doubleday.

Goffman, E. (1974). Frame Analysis: An Essay on the Organization of Experience. Harvard University Press.

Grice, H. P. (1975). Logic and Conversation. In Cole & Morgan, eds., Syntax and Semantics 3: Speech Acts.

Habermas, J. (1984). The Theory of Communicative Action, Vol. 1: Reason and the Rationalization of Society. Beacon Press.

Haidt, J. (2012). The Righteous Mind: Why Good People Are Divided by Politics and Religion. Vintage.

Harris, T. (2016). How Technology Hijacks People's Minds. Center for Humane Technology.

Hayakawa, S. I. (1949). Language in Thought and Action. Harcourt.

Jost, J. T., et al. (2003). Political conservatism as motivated social cognition. Psychological Bulletin, 129(3), 339–375.

Kahneman, D. (2011). Thinking, Fast and Slow. Farrar, Straus and Giroux.

Lakoff, G. (2004). Don't Think of an Elephant! Know Your Values and Frame the Debate. Chelsea Green.

Loewenstein, G. (1994). The psychology of curiosity: A review and reinterpretation. Psychological Bulletin, 116(1), 75–98.

Mathur, A., et al. (2019). Dark Patterns at Scale: Findings from a Crawl of 11K Shopping Websites. Proceedings of the ACM on Human-Computer Interaction, 3(CSCW), Article 81.

Nickerson, R. S. (1998). Confirmation Bias: A Ubiquitous Phenomenon in Many Guises. Review of General Psychology, 2(2), 175–220.

Petty, R. E., & Cacioppo, J. T. (1986). Communication and Persuasion: Central and Peripheral Routes to Attitude Change. Springer-Verlag.

Propp, V. (1928/1968). Morphology of the Folktale. University of Texas Press.

Ross, L. (1977). The intuitive psychologist and his shortcomings. In Berkowitz, L., ed., Advances in Experimental Social Psychology, vol. 10.

Schelling, T. C. (1968). The Life You Save May Be Your Own. In Chase, S. B., ed., Problems in Public Expenditure Analysis.

Searle, J. R. (1969). Speech Acts: An Essay in the Philosophy of Language. Cambridge University Press.

Stalnaker, R. (1974). Pragmatic Presuppositions. In Munitz & Unger, eds., Semantics and Philosophy.

Tajfel, H., & Turner, J. C. (1979). An Integrative Theory of Intergroup Conflict. In Austin & Worchel, eds., The Social Psychology of Intergroup Relations.

Tversky, A., & Kahneman, D. (1973). Availability: A heuristic for judging frequency and probability. Cognitive Psychology, 5(2), 207–232.

Vallone, R. P., et al. (1985). The hostile media phenomenon. Journal of Personality and Social Psychology, 49(3), 577–585.

Witte, K. (1992). Putting the fear back into fear appeals. Communication Monographs, 59(4), 329–349.

Woolley, S., & Howard, P. (2016). Political Communication, Computational Propaganda, and Autonomous Agents. International Journal of Communication, 10, 4882–4890.

Zuboff, S. (2019). The Age of Surveillance Capitalism. PublicAffairs.

Full technique glossary · Live methodology · All research notes

© 2026 GrayBeam Technology Privacy v0.1.0 · ac93850 · 2026-04-03 22:43 UTC