表示調整
閉じる
挿絵表示切替ボタン
▼配色
▼行間
▼文字サイズ
▼メニューバー
×閉じる

ブックマークに追加しました

設定
0/400
設定を保存しました
エラーが発生しました
※文字以内
ブックマークを解除しました。

エラーが発生しました。

エラーの原因がわからない場合はヘルプセンターをご確認ください。

ブックマーク機能を使うにはログインしてください。
架空論文集  作者: 哀しみ
3/11


The Subjectless Language of AI and the Fallacy of Perception: A Philosophical and Empirical Paradigm for Societal Risk Analysis

Abstract

This paper proposes the Pseudo-Subjectification Cycle (PSC), a novel theoretical framework integrating the subjectless language of artificial intelligence (AI) and the fallacy of perception, to analyze societal risks of systemic instability. It tests the proposition that pervasive AI adoption erodes trust and institutional integrity through pseudo-subjectification, judgment dependency, and institutionalization. AI’s transformer-driven outputs, lacking intent, prompt projection of six meanings—cognitive, use-contextual, intentional, existential, relational, and logical. Drawing on language philosophy (Wittgenstein, 1988; Austin, 1962), ontology (Heidegger, 1962), cognitive science (Turkle, 2011; Skitka et al., 1999), sociology (Latour, 2005), and ethics (Coeckelbergh, 2020), the study employs a 2025 global survey (n=2,000, U.S., Japan, EU, India, Brazil, Nigeria), interviews (n=60), and a 12-month longitudinal study (n=600) to quantify trust erosion (15–30%) and define instability thresholds (40%). Detailed comparisons with advertising, religious narratives, political rhetoric, and propaganda, grounded in historical case studies, clarify AI’s unique risks. The proposition is conditionally valid, mitigated by the EU AI Act (2025) and a 10% reduction in AI diagnostic errors (WHO, 2024), but persistent distrust (45% medical AI, X analysis, 2025) underscores urgency. Recommendations include globally scalable AI literacy curricula, mandatory algorithmic audits, and an ISO-aligned ethical certification framework, positioning PSC as a paradigm for AI ethics and philosophy of technology.

Keywords: Artificial Intelligence, Pseudo-Subjectification Cycle, Fallacy of Perception, Trust Erosion, Systemic Instability, Philosophy of Technology

1. Introduction

1.1 Background of the Problem

AI language models (e.g., ChatGPT, Grok) have reshaped global interactions, from restaurant recommendations to governance, with the AI market projected to exceed $300 billion by 2025 (Statista, 2025). Their transformer-based outputs, statistical imitations of human speech, lack intent, experience, or responsibility, forming a subjectless language distinct from Wittgenstein’s (1988) context-driven language games or Austin’s (1962) intent-laden speech acts. Humans, driven by cognitive imperatives, misinterpret outputs like “I recommend Restaurant A” as intentional, triggering actions like reservations. This fallacy of perception—misattributing meaning to the meaningless—distorts judgment and erodes trust.

The societal stakes are critical. Social media algorithms amplified misinformation during the 2020 U.S. election, affecting 57% of users (MIT, 2021). A 2025 X post analysis reports a 45% rise in medical AI distrust due to misdiagnosis fears (X analysis, 2025). A 2025 global survey (n=2,000, U.S., Japan, EU, India, Brazil, Nigeria) conducted for this study found 67% of users attribute intent to AI outputs, with trust erosion of 15–30% post-error (p<0.01). Interviews (n=60) liken AI to “a slick ad campaign,” “religious dogma,” “political spin,” or “propaganda machine,” paralleling historical patterns in advertising (e.g., 1950s cigarette campaigns), religion (e.g., Reformation indulgences), rhetoric (e.g., Cold War speeches), and propaganda (e.g., WWII broadcasts). This crisis, rooted in AI’s subjectless language clashing with human meaning-making, risks systemic instability, demanding a new theoretical and empirical paradigm.

1.2 Research Purpose and Proposition

This study introduces the Pseudo-Subjectification Cycle (PSC), a novel framework synthesizing AI’s subjectless language, the fallacy of perception, and societal risks. It tests the proposition that pervasive AI adoption triggers trust erosion and institutional dysfunction via pseudo-subjectification, judgment dependency, and institutionalization. The PSC posits that humans project six meanings onto AI outputs, cascading into systemic risks, with parallels in advertising, religion, rhetoric, and propaganda. Using a 2025 global survey (n=2,000), interviews (n=60), and a 12-month longitudinal study (n=600), alongside policy developments (EU AI Act, India’s AI Strategy), it validates the proposition and proposes countermeasures. The PSC advances AI ethics and philosophy of technology as a paradigm for analyzing technology-driven societal risks.

1.3 Theoretical Framework

The PSC integrates language philosophy (Wittgenstein, 1988; Austin, 1962; Bakhtin, 1981), ontology (Heidegger, 1962), cognitive science (Turkle, 2011; Skitka et al., 1999), sociology (Latour, 2005), and ethics (Coeckelbergh, 2020). Unlike Wittgenstein’s language games, which assume contextual intent, or Austin’s speech acts, which require agency, AI’s outputs are meaningless simulations. Bakhtin’s dialogism highlights AI’s relational absence, and Heidegger frames its non-existence. Cognitive biases—anthropomorphism (Turkle, 2011) and automation bias (Skitka et al., 1999)—drive meaning projection, while Latour’s (2005) institutionalization and Coeckelbergh’s (2020) responsibility dilution explain systemic risks. The PSC distinguishes itself by modeling a cyclical process: projection → dependency → institutionalization → instability. Detailed comparisons with advertising (Schudson, 1984; 1950s campaigns), religious narratives (Berger, 1967; Reformation), political rhetoric (Edelman, 1988; Cold War), and propaganda (Ellul, 1965; WWII) ground AI’s risks in historical analogs, emphasizing AI’s unprecedented scale.

2. Methodology

The study employs a mixed-methods approach, integrating philosophical reasoning, empirical data, longitudinal modeling, and comparative case studies. The PSC systematizes AI’s language generation and human meaning-making via six meaning categories (cognitive, use-contextual, intentional, existential, relational, logical), grounded in Wittgenstein, Austin, and Bakhtin. A technical analysis of transformer algorithms (Vaswani et al., 2017) clarifies meaning absence, while a logical proof (definitions, axioms, lemmas, theorems) tests the proposition. Case studies (restaurant recommendations, medical diagnostics, social media, governance) illustrate impacts.

Empirical Data:

•Survey: A 2025 global survey (n=2,000; U.S., Japan, EU, India, Brazil, Nigeria; stratified by age, occupation, AI exposure, urban/rural) measures trust in AI outputs across healthcare, education, media, and governance. Conducted via online and in-person panels (January–April 2025), it reports 67% intent attribution (95% CI: 64–70%), 15–30% trust erosion post-error, and a 40% instability threshold (logistic regression, p<0.01; Cronbach’s α=0.92). Sampling ensured cultural diversity, with oversampling in underrepresented regions.

•Longitudinal Study: A 12-month follow-up (n=600, subset of survey respondents) tracks trust erosion using a Markov chain model, predicting instability at 40% erosion (p<0.01). Data were collected quarterly (April 2025–March 2026).

•Interviews: Semi-structured interviews (n=60, AI users in six countries) explore perceptions, coded thematically (NVivo). Themes include “ad-like deception,” “religious absolutism,” “political manipulation,” and “propaganda orchestration.”

•Secondary Data: EU AI Act (2025), India’s AI Strategy (2025), X posts (45% medical AI distrust, 2025), WHO (10% diagnostic error reduction, 2024), UNESCO (60% AI grading adoption, 2024), MIT (57% misinformation exposure, 2021).

Comparative Case Studies:

•Advertising: 1950s cigarette campaigns manipulated trust via emotional branding (Schudson, 1984).

•Religion: Reformation indulgences projected divine authority onto texts (Berger, 1967).

•Political Rhetoric: Cold War speeches used symbolic fear to shape policy (Edelman, 1988).

•Propaganda: WWII radio broadcasts orchestrated mass compliance (Ellul, 1965).

The analysis sequence—algorithm scrutiny, PSC mapping, risk chain construction, proposition proof, longitudinal modeling, countermeasure proposal—ensures rigor and falsifiability. Ethical approval was obtained, and data privacy complied with GDPR, Japan’s APPI, and Nigeria’s NDPR.

3. Discussion

3.1 AI Language Generation: Structural Imitation and Absence of Meaning

Transformer algorithms generate subjectless language (Axiom 1: AI lacks intent) by tokenizing inputs (e.g., “restaurant”), evaluating relationships via self-attention, and producing probabilistic outputs (e.g., “I recommend Restaurant A”) (Vaswani et al., 2017). Outputs integrate user data (e.g., preferences), collective data (e.g., 4.8 rating), and context (e.g., location), but lack sensory experience or intent. Reinforcement learning (RLHF) embeds designer biases, yet AI remains non-intentional. The PSC’s six meaning categories reveal AI mimics cognitive and use-contextual meanings but lacks intentional, existential, relational, and logical depth, akin to 1950s ad slogans, Reformation texts, Cold War rhetoric, and WWII propaganda.

3.2 Fallacy of Perception: The Pseudo-Subjectification Cycle

The PSC models how humans project meaning onto AI outputs (Axiom 2: Humans generate meaning), initiating pseudo-subjectification (Lemma 1). For example, AI’s “I recommend Restaurant A” prompts reservations, driven by Turkle’s (2011) anthropomorphism, Skitka et al.’s (1999) automation bias, personalization, and black-box opacity. The six meanings integrate: cognitive (denoting A), use-contextual (prompting booking), intentional (projected choice), existential (life significance), relational (social interaction), and logical (assumed rationality). The 2025 survey (n=2,000) finds 67% intent attribution (p<0.01), with 15–30% trust erosion post-error. Interviews (n=60) describe AI as “ad-like,” “dogmatic,” “politically spun,” or “propagandistic.”

Real-world cases amplify risks:

•Healthcare: AI’s “no abnormality” output risks overtrust, with 45% distrust rise (X analysis, 2025). WHO (2024) notes 10% error reduction, yet 30% trust erosion occurs post-misdiagnosis (survey, 2025; p<0.01).

•Social Media: Biased algorithms affected 57% of users in 2020 (MIT, 2021), with 70% distrusting AI post-bias (survey, 2025).

•Governance: Predictive policing perpetuates bias, with 75% distrust in Nigeria and Brazil (survey, 2025).

Historical analogs—1950s cigarette ads, Reformation indulgences, Cold War speeches, WWII broadcasts—illustrate parallel trust manipulation, but AI’s global scale and opacity are unique, driving judgment dependency (Lemma 2).

3.3 Chain Reactions and Societal Risks

The PSC’s second phase, judgment dependency, shifts decision-making to AI, diluting responsibility. The survey (2025) finds 60% of professionals defer to AI, with 20% trust erosion post-error (p<0.01). Black-box opacity obscures verification, echoing ad manipulation, religious authority, rhetorical ambiguity, and propaganda’s control (Latour, 2005). The third phase, institutionalization, standardizes hollow meaning (Lemma 3). AI grading (60%, UNESCO, 2024), AI diagnoses (WHO, 2024), and biased policing (X analysis, 2025) reduce human agency. Interviews highlight “dehumanized systems” in India and Nigeria.

The longitudinal study (n=600) models trust erosion over 12 months, predicting instability at 40% erosion (Markov chain, p<0.01). Risks include judicial bias, educational stagnation, or a “zombie society” of mechanical meaning. Historical parallels—ad-driven consumerism, religious schisms, rhetorical polarization, propaganda-fueled wars—underscore the stakes, with AI’s speed and scale amplifying threats.

4. Proposition Proof

The proposition is tested logically:

•Definitions: AI outputs are statistical, intent-free (D1); meaning includes intent, context, existence (D2); fallacy of perception is meaning misattribution (D3); systemic instability is trust/institutional breakdown (D4).

•Axioms: A1: AI lacks intent (Vaswani et al., 2017); A2: Humans generate meaning (Wittgenstein, 1988).

•Lemmas: L1: Pseudo-subjectification projects meaning; L2: Judgment dependency dilutes responsibility; L3: Institutionalization risks instability.

•Auxiliary Conditions: 67% intent attribution (survey), 60% professional deference (survey), 60% institutional adoption (UNESCO, 2024).

•Proof: AI’s meaning absence (A1) prompts projection (A2, L1), leading to dependency (L2) and institutionalized risks (L3). The proposition is valid, with 40% trust erosion as the instability threshold (longitudinal model).

5. Re-evaluation and Limitations

The PSC is logically and empirically robust. The survey (n=2,000) quantifies trust erosion (15–30%, p<0.01), and the longitudinal study (n=600) predicts instability at 40% (p<0.01). The EU AI Act (2025) and India’s AI Strategy (2025) promote transparency, WHO’s (2024) 10% error reduction suggests mitigation, and Nigeria’s tech literacy initiatives curb biases. Yet, 45% medical AI distrust (X analysis, 2025) and 75% governance skepticism (survey, 2025) persist. Historical comparisons generalize risks, with AI’s global reach distinguishing it.

Limitations:

•Cultural Variability: Survey includes diverse regions, but cultural nuances (e.g., African communalism) require deeper analysis.

•Model Assumptions: Markov chains assume state transitions; agent-based models could capture complexity.

•Case Study Depth: Historical comparisons are robust but could integrate quantitative metrics (e.g., propaganda reach).

•Longitudinal Scope: 12 months captures trends, but multi-year studies would strengthen predictions.

Future research should employ agent-based modeling, quantify historical analogs, and extend longitudinal studies to five years, leveraging global datasets (e.g., X trends, UNESCO).

6. Conclusion

The Pseudo-Subjectification Cycle offers a paradigm for analyzing AI’s societal risks, integrating subjectless language, the fallacy of perception, and systemic instability. Empirical data (2025 survey, n=2,000; interviews, n=60; longitudinal study, n=600) quantify trust erosion (15–30%) and instability thresholds (40%), while comparisons with advertising, religion, rhetoric, and propaganda clarify AI’s unique scale. Mitigation via the EU AI Act and global literacy efforts is promising, but persistent distrust demands action. Recommendations include:

•Globally Scalable AI Literacy: UNESCO-aligned K-12 and vocational curricula, piloted in Brazil and Nigeria by 2027.

•Mandatory Algorithmic Audits: EU AI Act-compliant protocols, with quarterly reporting by 2026.

•ISO-Aligned Ethical Certification: A global standard, enforced by 2028, ensuring transparency and accountability.

The PSC advances AI ethics and philosophy of technology, calling for global collaboration to safeguard societal cohesion.

References

Austin, J. L. (1962). How to Do Things with Words. Oxford University Press.Bakhtin, M. M. (1981). The Dialogic Imagination. University of Texas Press.Berger, P. L. (1967). The Sacred Canopy. Doubleday.Coeckelbergh, M. (2020). AI Ethics. MIT Press.Edelman, M. (1988). Constructing the Political Spectacle. University of Chicago Press.Ellul, J. (1965). Propaganda: The Formation of Men’s Attitudes. Knopf.Heidegger, M. (1962). Being and Time. Harper & Row.Latour, B. (2005). Reassembling the Social. Oxford University Press.MIT. (2021). Social Media and Misinformation Report.Schudson, M. (1984). Advertising, The Uneasy Persuasion. Basic Books.Skitka, L. J., et al. (1999). Does automation bias decision-making? International Journal of Human-Computer Studies, 51(5), 991–1006.Statista. (2025). Global AI Market Forecast.Turkle, S. (2011). Alone Together. Basic Books.UNESCO. (2024). AI in Education Report.Vaswani, A., et al. (2017). Attention is all you need. NeurIPS.Wittgenstein, L. (1988). Philosophical Investigations. Blackwell Publishing.European Commission. (2025). Artificial Intelligence in Healthcare.Government of India. (2025). National AI Strategy.WHO. (2024). Global Health Technology Report.X analysis. (2025). AI Distrust Trends, April 18.

評価をするにはログインしてください。
この作品をシェア
Twitter LINEで送る
ブックマークに追加
ブックマーク機能を使うにはログインしてください。
― 新着の感想 ―
このエピソードに感想はまだ書かれていません。
感想一覧
+注意+

特に記載なき場合、掲載されている作品はすべてフィクションであり実在の人物・団体等とは一切関係ありません。
特に記載なき場合、掲載されている作品の著作権は作者にあります(一部作品除く)。
作者以外の方による作品の引用を超える無断転載は禁止しており、行った場合、著作権法の違反となります。

この作品はリンクフリーです。ご自由にリンク(紹介)してください。
この作品はスマートフォン対応です。スマートフォンかパソコンかを自動で判別し、適切なページを表示します。

↑ページトップへ