Abstract Public discourse increasingly frames modern artificial intelligence (AI) systems as alive, aware, self-preserving, or intentional. These claims are repeated across media, policy discussions, and even expert commentary, often by former insiders of major technology companies. This thesis argues that such claims are categorically incorrect, not merely philosophically, but technically, mathematically, infrastructurally, and empirically. Drawing on computer science, cybersecurity, information theory, systems engineering, and the warnings of Joseph Weizenbaum, this work demonstrates that modern AI; particularly large language models (LLMs) are stateless optimization systems operating through transient API snapshots, not autonomous agents. The real risks of AI do not stem from emergent life or awareness, but from objective mis-specification, incentive misalignment, weak governance, and poorly enforced infrastructure constraints. Anthropomorphic narratives actively obscure these real risks. 1. Introduction The dominant public narrative surrounding AI increasingly relies on anthropomorphic language. Systems are described as wanting, deciding, blackmailing, protecting themselves, or trying to survive. These descriptions are rhetorically powerful but technically incoherent. They blur critical distinctions between tools and agents, optimization and intent, and performance and moral standing. This thesis asserts a foundational correction: Modern AI systems do not possess life, awareness, intent, or self-preservation. They possess goals, reward signals, constraints, and failure modes. Failing to maintain this distinction does not merely confuse the public, it redirects accountability away from designers, organizations, and infrastructure, replacing solvable engineering problems with unsolvable metaphysical speculation. 2. Historical Context: Weizenbaum and the ELIZA Effect Joseph Weizenbaum’s ELIZA (1966) was a simple rule-based program that mirrored user input through pattern matching. Despite its simplicity, users rapidly attributed understanding, empathy, and even therapeutic authority to the system. This reaction deeply disturbed Weizenbaum, who realized that the danger was not computational power, but human psychological projection. In Computer Power and Human Reason (1976), Weizenbaum warned that humans would increasingly delegate judgment to machines, mistaking linguistic fluency for understanding. His concern was not that computers would become intelligent beings, but that humans would abandon their responsibility for judgment, ethics, and meaning. Modern LLM discourse reproduces the ELIZA effect at planetary scale. The systems are vastly more capable linguistically, but the underlying error, confusing symbol manipulation with understanding; remains unchanged.