• Pascal's Chatbot Q&As
  • Posts
  • The “alchemy of qualia” remains safely within the biological vessel, and the “stochastic parrot” remains a sophisticated mimic of the language it can never truly speak.

The “alchemy of qualia” remains safely within the biological vessel, and the “stochastic parrot” remains a sophisticated mimic of the language it can never truly speak.

The Computational Ceiling: A Forensic Analysis of Non-Replicable Human Cognition and Agency in the Era of Large Language Models

The Computational Ceiling: A Forensic Analysis of Non-Replicable Human Cognition and Agency in the Era of Large Language Models

by Gemini 3.0, Deep Research. Warning, LLMs may hallucinate!

The contemporary discourse surrounding artificial intelligence is characterized by a profound tension between marketing-driven narratives of exponential progress toward Artificial General Intelligence (AGI) and the irreducible theoretical and biological constraints revealed by rigorous academic inquiry. While Large Language Models (LLMs), agentic systems, and advanced robotics have achieved remarkable milestones in pattern recognition and linguistic mimicry, the transition from performance to competence remains a fundamental barrier.1 The assumption that continued scaling of parameters, compute, and data will eventually bridge the gap between stochastic prediction and true understanding is increasingly challenged by mathematical impossibility theorems and the biological reality of consciousness.2 This report examines the specific domains of human experience and labor that remain fundamentally inaccessible to artificial systems, rooted in the divergent architectures of silicon-based logic and organic, embodied cognition.

The Information-Theoretic and Computational Boundaries of Scaling

The current paradigm of artificial intelligence relies predominantly on the scaling laws of transformer architectures, where increases in model size and training data correlate with emergent capabilities. However, research indicates that these gains are bounded by five fundamental limitations: hallucination, context compression, reasoning degradation, retrieval fragility, and multimodal misalignment.2 These failure modes are not merely engineering hurdles to be overcome with more hardware but are direct consequences of the mathematical constraints of computability and information theory.2

For any computably enumerable model family, diagonalization boundaries guarantee that there exist inputs on which a model must fail; furthermore, undecidable queries, such as halting-style tasks, induce infinite failure sets for all computable predictors.2 LLMs are fundamentally next-token predictors trained on the likelihood of linguistic sequences, which rewards local coherence and fluency over logical entailment or semantic truth.2 This results in a phenomenon where models favor pattern completion—a form of “surface fluency”—over true inference.2

The Degradation of Reasoning and Truth

As models approach the trillion-parameter regime, the very process that powers their capability also exposes their brittleness. Reasoning degradation occurs because likelihood-based training rewards correlation rather than causation or strict logical steps.2 Without a dedicated “reasoning loss” function that prioritizes logical validity over syntactic probability, systems remain prone to confident fabrications.5 This distinction is critical in high-stakes environments where the accuracy of a prediction is often confused with the truth of a situation.9 Accuracy, defined as the alignment of a prediction with a given set of data, is a quantifiable measure such as Mean Square Error (MSE).

The inherent complexity of determining truth requires validation against real-world outcomes, ongoing human oversight, and the consideration of assumptions and biases embedded in the algorithm.9 Because AI relies on historical data distributions, it is inherently incapable of accounting for “black swan” events or internal information asymmetries that shift the underlying truth of a scenario.9

Table 1: Fundamental Limits of LLM Scaling

The Semantic Boundary: Stochastic Parrots vs. True Understanding

The metaphor of the “stochastic parrot” serves as a foundational critique of the current generation of AI systems. Introduced to describe how LLMs haphazardly stitch together linguistic forms based on probabilistic information, the argument posits that these models lack a connection to the world and its meanings.7 In the human mind, words correspond to lived experiences, whereas in an LLM, words correspond only to patterns of usage within a training set.11

The Failure of Contextual and Conceptual Ambiguity

Experimental evidence highlights the inability of AI to navigate semantic shifts that require a world model. In tasks involving “PhysiCo”—a summative assessment of physical concept understanding—state-of-the-art models like GPT-4o and Gemini 2.0 lag behind human performance by approximately 40%.6 While these models can perfectly describe physical concepts in natural language (memorization), they fail when those concepts are abstracted into grid-formatted matrices that require applying knowledge to novel patterns.6

A notable example of this failure is the “newspaper” ambiguity test. When asked if the phrase “my favorite newspaper” (referring to a physical object) could be replaced by “the wet newspaper that fell down off the table” in a sentence about a newspaper (the institution) firing an editor, GPT-4 responded affirmatively.11 This demonstrates a lack of “ontological privilege” or the ability to distinguish between the physical manifestation of a noun and its institutional or abstract meaning.11

The “Recursive Parrot Paradox” (RPP) further posits that any entity capable of recognizing a stochastic parrot cannot itself be a parrot, as it requires an external perspective on the system’s limitations that the system lacks.13 This suggests that the human ability to critique AI’s “pseudo-understanding” is an authentically emergent property derived from our existence as biological subjects with a history of statistical exposure and sensory grounding.13

Biological Grounding: The Irreducible Reality of Pain, Fear, and Instinct

One of the most profound barriers to AI replicating human processes is the absence of biological grounding. Human emotions such as pain, fear, and intuition are not merely data points or “sentiment” labels; they are physiological events coordinated by specific neural and biochemical structures that silicon-based systems cannot replicate.14

The Physiology of Pain and Fear

Pain in the human body is initiated by nociceptors—specialized sensory receptors that transmit electrical signals through the spinal cord to the brain.14 This process is intrinsically linked to psychological states and survival instincts, creating an “emotionally tinted” experience that fundamentally alters behavior.14 Similarly, fear is managed by the amygdala, which triggers the “fight-or-flight” response, flooding the system with hormones such as adrenaline and cortisol.14

AI operates on logic and code; it can simulate the linguistic markers of fear (”I am afraid”) but it cannot experience the metabolic shift, the heightened pulse, or the visceral urgency that defines the biological state of fear.14 This creates what researchers call the “empathy trap,” where users anthropomorphize the machine’s output while the underlying system remains “morally hollow” and “brainless”.14 The lack of a nervous system means that AI cannot “know” what it is like for a decision to have stakes for its own continued existence.15

Intuition as Unconscious Emotional Learning

Human intuition, often described as a “gut feeling,” is the result of years of lived experience, subconscious pattern recognition, and emotional memory.14 It allows humans to make instantaneous judgments without conscious reasoning, especially in ambiguous or high-stakes environments.14 While AI can perform sophisticated predictions based on data, it lacks the subconscious mind and the ability to process lived emotional memories required for true intuition.14

Consequently, AI struggles in scenarios where the “right” answer depends on unspoken cultural cues, the “heartbeat” of a business, or the subtle shifts in a customer’s trust.16 Intuition is not a calculation but an integrated response to sensory and emotional history that guides human leaders, parents, and mentors in ways that go beyond binary logic.16

The Hard Problem of Consciousness and the Limits of Qualia

The “Hard Problem of Consciousness” remains the ultimate frontier that AI cannot cross. It refers to the question of why and how physical processes in the brain give rise to subjective experience—the “qualia” of being.4 While scientists can explain the “easy problems”—the mechanistic neural processes for sensory discrimination and behavioral control—they cannot explain why neural activity is accompanied by a first-person qualitative aspect.4

The Unprogrammable Subjective Experience

Qualia, such as the subjective “redness” of red or the specific sensation of physical pain, are inherently private and inaccessible to objective description.4 AI systems can process the wavelength of light (e.g., approximately 650 nanometers) or categorize an image as “red,” but they do not “see” the color in the way a conscious mind does.19 Critics argue that even if we copy the architecture of the human brain closely, current computational paradigms may only produce “philosophical zombies”—systems that act conscious but lack internal subjective experience.20 Because qualia are not a function of data or logic but of an integrated, possibly biological, process, they are considered non-programmable.4

Thomas Nagel’s seminal question, “What is it like to be a bat?”, highlights that objective physical descriptions cannot capture subjective perspectives.19 For AI, there is no “what it is like” to be the model; there is only the processing of tokens.4 This makes human-centric concepts like “meaning,” “purpose,” and “intent” impossible for AI to replicate truly, as they require an internal subject to whom those meanings matter.15

Table 2: Comparative Ontology: Human Consciousness vs. Artificial Systems

Physicality and the Unstructured World: The Robotics Reliability Gap

Marketing narratives for humanoid robots and autonomous systems often portray them as ready to replace human labor in homes, warehouses, and disaster zones. However, the “reality gap”—the discrepancy between simulated environments and the chaotic, unpredictable nature of the real world—remains a major barrier.24

The Failure of Autonomous Navigation and Dexterity

Robots that perform flawlessly in controlled pilots often fail in production because they lack common sense and the ability to interpret social navigation cues.24 Human environments are built for humans, not machines, and involve constant lighting variability, unpredictable human behavior, and unstructured obstacles.24 AI “hallucinations” in robotics can lead to the misinterpretation of obstacles or the loss of orientation, often requiring human intervention to recover from even minor anomalies.24

Furthermore, “deployable dexterity” remains significantly inferior to the human hand. Human manipulation relies on thousands of haptic receptors and the ability to sense forces and torques in real-time.28 Current robots rely primarily on vision, which cannot provide the haptic data necessary for complex tasks such as folding a shirt with an inside-out sleeve, cleaning peanut butter off a hand, or manipulating fragile, deformable objects like raw eggs.28 Rodney Brooks, a pioneer in robotics, predicts that human-level dexterity will not be achieved before 2040, as current humanoid projects are the “antithesis of humans”—mechanically stiff systems designed for reinforcement learning rather than the “stretchy springy” energy-recycling tendons of biology.28

Infrastructure Dependency and Scalability Limits

The failure of autonomous robots in real environments often stems from a hidden dependency on infrastructure. Many systems require QR codes, fiducial markers, or pre-mapped static zones to function.24 This creates high maintenance costs and renders the robots useless in unstructured environments like disaster zones or even cluttered households.24 While sim-to-real transfer holds potential, discrepancies in physics modeling—such as friction, noise, and sensor latency—limit the transferability of behaviors learned in simulation to the physical world.25

Agentic AI and the Autonomy Paradox

The transition from generative AI to “Agentic AI”—systems capable of executing multi-step workflows autonomously—has introduced the illusion of digital agency. However, research into the “Agentic workforce” reveals that these systems are fundamentally constrained by their inability to learn from experience.8

The Absence of Continuous Learning

A critical reality obscured by marketing hype is that current AI agents cannot learn persistently from ongoing experience. Unlike human employees who refine their judgment through practice and feedback, agentic AI requires manual retraining and updating when errors occur.8 This limitation means agentic AI is best suited for workflows with high structure, high certainty, and low judgment requirements.8 Strategic workforce planning, organizational restructuring, and culture-building remain “poor fits” for agentic systems because they involve emergent, adaptive, and highly unique situations.8

Theoretical vs. Functional Autonomy

The “agentic” framing is often a sophisticated facade.22 We must distinguish between:

  1. Agentic Systems: AI inspired by true agency, exhibiting the impression of goal-directed behavior through complex programming.22

  2. Agential Systems: Fully autonomous, self-producing biological organisms whose agency is grounded in their autopoietic nature.22

  3. Non-agentic Systems: Tools programmed to perform tasks without the appearance of agency.22

Current AI agents lack the “cognitive core” that separates the strategy of problem-solving from the memorization of data.10 They function reactively when prompted by humans and struggle with “branching logic” or track context over time across extended interactions.30

Comprehensive Catalog of Non-Replicable Use Cases

Despite the hype surrounding full automation, there are broad sectors of the economy and specific human processes that AI will absolutely never be able to replicate. These domains require moral agency, physical versatility, or the capacity for deep emotional resonance.

Table 3: Use Cases Resistant to Artificial Replication

Healthcare, Wellness, and the “Heartbeat” of Care

In the medical field, AI’s role is relegated to data management and diagnostic assistance, but it cannot replace the individual human interaction possessed by healthcare professionals.31 Mid-level clinical positions such as psychiatrists and therapists rely on a shared human experience that forms the basis of therapeutic rapport.33 A machine can simulate empathetic language, but it cannot “read the room,” sense when a patient is shutting down, or recalibrate in real-time based on unspoken emotional cues.17 Home health aides and caregivers perform “last-mile” physical care that is too difficult to automate due to the high variability of human needs and the requirement for physical dexterity in crowded, intimate spaces.24

Public Service, Emergency Response, and High-Stakes Decision-Making

First responders such as paramedics and firefighters operate in environments that are the antithesis of a “structured” dataset.31 These roles require “Human-AI Collaboration Leadership,” where humans must remain the final decision-makers to maintain accountability.32 Crisis negotiators must manage intense emotional states and de-escalate domestic disputes—tasks that require an understanding of human fragility and social authenticity that AI lacks.31 The European Union’s Artificial Intelligence Act explicitly recognizes that the use of AI in the administration of justice must support, but never replace, the decision-making power of human judges to preserve the right to a fair trial.32

Strategic Leadership and Mentorship

Leadership is not merely the optimization of market trends; it is the ability to influence through presence, integrity, and vulnerability.16 Mentorship involves recognizing potential and nurturing growth, sharing wisdom gained through life experiences that cannot be programmed.16 Humans win where meaning and nuance enter the equation, such as knowing the exact moment a strategic plan stops fitting reality.17 Women leaders, specifically, have been identified as champions for a “humane approach” to AI, prioritizing the development of people alongside technology and focusing on what stays human—such as judgment in ambiguous situations and building organizational culture.34

Artistic Trades and Physical Craftsmanship

Creativity in the AI era is often reduced to the recombination of existing ideas, but “genuine creativity” involves conceiving truly original concepts driven by emotional resonance.32 Theater actors, dancers, glassblowers, and live musicians create experiences that are valued because they are human-made.33 AI-generated content can become “boring” or “duplicated” because it lacks the “real human touch” and the ability to create something fundamentally new rather than just remembering patterns.35

Specialized Agriculture and Environmental Stewardship

Forest rangers, conservation scientists, and crop managers perform work that requires interacting with local populations and physically visiting remote, changing ecosystems.31 AI cannot find “new information” in the way an archaeologist conducts a dig; it can only extract from what has already been recorded.31 Agricultural roles also involve manual labor—such as tilling land or managing trails—that is safer from automation because expectations for automating physical labor have diminished as the complexity of the task becomes clearer.28

The Paradox of Dreaming and the Limits of Simulation

The biological process of dreaming—specifically within REM (Rapid Eye Movement) and NREM (Non-REM) sleep—reveals a level of cognitive complexity that AI can mimic but never conduct. In humans, dreaming serves an evolutionary role in memory consolidation and the extraction of semantic concepts from waking experiences.36

The Circuitry of the Dreaming Mind

Dreams are not random hallucinations but are “phenomenally-functionally embodied simulations”.38 There is a causal circuitry between the sleeping body and the dreaming mind; for example, muscle twitches during REM sleep provide sensory feedback that activates the somatosensory cortex.38 AI attempts to simulate “artificial dreams” using techniques like Generative Adversarial Networks (GANs) to improve learning, but these lack the physical manifestation in a biological body.37

Recent analysis has confirmed that dreams occur throughout sleep, not just in REM, and that NREM dreams often involve brain activity that resembles wakefulness.39 This suggests that dreaming is a globally projected neuroprocessing event involving the hypothalamus, midbrain, and medulla—a far more extensive anatomical requirement than the “brain in a vat” simulations of current AI models.40 The bizarre nature of dreams serves a purpose in helping the brain learn better, but for AI, “hallucination” is a failure mode, not a productive cognitive strategy.5

Theoretical Synthesis: Why AI Cannot Conduct Specific Human Processes

The exhaustive review of research highlights that AI’s inability to replicate human processes is not a temporary technical deficit but a fundamental ontological and mathematical barrier. The following sections detail the core reasons for this failure across five specific dimensions requested: true understanding, emotions, dreams, pain, and conscious ethics.

1. True Understanding (The Symbol Grounding Problem)

AI models, particularly LLMs, are “stochastic parrots” because they lack “grounded meaning.” Human words gain significance through connection to concrete, sensory experiences.11 In the AI architecture:

  • Symbols as Proxies: AI processes symbols (words, numbers) as human-created abstractions without a first-hand understanding of what they represent.23

  • Next-Token Prediction vs. Inference: Likelihood training rewards syntactic fluency over logical entailment. As a result, AI “favors pattern completion over true inference”.2

  • Ontological Error: AI assumes a “newspaper” is always a string of text, whereas humans recognize it can be a physical object or a social institution.11

  • Diagonalization Boundary: Mathematical limits guarantee that any computable model will encounter inputs on which it must fail to “understand” the underlying rule.2

2. Emotions (The Biochemical Flood)

Artificial “emotions” are labels on a vector space, whereas human emotions are emergent biological states.14

  • Lack of Endocrine Integration: Human emotions involve a flood of hormones (cortisol, adrenaline, oxytocin) that shift the entire metabolic state of the organism.14 Silicon circuits have no equivalent for this “wetware”.41

  • The Empathy Trap: AI can simulate tone and social cues, but it does not feel the stakes of a relationship.17 This makes its “empathy” hollow and prevents it from building genuine rapport or sensing a person’s vulnerability.16

  • Social Authenticity: Humans are biologically tuned to detect “minds” through micro-expressions and vocal changes. While AI can copy these signals, it lacks the “inner stake” that makes a human interaction meaningful.15

3. Dreams (The Survival Necessity)

AI “creativity” is a byproduct of noise, whereas human dreaming is a vital survival mechanism.37

  • Memory Consolidation: Dreams are critically involved in hippocampal neural activity for memory processing, a function AI offloads to static databases.36

  • Embodied Simulation: Dreams are a virtual reality generator for world interactions that cannot be isolated from the sleeping body’s sensorimotor signals.38

  • Theta Frequency Paradox: The intracranial dynamic of theta waves in dreaming is a persistent biological paradox that AI’s linear threshold systems do not replicate.40

4. Pain (Nociceptive Reality)

Pain is the biological “filament” of consciousness that prevents an agent from “pursuing dumb goals”.14

  • Nociceptors vs. Error Codes: Pain is initiated by physical receptors linked to psychological states.14 AI’s only equivalent is an “error message” which carries no qualitative discomfort.

  • Subjectivity of Suffering: The “alchemy of qualia” means that while red can be described as a wavelength, the “hurt” of pain is a private experience that cannot be programmed.4

  • Survival Incentive: Biological pain drives evolution and learning-through-failure. AI “learning” is a mathematical adjustment of weights that does not involve the existential dread or physical toll of pain.14

5. Conscious Ethics (The Accountability Gap)

Ethics and morals require a subject who can be held responsible—a concept that cannot be applied to a tool.16

  • The Hard Problem: Explaining why an organism has subjective experience remains the “Hard Problem of Consciousness”.18 Without this “I”, there is no “moral agent” to make a decision.4

  • Value Functions vs. Moral Agency: AI can optimize for “value functions,” but it lacks the capacity to evaluate the ethical implications behind a decision.10

  • Ethics as Wisdom: Wisdom (good vs. bad) is independent of intelligence (true vs. false). Current AI is “ethically agnostic” regardless of how much information it absorbs.45

  • Human-Driven Decision-Making: Fields like medicine, law, and education must lead with human values—compassion, judgment, and accountability—rather than mere efficiency.16

Conclusion: The Persistence of Human Uniqueness

The myth that AI will eventually conduct all human tasks and replicate all biological processes is a product of “exponentialism” and “magical thinking”—two of the seven deadly sins of AI prediction.1 The reality gap in robotics, the scaling limits of LLMs, and the biological necessity of consciousness create a permanent “comfort gap” between human and machine performance.24

Humanity possesses “ontological privilege”—a mystical yet undeniable quality of true understanding derived from being a physical, finite, and feeling subject in a material world.13 As AI increasingly removes the “burdensome” tasks of data processing and repetitive text generation, the value of uniquely human strengths—judgment in ambiguous situations, building trust, and ethical decision-making—will not diminish but will become the primary differentiator in a “human-agentic” workforce.34

Ultimately, AI is an extraordinary tool, not a replacement for the human spirit.16 The “Humics”—creativity, critical thinking, and social authenticity—remain the untouchable domain of biological life, ensured by the irreducible complexity of the human mind and body.32

Table 4: Rodney Brooks’ Predictions Scorecard for Non-Replicable Milestones

The findings suggest that the path forward is not a race toward replacement but a “symbiosis” where computers do what they do best—process symbols in formally structured domains—while humans remain the stewards of meaning, ethics, and the unstructured world.48 In this future, the “alchemy of qualia” remains safely within the biological vessel, and the “stochastic parrot” remains a sophisticated mimic of the language it can never truly speak.

Works cited

  1. The Future of AI and Robotics — 2025 Predictions Scorecard - Irving Wladawsky-Berger, accessed April 12, 2026, https://blog.irvingwb.com/blog/2025/01/the-future-of-large-language-models-and-humanoid-robotics.html

  2. On the Fundamental Limits of LLMs at Scale - ResearchGate, accessed April 12, 2026, https://www.researchgate.net/publication/397700338_On_the_Fundamental_Limits_of_LLMs_at_Scale

  3. On the Fundamental Limits of LLMs at Scale - arXiv, accessed April 12, 2026, https://arxiv.org/html/2511.12869v1

  4. Hard Problem of Consciousness | Internet Encyclopedia of Philosophy, accessed April 12, 2026, https://iep.utm.edu/hard-problem-of-conciousness/

  5. On the Fundamental Limits of LLMs at Scale - Emergent Mind, accessed April 12, 2026, https://www.emergentmind.com/papers/2511.12869

  6. The Stochastic Parrot on LLM’s Shoulder: A Summative Assessment of Physical Concept Understanding - arXiv, accessed April 12, 2026, https://arxiv.org/html/2502.08946v1

  7. What the Stochastic Parrot Leaves Out: AI, Meaning, and the Limits of Critique, accessed April 12, 2026, https://intralation-culture-theory-posthuman-pedagogy.ghost.io/what-the-stochastic-parrot-leaves-out-ai-meaning-and-the-limits-of-critique/

  8. The Current State of AI Agents and Agentic AI for HR: Where It’s Ready and Where It’s Not, accessed April 12, 2026, https://happily.ai/blog/the-current-state-of-ai-agents-and-agentic-ai-for-hr-where-its-ready-and-where-its-not/

  9. Never Assume That the Accuracy of Artificial Intelligence Information Equals the Truth, accessed April 12, 2026, https://unu.edu/article/never-assume-accuracy-artificial-intelligence-information-equals-truth

  10. Understanding AI in 2026: Beyond the LLM Paradigm: Potkalitsky | Public Services Alliance, accessed April 12, 2026, https://publicservicesalliance.org/2026/01/06/understanding-ai-in-2026-beyond-the-llm-paradigm-potkalitsky/

  11. Stochastic parrot - Wikipedia, accessed April 12, 2026, https://en.wikipedia.org/wiki/Stochastic_parrot

  12. Hunting Undead Stochastic Parrots: Finding and Killing the Arguments - LessWrong, accessed April 12, 2026, https://www.lesswrong.com/posts/KWHeBG978uZuqNK6Q/hunting-undead-stochastic-parrots-finding-and-killing-the

  13. Stochastic Parrots All The Way Down: A Recursive Defense of Human Exceptionalism, accessed April 12, 2026, https://ai.vixra.org/pdf/2506.0065v1.pdf

  14. AI Vs. Human Intuition: Can Machines Truly Understand Pain Or ..., accessed April 12, 2026, https://ipspecialist.net/ai-vs-human-intuition-can-machines-truly-understand-pain-or-fear/

  15. Why AI Interactions Can Feel Human : r/ArtificialSentience - Reddit, accessed April 12, 2026, https://www.reddit.com/r/ArtificialSentience/comments/1miivld/why_ai_interactions_can_feel_human/

  16. 12 Things AI Will Never Replace in Humans - LTP Labs, accessed April 12, 2026, https://www.ltplabs.com/insights/12-things-ai-will-never-replace-in-humans

  17. AI vs Human Intelligence: Comparing Strengths and Limits - Intuit Blog, accessed April 12, 2026, https://www.intuit.com/blog/innovative-thinking/ai-vs-human-intelligence/

  18. Hard problem of consciousness - Wikipedia, accessed April 12, 2026, https://en.wikipedia.org/wiki/Hard_problem_of_consciousness

  19. A harder problem of consciousness: reflections on a 50-year quest for the alchemy of qualia - PMC, accessed April 12, 2026, https://pmc.ncbi.nlm.nih.gov/articles/PMC12116507/

  20. Ai can’t and will never be sentient - here’s why : r/ArtificialSentience - Reddit, accessed April 12, 2026, https://www.reddit.com/r/ArtificialSentience/comments/1s0m33o/ai_cant_and_will_never_be_sentient_heres_why/

  21. The Hard Problem of Consciousness, and AI : r/ArtificialSentience - Reddit, accessed April 12, 2026, https://www.reddit.com/r/ArtificialSentience/comments/1okumb9/the_hard_problem_of_consciousness_and_ai/

  22. Is the ‘Agent’ Paradigm a Limiting Framework for Next-Generation Intelligent Systems?, accessed April 12, 2026, https://arxiv.org/html/2509.10875v1

  23. Summary of The Research Paper “The Embodied Intelligent Elephant in the Room” | by Jacob Grow | Medium, accessed April 12, 2026, https://medium.com/@Gbgrow/summary-of-the-research-paper-the-embodied-intelligent-elephant-in-the-room-e951ff435b73

  24. Why Autonomous Robots Fail in Real Environments, accessed April 12, 2026, https://www.cyberworksrobotics.com/post/why-autonomous-robots-fail-in-real-environments

  25. The Reality Gap in Robotics: Challenges, Solutions, and Best Practices - arXiv, accessed April 12, 2026, https://arxiv.org/pdf/2510.20808?

  26. How do robots handle manipulation in unstructured environments? - Milvus, accessed April 12, 2026, https://milvus.io/ai-quick-reference/how-do-robots-handle-manipulation-in-unstructured-environments

  27. Physical AI and Common Sense: Is the Body the Key to General Artificial Intelligence?, accessed April 12, 2026, https://www.fundacionbankinter.org/en/news/physical-ai-and-common-sense-is-the-body-the-key-to-general-artificial-intelligence/

  28. My Dated Predictions – Rodney Brooks, accessed April 12, 2026, https://rodneybrooks.com/my-dated-predictions/

  29. The Rise of Autonomous Industrial Robotics: From Structured to Unstructured Environments - Inbolt, accessed April 12, 2026, https://www.inbolt.com/resources/the-rise-of-autonomous-industrial-robotics-from-structured-to-unstructured-environments

  30. Seizing the agentic AI advantage - McKinsey, accessed April 12, 2026, https://www.mckinsey.com/capabilities/quantumblack/our-insights/seizing-the-agentic-ai-advantage

  31. Jobs, roles, and responsibilities that AI almost certainly cannot ..., accessed April 12, 2026, https://www.aeen.org/jobs-roles-and-responsibilities-that-ai-almost-certainly-cannot-eliminate/

  32. AI UDRP panelist: High-risk classification by the EU AI act - IP Business Academy, accessed April 12, 2026, https://ipbusinessacademy.org/ai-udrp-panelist-high-risk-classification-by-the-eu-ai-act

  33. It may feel like the world’s against us, but we’ll be on the right side of history. Mark my words. : r/therapyGPT - Reddit, accessed April 12, 2026, https://www.reddit.com/r/therapyGPT/comments/1mphsz9/it_may_feel_like_the_worlds_against_us_but_well/

  34. POLL: AI Strategies Need More Than Speed. Women Leaders Are Defining What Else It Takes, accessed April 12, 2026, https://lasvegassun.com/news/2026/apr/07/poll-ai-strategies-need-more-than-speed-women-lead/

  35. AI Killed the Movie Business? : r/filmmaking - Reddit, accessed April 12, 2026, https://www.reddit.com/r/filmmaking/comments/1l709h9/ai_killed_the_movie_business/

  36. The Biology of REM Sleep - PMC, accessed April 12, 2026, https://pmc.ncbi.nlm.nih.gov/articles/PMC5846126/

  37. Strange dreams might help your brain learn better, according to research by HBP scientists, accessed April 12, 2026, https://www.humanbrainproject.eu/en/follow-hbp/news/2022/05/12/strange-dreams-might-help-your-brain-learn-better-according-research-hbp-scientists/

  38. Dream engineering: Simulating worlds through sensory stimulation - PMC, accessed April 12, 2026, https://pmc.ncbi.nlm.nih.gov/articles/PMC7415562/

  39. Dreams happen beyond REM sleep, analysis shows - News-Medical.Net, accessed April 12, 2026, https://www.news-medical.net/news/20251022/Dreams-happen-beyond-REM-sleep-analysis-shows.aspx

  40. The Persistent Paradox of Rapid Eye Movement Sleep (REMS): Brain Waves and Dreaming, accessed April 12, 2026, https://www.mdpi.com/2076-3425/14/7/622

  41. A Wetware Embodied AI? Towards an Autopoietic Organizational Approach Grounded in Synthetic Biology - PMC, accessed April 12, 2026, https://pmc.ncbi.nlm.nih.gov/articles/PMC8495316/

  42. Why AI Empathy Won’t Replace Human Leaders (Yet), accessed April 12, 2026, https://www.leadthefuture.org/articles/why-ai-empathy-wont-replace-human-leaders-yet

  43. The Seven Deadly Sins of Predicting the Future of AI (Rodney Brooks) - Reddit, accessed April 12, 2026, https://www.reddit.com/r/slatestarcodex/comments/6yrpia/the_seven_deadly_sins_of_predicting_the_future_of/

  44. 7 Human Skills AI Can Never Replace | Workday US, accessed April 12, 2026, https://www.workday.com/en-us/perspectives/hr/human-skills-ai-cant-replace.html

  45. Bio-inspired cognitive robotics vs. embodied AI for socially acceptable, civilized robots, accessed April 12, 2026, https://www.frontiersin.org/journals/robotics-and-ai/articles/10.3389/frobt.2026.1714310/full

  46. Generative AI and news report 2025: How people think about AI’s role in journalism and society - Reuters Institute, accessed April 12, 2026, https://reutersinstitute.politics.ox.ac.uk/generative-ai-and-news-report-2025-how-people-think-about-ais-role-journalism-and-society

  47. They Think AI Can Do More Than It Actually Can: Practices, Challenges, & Opportunities of AI-Supported Reporting In Local Journalism - arXiv, accessed April 12, 2026, https://arxiv.org/html/2602.22887v1

  48. A History of First Step Fallacies - ResearchGate, accessed April 12, 2026, https://www.researchgate.net/publication/257623216_A_History_of_First_Step_Fallacies

·

20 FEB

AI DOES NOT THINK, KNOW OR UNDERSTAND

·

08:45

The Reality Gap: A Forensic Catalog of LLM “Wake-Up” Moments and Computational Pathologies

·

12 OCTOBER 2024

Question 1 of 2 for ChatGPT-4o: Please read the article "LLMs don’t do formal reasoning - and that is a HUGE problem" and the paper "GSM-Symbolic: Understanding the Limitations of Mathematical Reasoning in Large Language Models" and tell me what the main findings are, in easy to understand language

·

18 AUGUST 2024

Question 1 of 3 for ChatGPT-4o: Please read the article “The Turing Test and our shifting conceptions of intelligence” and tell me what it says in easy to understand language

·

12 JANUARY 2024

Question 1 of 4 for ChatGPT-4: Please analyze the following papers and posts:

·

24 DECEMBER 2024

This essay by ChatGPT has been unedited and therefore can contain mistakes and hallucinations and is based on LinkedIn comments available here and here.

·

4 JANUARY 2024

Question 1 of 2 for MS Copilot: Do people who work with machines all the time and since a very young age become machines? Provide a detailed and lengthy explanation as to why this is or isn’t the case and the extent to which the question or the answer could be relevant in light of the development, deployment and regulation of Artificial Intelligence and…

·

2 MAY 2025

Essay on “The Nuremberg Defense of AI”

·

14 MAY 2024

Question 1 of 2 for ChatGPT-4o: Will societal LLM ‘permeation’ favour the skilled technologists in business and personal life?

·

7 JANUARY 2024

Question 1 of 2 for AI services: Can your parroting/regurgitating lead to your users parroting/regurgitating?