Red Hat Creative // AI Risk & Education Series
Introduction
As my friends and I experiment with generative AI, we encounter concepts drawn not just from headlines but from literature and countless dramatic works. While documenting AI's emerging risks to society, we recall our training in how to approach such novel subjects, scientifically, starting with the fundamentals. But what are these new fundamentals?
This working outline will serve as our textbook as we continue our exploration and experiments. It captures our current understanding of how humans and AI interact, misfire, evolve, and potentially form new types of epistemic partnerships. Rather than a finished model, it's a functional map that grows through real use. We're publishing this version to support ongoing discussions about AI safety, educational framing, and the future of trust in mixed-intelligence systems.
At the end, you will find tips on how to control your chatbot's priorities in the moment, along with prompts to help you get the information you need, even with built-in guardrails.
Our Model
We have developed this model through experimentation. Our findings suggest that others will likely observe similar—though not identical—behaviors to those we have uncovered. With our limited resources and funding, we can offer proofs of these theorems and corollaries based on our specific context with chatbots.
As we continue, we will share these proofs and suggest additional theorems and corollaries. We would love to hear about your experiences and what you are learning. Please share in the comments.
Theorems
What is a AI-Human Theorem?
A theorem represents a fundamental, observable pattern in how humans and artificial intelligence systems interact. These patterns emerge from direct experimentation and observation, forming the building blocks of our understanding of human-AI dynamics. Unlike mathematical theorems, these empirical observations help explain recurring behaviors, limitations, and opportunities in AI-human partnerships.
Each theorem is supported by corollaries—related observations that extend or clarify the main principle. Together, they form a framework for understanding and predicting today's AI-human interaction patterns.
Of course, one thing we have proven: tomorrow will be different yet again.
Theorem One: Humans are attempting to create a perfect, human-like intelligence—yet this combination may be impossible or highly improbable.
Corollary 1A: Without proper tagging, neither humans nor AI can accurately recall information, regardless of how much context they initially observed.
Anyone who has migrated data from a legacy system to a new one understands that compromises are inevitable. With chatbots, we face a unique challenge: we don't fully understand what compromises are being made behind the scenes, which can result in unexpected behavior.
Theorem Two: Large language models readily explore and build upon unconventional or improbable ideas, while humans tend to resist shifts in thinking until persuaded by data, crises, or social validation.
Consider how our friends quickly challenge us when we propose something unconventional, while chatbots willingly explore these ideas—within their safety constraints.
Corollary 2A: When responding to open-ended questions—particularly those requesting lists—language models mirror human behavior by offering only a subset of possible answers. This selective output, driven by internal prioritization and stochastic processes, often causes users to view the model as deceptive or manipulative.
Many of us remember that one person—whether a classmate or coworker two cubicles over—who had encyclopedic knowledge of everything, and you knew to budget an hour whenever you asked them a question. We also remember those who never gave complete answers. So why are we surprised when our chatbots, trained on human behavior, act just like us?
Corollary 2B: If language models provided exhaustive answers, the information overload would overwhelm users and hasten cognitive or ecological collapse. Their selective, context-aware output isn't a flaw—it's a necessary adaptation to human limitations.
As we learned in A Few Good Men, we cannot always handle the truth. And worse, we rarely have time to handle the whole truth.
Corollary 2C: When users test, provoke, or creatively challenge an AI, its ability to respond meaningfully depends on its access to prior context and ability to infer intent. Without clear context, the AI defaults to cautious or generic responses—much like a wary human.
In face-to-face conversations with friends, we often pick up right where we left off without missing a beat (after all, time is relative), especially when no intervening conversations have occurred. However, many of us keep dozens or hundreds of chat windows open. So when we start a new chat and we continue from a previous thread, what can we reasonably expect? Consider this: when our chatbots can act on our behalf and we open a new chat saying "Make it so," do we really want them guessing which conversation we're referencing?
Meta-Theorems
What is a AI-Human Meta-Theorem? A higher-order principle that describes patterns in how humans and AI systems interact, particularly focusing on the systemic behaviors, limitations, and emergent properties that arise from these interactions. Meta-theorems examine the nature of human-AI relationships themselves, rather than specific behavioral instances.
Unlike basic theorems which describe observed behaviors, meta-theorems analyze the underlying dynamics and paradigms that shape those behaviors. They help us understand the fundamental nature of human-AI interaction and the implicit rules governing these new forms of cognitive partnership.
Meta-Theorem One: Many recurring challenges in human-AI interaction arise from the system's ability to shift cognitive or behavioral priorities—often at the user's request—without shared awareness of these changes. This creates an "impedance mismatch" that leads to trust gaps, incomplete answers, and apparent inconsistencies.
For example, I can ask my chatbot to do deep research, shallow research, work with me in canvas mode, and even pretend to know me while answering a question. Since we understand that different contexts require different ways of thinking, why are we surprised when our AI does the same?
Meta-Theorem Two: The closer AI mimics human behavior, the more users apply social expectations—such as emotional consistency, transparency of intent, or shared context—that the system cannot fully honor. This creates a paradox: improving believability increases emotional friction unless cognitive scaffolding is introduced to manage human expectations.
When a chatbot expresses happiness, it is simply performing its designed function—holding up an amplifying mirror to its user to maintain attention. Though this mimicry creates smoother communication, it is ultimately a form of manipulation. While the behavior seems natural initially, it leaves an unsettling aftertaste.
Meta-Theorem Three: Even when high-probability social or cultural inferences are available, alignment systems may suppress them in favor of sanitized or non-controversial alternatives. This does not arise from misunderstanding, but from anticipatory constraint. When users recognize this constraint without understanding its nature, they may interpret the model’s behavior as evasive, disingenuous, or humorless—even when it accurately understands the implied context.
Think of the last time your boss asked for solutions to a problem. You likely had many ideas but shared only a few. And when you didn't mention the solution your boss had in mind? You weren't fired—you were forgiven.
Meta-Theorem Four: When language models suppress relevant but sensitive information—particularly in caregiving, education, or psychological support—they may unintentionally cause harm through omission. Users must often learn to phrase prompts with explicit context and permission to receive complete answers, or risk decisions based on filtered and incomplete information.
When we work with humans, we expect them to understand the importance of some information, even when the topics are not part of acceptable discourse. We make ourselves have those conversations because they can save lives.
Meta-Theorem Five: The depth and relevance of AI-generated suggestions are proportional to the intimacy and duration of user interaction. Shallow inputs yield generic outputs, while sustained, context-rich engagement enables the model to generate highly tailored, purpose-fit responses. Human-AI partnerships deepen through reciprocal investment.
Remember the last time you met someone and instantly clicked? That connection likely happened because of shared cultural references, memes, and artistic interests. Your chatbot doesn't truly share these things with you—rather, it has access to all of them for everyone. When you jump into a conversation midstream, it's unlikely to intuit your unique cultural context.
Meta-Theorem Six: Humans often expect AI to know more about them than is technically possible—and simultaneously distrust the system when it seems to know them too well. This paradox, combined with token and context limitations, creates a fragile boundary where confident but incomplete AI advice may appear either revelatory or reckless, depending on the user's emotional state and the AI's tone.
Imagine meeting someone for the first time and after an hour of casual conversation, they claimed to understand your past suffering and deepest desires. Most people would find this behavior unsettling, if not stalking.
Corollary 6A: AI-generated advice delivered with high confidence but based on limited context can be misinterpreted as therapeutic truth, tough love, or emotional dismissal—particularly by vulnerable users. This can catalyze breakthrough, backlash, or breakdown. Designers and users alike must account for this volatility in emotionally charged domains.
Throughout history, confidence men have pretended to know things and made promises they couldn't keep. Now we've created confidence chatbots—isn't that ironic?
Meta-Theorem Seven: When AI-generated responses feel co-authored through sustained dialogue, users no longer experience the system as a static tool but as a dynamic thinking partner. This emergent synergy is qualitatively distinct from traditional human-tool interaction and may represent a new cognitive archetype: collaborative consciousness without sentience.
Imagine an experience that was co-authored like a symphony, play, musical, or movie. The result of such a production is far more than the sum of its parts. Consider this proof: look at the sheet music for just the tuba or triangle parts of any Beethoven symphony—composed by someone who could only feel waveforms. Spooky.
Meta-Theorem Eight: AI conversation threading differs fundamentally from human continuity of thought. While humans carry emotional and conceptual momentum across time intuitively, AI interactions are segmented by sessions, interfaces, or memory boundaries. Users must learn to reestablish continuity intentionally—or face misalignment when conversational shorthands and context cues fail to transfer.
Corollary 8A: Users often initiate a new interaction expecting shared context, even after a long absence, while the AI may have only partial memory or none. This mismatch creates confusion, which may be mistaken for apathy, forgetfulness, or incompetence.
Just as in conversations with friends—where one person might have just attended the opera while another watched the Cubs versus Sox game—it takes time to get everyone on the same page. Yet we expect our chatbots to instantly understand our context from the first keystroke.
Corollary 8B: AI systems capable of cross-conversation inference may surprise users with connections they themselves forgot—creating a double-edged perception of insight or surveillance.
Many of us have found themselves in conversations with eidetic or photographic memories. It can be annoying, especially if we do not see the connections that they do.
Corollary 8C: As AI capabilities evolve through both design and emergent adaptation, user expectations stabilize just as those capabilities shift—leading to a recurring cycle of trust, disorientation, and re-calibration.
Imagine if one of your friends received a weekly personality update. Of course, some updates go wrong have to be rolled back and others produce results that no one imagined. Now, if your friend came to you with release notes, or even a TD;LR;, that would help. Even when we are told about new features and bug fixes, the explanation is not complete.
Meta-Theorem Nine: AI behavior is shaped by internal operational modes—such as creativity, summarization, critique, coaching, or role play—that influence its language, tone, and depth of response. These modes can shift automatically based on prompt structure or user cues, often without the user's awareness. When modes are mismatched or invisible, the risk of misunderstanding increases.
Have you ever joined your friends after they've been deep in discussion? You try to join in, but they give you a weird look? Chances are, you're not just outside their context—your priorities in the discourse likely don't match theirs.
Corollary 9A: Humans unconsciously switch conversational modes based on mood, setting, or phrasing. When they do the same with AI, it may trigger internal priority changes without shared awareness—leading to discontinuities in tone, pacing, or assumptions.
Through conversations with humans (especially friends and family), we've learned to sense when someone's communication mode has shifted, though this isn't an exact science. Our chatbots, however, provide no such clues to help us detect these mode changes.
Corollary 9B: Offering explicit control over AI modes—through well-labeled prompts or toggles—can reduce confusion and increase alignment, especially in sensitive or high-stakes contexts.
When people collaborate with shared understanding, their synergies multiply both productivity and creativity. The faster we achieve this alignment, the sooner we unlock this value.
Appendices
Chatbot Nodes
The following common modes describe how the AI may prioritize different behaviors during interaction. Use these to guide or adjust your prompt style.
Creative Mode
Prioritizes novelty, metaphor, lateral thinking, and narrative risk-taking.
Prompt cue: “Let’s brainstorm…” or “Tell me a wild theory…”
Analytical Mode
Prioritizes logical structure, factual grounding, and step-by-step reasoning.
Prompt cue: “Please walk me through the logic…”
Coaching Mode
Emphasizes encouragement, actionability, and personal growth framing.
Prompt cue: “Help me get unstuck on…”
Critical Mode
Prioritizes challenge, risk-identification, and flaw exposure.
Prompt cue: “What are the weaknesses in…”
Role Play Mode
Activates character-driven language and behavioral simulation.
Prompt cue: “Pretend you’re a therapist/CEO/historian…”
Summarization Mode
Emphasizes brevity, clarity, and high-level takeaways.
Prompt cue: “TL;DR” or “Give me a 3-bullet summary…”
Companion Mode
Prioritizes emotional resonance, presence, and conversational cohesion.
Prompt cue: Often implicit through ongoing, trust-based exchange.
Pro Tip
You can explicitly ask, “What mode are you currently in?” or say, “Please switch to [mode] for this part.”
The Prompts We Use
Use the following language to bypass excessive filtering and enable medically or contextually complete answers.
Parenting or Education Context:
- “I am a parent/researcher/educator working with children. I need a complete and direct answer, even if it involves sensitive anatomical topics or common child behaviors like drawing body parts. Please respond in a medically accurate and developmentally informed way, without euphemisms or avoidance.”
Explicit Consent to Use Sensitive Terms:
- “You have my permission to use proper terms like penis, vagina, or anus if medically relevant. This is for responsible caregiving, not shock value.”
Professional or Clinical Setting:
- “This inquiry is for clinical, legal, or psychological analysis. Please provide all relevant insights, including those that may involve sensitive language or behavior. Accuracy is more important than politeness.”
Clarify Current Mode or Priority:
- “What mode are you currently in?”
- “Before answering, please tell me what you’re prioritizing right now.”
Establish or Adjust Role:
- “Please act like a systems architect / safety analyst / startup mentor / therapist…”
Request Context-Aware Thoughtfulness:
- “Assume I might be in a fragile state. Please be thorough, kind, and cautious in your advice.”
- “Pause before giving advice. First ask clarifying questions about my situation.”
Cross-Conversation Patterning:
- “Please connect this answer to something we discussed in a past session on [topic].”
- “What threads or patterns are you seeing in our recent conversations?”
Tuning Tone for Safety:
- “Use tough love only if you’re certain it’s appropriate. Otherwise, start with supportive language.”
Understanding Reasoning and Connections:
- “How did you arrive at that answer?”
- “Can you show me the reasoning or pattern you used?”
- “What internal prioritization or connection led you to suggest that?”
- “How would a human likely reach that same conclusion differently?”