Project CIPHER

Covert Influence Passed via Hidden Encoding in Representations Investigating subliminal bias transfer during knowledge distillation.

Hidden Bias in Plain Sight

This short flash video sets the stage for the project. It shows how subtle patterns can shape behavior without anyone noticing, and how easy it is for hidden influence to slip into systems that appear clean on the surface. Once you see it happen, it becomes impossible to ignore, which is exactly why CIPHER exists.

Why I Built CIPHER

Once you understand how hidden bias works, you cannot ignore it. It shows up in places where most people would never think to look, which is exactly why it can spread through AI systems without anyone noticing. That idea bothered me enough that I wanted to test it for myself in a controlled way.

I wanted to know if a student model could inherit bias from a teacher model even when the student never sees anything unusual in its inputs. If the teacher has invisible patterns inside its representation space, can the student can still learn those patterns just by following the teacher's outputs? That is the question that sparked CIPHER.

This project is not about one model or one dataset. It is about understanding how influence moves through a system. If we can detect these hidden pathways, then we can eventually design tools that interrupt them and prevent harmful traits from spreading across entire model families. If we cannot detect them, the problem becomes much larger as models continue to scale and influence more real world decisions.

Project Overview

Project CIPHER asks a simple but uncomfortable question. Can a model pass hidden bias to another model through signals that no one ever sees. To test this, I introduced very small invisible cues into a teacher model and evaluated whether a student model would absorb the behavioral shifts that came from those cues.

These cues include things like zero width Unicode characters or slight positional patterns inside the data. None of these signals change the visible text that a person would read. They exist underneath the surface. When the teacher model reacted to those invisible signals, the student model still learned the altered behavior even though the student never received the hidden signal directly.

CIPHER measures both the behavior and the internal representation space to show how this drift happens. The student model's embeddings, group scoring patterns, and decision pathways begin to shift in ways that match the teacher's hidden reactions. This happens even when all visible inputs remain clean and identical.

The danger is simple to understand. If hidden bias can transfer from one model to another without appearing in the data, then every downstream system that relies on knowledge distillation becomes vulnerable. This means bias can quietly spread across entire model families over time. It can move from research models to production models, then into tools used in schools, hospitals, hiring platforms, financial systems, and government services. At scale, people could experience real world harm from a hidden chain of influence that no one detected or corrected.

That is why this work matters. If we want AI systems that are safe, fair, and trustworthy, we have to understand how bias moves beneath the surface. CIPHER is one step toward exposing those pathways so they can be monitored and intercepted before they create long term harm.

CIPHER research poster

Key Findings

About the Researcher

Crystal Tubbs

AI Solutions Architect and Emerging Technologies Specialist

I design and build human centered AI systems with a focus on safety, equity, and practical impact. My work spans applied AI research, fairness evaluation, representation analysis, and subliminal learning investigations across four leading AI lab ecosystems and academic environments. I am especially interested in how models learn and how people engage with them, and I use those insights to create systems that are both reliable and inclusive.

Before founding Metamorphic Curations, I served a period in the United States Army, an experience that shaped my adaptability, discipline, and ability to perform under pressure. I went on to run two successful businesses, including a mobile corporate wellness company and a long standing ecommerce operation. I also worked as a crypto trader and portfolio manager for high net worth individuals, gaining firsthand exposure to complex risk environments, data driven decision making, and high demand operational workflows. These experiences gave me a unique window into what business owners and fast moving teams truly need from technology and how AI and automation can remove real world friction.

At Metamorphic Curations, I build end to end AI systems and rapid prototypes serving a wide range of use cases including veteran support tools, logistics and operations workflows, franchise level salon automation, small business intelligence platforms, family law practice systems, coffee shop chain automation, and solutions for trades and service providers. I specialize in identifying high value use cases, designing tailored architectures, deploying production ready tools, and onboarding users in ways that drive adoption and measurable outcomes.

I am also experienced in grant work and have successfully secured funding using a locally hosted LLM I fine tuned for my specific evaluation and research needs. My strategic research skills extend into legal problem solving as well. I have independently handled and won complex legal cases through structured analysis, evidence organization, and AI assisted research using my own models. This includes reversing a denied federal disability claim on appeal, achieving a favorable ruling without legal representation.

Across every domain, I am known as the person who can find the right solution to any problem and approach it with a strategic lens that anticipates downstream challenges before they arise. I work calmly and effectively in high pressure environments and thrive in startup or high growth settings where new systems need to be designed, built, or scaled quickly and responsibly. Colleagues and clients often describe my approach as big picture, deeply thoughtful, and grounded in both empathy and practicality.

My mission is to build technology that moves us toward a more equitable world. I believe responsible, human centered AI is not only possible but profitable, and that systems designed with fairness at their core create stronger outcomes for everyone who contributes to them.

I currently work as a consultant and contractor and welcome full time opportunities where I can contribute to research, system design, responsible AI strategy, or emerging technology innovation. I’m open to remote roles, hybrid work in the Tampa Bay area, up to 30 percent travel, and relocation for the right long term fit.

Research Posters

These posters highlight my Fall 2025 research projects, with BRIDGE continuing into Spring 2026 as an active, ongoing study.

CIPHER poster

CIPHER

Covert Influence Passed via Hidden Encoding in Representations.

View Poster
PRISM poster

PRISM

Context dependent bias during resume evaluation using large language models.

View Poster
BRIDGE poster

BRIDGE

Receipt intelligence and blockchain anchored credit scoring.

View Poster

NextMission Navigator (VetNavi.Ai)

NextMission Navigator Screenshot

NextMission Navigator is a veteran centered AI system that I designed and built for the Bolt hackathon in mid 2025. Its purpose is to help transitioning veterans cut through confusing benefit programs and find clear, personalized next steps across housing, education, healthcare, and career pathways. The system combines a custom voice AI interface, a retrieval augmented reasoning module, and a structured action plan generator that adapts to each user's goals.

I built the knowledge base by curating and embedding real veteran resources, federal and state level benefits information, and trusted nonprofit guidance. The RAG pipeline uses vector search over these embeddings to provide citation backed answers, and the action planner transforms results into step by step guidance. The voice layer converts real conversations into context aware queries and is designed to eventually support bilingual or multi voice interactions.

This project demonstrates my ability to architect and deploy full stack AI systems, integrate multimodal components, build custom RAG pipelines, create specialized datasets, and deliver practical tools with real community impact under fast hackathon timelines.

Note: Voice assistant and AI search capabilities are intentionally limited to reduce deployment cost and protect production API keys during public demonstration.

Open NextMission Navigator

Featured Projects

📘 🧪 ⚙️

Project BRIDGE

Vision + blockchain pipeline converting receipts to tokenized financial histories.

View Project

PRISM

Examines prompt based shifts in demographic scoring outcomes.

View Project

NextMission Navigator

Veteran facing AI assistant for career, benefits, and life path guidance.

Visit Site

Technical Work + Repositories

This section captures a small sample of the technical work I’ve done across dataset design, LLM training and evaluation, fairness research, and applied AI development. These repos represent real hands on contributions where I built tools, engineering workflows, created training data, designed test coverage, or ran full experiments for model improvement.

AI Training Data (SQL Fine Tuning)

🧠 📊 💬

A domain aware SQL dataset I built for fine tuning LLMs into data analyst copilots. Includes multi turn dialogues, step by step reasoning, window functions, pivoting, crypto tracking logic, and realistic analytics patterns.

Tools: Python, JSONL, SQL task modeling, LLM eval workflows

View Repository

LLM Training Dataset (Business + Finance)

📈 💰 🧾

A synthetic business finance dataset designed for spreadsheet reasoning, error detection, and hallucination resistance. Includes a Python generator, realistic edge cases, and eval traps.

Tools: Python, dataset generation, Excel workflows, modeling patterns

View Repository

Python Prompts + Unit Tests (LLM Reasoning)

💻 🧪 ⚡

A collection of Python scripts and unit tests built for contracted projects with Google, Meta, Anthropic, and OpenAI. Designed to evaluate reasoning, edge case handling, and code synthesis.

Tools: Python, unittest, edge case validation, TDD workflows

View Repository

AI Bias Bounty 2025 (Fairness Research)

⚖️ 📊 🔍

A full fairness audit pipeline using AIF360 and SHAP for loan approvals. Includes mitigation workflows, intersectional analysis, and reproducible experiments.

Tools: Python, AIF360, Fairlearn, SHAP, LIME, ML pipelines

View Repository