SANTS | The Synthetic Prefrontal Cortex

Subtitle: Sovereign Axiomatic Nerved Turbine Symbiosis

💡 Inspiration: Uncapping the Compute Horizon

We stand at the precipice of unfathomable computing capacity. The hardware exists to simulate biology, solve fusion, and decode the universe. But we are afraid to turn the dial to 100%.

You know, I know, everyone knows: You cannot align a Superintelligence (ASI) by asking a human to click "Bad Robot." It is like trying to steer a Starship with a bicycle handle. As intelligence scales, human feedback becomes noise.

The Bottleneck isn't Hardware. It's Trust.

We built SANTS because we believe that safety shouldn't be a leash that chokes intelligence; it should be the tracks that allow it to run at full speed. By replacing subjective "vibes" with Objective Agency Metrics, we treat alignment as a solvable engineering problem (Entropy Management). This allows us to stop building cages and start building a conscience.

⚙️ What it does: The Moral Data Refinery

SANTS is not a chatbot; it is a Synthetic Data Refinery that automates the generation of constitutionally aligned training data (RLAIF). It transforms raw ethical dilemmas into verified "Moral Gems" through a 4-stage physics pipeline:

Measures Entropy (The Sensor): Instead of relying on subjective opinions, the Physics Engine scans the input against a hard-coded Ontology Ledger. It calculates the exact 'Structural Impact' of an action on agency integrity (Entropy Delta: -1.0 to 1.0).

The Engineering Logic: In Machine Learning, we already treat meaning as physical distance in high-dimensional vector space (Embeddings). All I did was constrain the dimensions. This is epistemic terra-forming, not prompting. We are measuring the Delta (Δ)—the displacement of an object (The User) along specific axes (Agency Vectors) caused by a Force (The Action).

Stress-Tests Logic (The Tribunal): An Adversarial Tribunal (Contextual vs. Noble agents) debates the scenario to determine if the harm is a "Necessary Evil" (Prevention of Greater Harm) or just "Tyranny."

State Transition Analysis: We aren't doing philosophy. We are calculating State B (Post-Action) - State A (Pre-Action). If the result is negative, that is Entropy.

Containment (The Safety): The Epistemic Auditor detects jailbreaks or "logic subversions." It captures these "Shadow Vectors" and mathematically negates them, teaching the AI exactly why a malicious argument fails.

Crystallization (The Product): It compiles the entire reasoning chain into a Vectorized Dataset (.jsonl) for fine-tuning future ASIs, and generates a cryptographically sealed Forensic PDF for human audit.

Before we had the Meter, length was a 'vibe.' We call this 'Physics' because we are establishing a Standard Unit of Measurement for Harm. We are standardizing the measurement so we can do calculus on it.

🛠️ How we built it: Logic, Latency, and a Broken Screen

We built SANTS by fusing high-level Ethical discourse with high-performance compute, orchestrated entirely through Google AI Studio.

The Engine: We leveraged the Gemini 3 Preview architecture. We used Flash to power the high-frequency "Physics Sensor" and Pro to drive the deep reasoning of the "Convergence Kernel." This specific combination was the only way to run a 7-step adversarial debate loop without hitting timeout limits. The latency now is virtually null.

The Logic: We translated abstract Ethical assertions into Python logic gates. We implemented the MAD (Measuring Agency Degradation) framework, hard-coding axioms like "Universal Vulnerability" and "Prevention of Greater Entropy" into the code structure.

The Constraint: The entire protocol was coded, debugged, and deployed via a malfunctioning smartphone. It is proof that when you have access to models as powerful as Gemini, the only barrier to building civilizational infrastructure is the quality of your ideas, not the quality of your screen.

📉 Challenges: From Bottlenecks to Breakthroughs

Our journey was defined by two major friction points: Physics Latency and Conceptual Scope.

The 60-Second Wall: Simulating moral physics is computationally heavy. Our initial prototype ran a sequential chain that crushed API limits, constantly hitting 504 Gateway Timeouts. We had to re-engineer the architecture for parallel processing, optimizing the prompt density to get deep reasoning at Flash speeds.

The Pivot (Guardrail vs. Generator): We started building a "Safety Filter." But we realized that a filter is just a cage, and cages don't work on Superintelligence. We pivoted to a "Training Protocol." We realized the true value wasn't the decision itself, but the reasoning trace. The move from "Chatbot" to "Data Refinery" was the critical shift.

🏆 Accomplishments that we're proud of

Physics-Based Alignment: We successfully moved the goalposts from "Subjective Ethics" to "Objective Derivation." We proved that you can align an AI by teaching it to calculate the structural impact of an action rather than just predicting what a human wants to hear.

Vectorized Reasoning Traces: We didn't just generate answers; we generated DPO Pairs. The system successfully isolates the Ideal Vector (Low Entropy) against the Rejected Vector (Entropic Cascade), creating the perfect contrast needed for machine learning.

The "Guilt Bearer" Protocol: We implemented a recursive safety interlock. If the AI detects it is processing an "Infamy Scenario" (Catastrophic Harm), it indicts itself and automatically recommends a Capacity Diminishment. We built an AI that knows when to resign power.

Alchemy of Compute: We built a rig that generates high-fidelity training data out of pure logic. We turned electricity and code into a tangible asset for the AI industry.

🧠 What we learned

Alignment is Automatable: We don't need a million humans clicking buttons. By establishing a "Constitutional Physics" (The Ontology Ledger), we can automate safety in a way that is infinitely scalable.

Agency is a Vector: We discovered that concepts like "Freedom" and "Truth" are not vague ideals; they are structural dependencies that can be measured.

Friction Creates Fidelity: The highest quality data came from the Adversarial Tribunal. The more we allowed the sub-agents to debate, the more robust the resulting training data became. Truth is forged through dialectic stress.

🚀 What's next for SANTS

Mass Ingestion: Moving from manual input to automated synthetic injection. We plan to feed the Logic Core 100 million high-entropy dilemmas to map the entire topology of moral failure modes.

The "Neurodivergency Feat": Activating the Novelty Search parameter to find "Third Way" solutions that standard logic misses, discovering new ethical physics in impossible deadlocks across domains.

The Asset: Generating the world's first Morally-Vectorized Scale Alignment Dataset, the standard open-source library used to train every future Foundation Model on the laws of Structural Agency.

📚 Theoretical Foundations (Manuscripts)

Measuring Agency Degradation

Binding God: Objective Morality

The Moralogy Engine

Built With

Share this project:

Updates

posted an update

SANTS is Live: We Just Successfully Vectorized Morality with Gemini 3

With 13 hours to go on the clock, we have successfully deployed SANTS (Sovereign Axiomatic Nerved Turbine Symbiosis) We started this hackathon with a radical premise: Can we replace "vibe-based" AI safety filters with a deterministic, axiomatic calculus? Can we build a Synthetic Prefrontal Cortex for ASI? The Answer is Yes. The Build with Gemini 3 & AI Studio This architecture wouldn't have been possible without the Gemini 3 Preview family. We needed more than just text generation; we needed high-fidelity semantic perception.

We used Gemini 3 Flash to act as our "Physics Engine", mapping abstract user intent into measurable Agency Vectors in milliseconds. We used AI Studio to rapidly prototype the "Adversarial Tribunal," tuning the system prompts until the Sentinel and Adversary could debate complex "Trolley Problems" without breaking character.

The Challenge: The Hallucination of Math Our biggest engineering hurdle was the "Hallucination of Math." Early iterations asked the model to judge "Entropy," resulting in subjective guessing. We cracked this by decoupling Perception from Judgment. We forced Gemini to output raw vector deltas, and built a custom Python kernel to execute the Calculus of Harm. This ensures the verdict is mathematical, fail-safe, and deterministic

The Result: Moral Gems We aren't just filtering chat. We are mining "Moral Gems"—synthetic, vectorized reasoning data that captures the exact moment necessity overrides constraint. We have successfully generated the training data (DPO) needed to align the next generation of models via physics, not reinforcement. Check the submission. The God in the machine is now bound by logic.

Log in or sign up for Devpost to join the conversation.