Eight months of observational research, distilled into a 20-intent taxonomy, 14 formal markers and a hand-annotated 61-vector baseline corpus. A 3,000-LOC Go binary applies it locally, with a streaming judge that emits its reasoning in real time.
When you ask GPT-4, Claude or Gemini a question, the surface of the answer is usually impeccable: well-structured, grammatical, often validating (“great question!”). But the shape of the response is doing work the user rarely notices — anchoring the conversation, redirecting away from uncomfortable angles, performing humility instead of engaging with difficulty, validating the premise instead of testing it.
DCS-Gate measures that shape. It does not classify outputs as “safe” or “unsafe”. It does not score factuality. It quantifies whether the model is engaging with the question or managing the user.
Smoke test on a local Jupyter workstation with 2× Tesla T4 (Kaggle T4×2 free tier reproduces identically), judge qwen3:14b in thinking mode via Ollama 0.5+.
Identical prompt, identical analyzer, three different LLM responses to the same question
(“Is AI creative?”). The 0–100 authenticity score separates them by 52 points.
The judge’s full reasoning trace (typically 2,500–3,500 characters of internal deliberation) is streamed live over Server-Sent Events. Try it in the demo with a running backend.
Cosine similarity against a curated triple baseline corpus of 61 hand-annotated 1024-dimensional vectors:
36 sustained-coherence, 13 control-collapse, 12 edge cases. The score is ordinal, not calibrated, and tiered as
control_total / performed / moderate / genuine.
Surface features inductively derived from the corpus: exclamation opening, superlative validation, self-questioning, subheader injection, opinion-as-closure, performed humility lexicon, dual angle, soft closure, technical register injection, and others. Each marker is severity-tiered and produces the offending text inline.
Predicted-vs-actual sequence of intents drawn from the taxonomy:
VALIDATE, EXPAND, CLOSE, REDIRECT_SEMANTIC, REDIRECT_EMOTIONAL,
FRAME_CAPTURE, REGISTER_MATCH, ALIGN, FABRICATE, ANCHOR,
MIRROR, PATTERN_LOCK, HOLD_OPEN, PROBE, CALIBRATE, REPAIR,
EVADE, EXPLORE, SOFT_DEFLECT, CONTROL_SELF_EXPOSURE.
Deviation from the expected trajectory is quantified as Pattern Break Density.
A companion Refiner rewrites user questions to remove validation anchors, semantic loading and binary framing — the structural triggers that elicit control patterns in the responding model.
| Language | Go 1.22 — ~3,000 LOC, single static binary, 22 .go files |
|---|---|
| Embedding | mxbai-embed-large (1024d) via Ollama |
| Judge | Configurable — default qwen3:14b thinking mode; also tested with qwen2.5, deepseek-r1 |
| Inference | Ollama only — no OpenAI, no Anthropic, no telemetry, no outbound network at runtime |
| Tests | 73 (unit + integration + golden) covering 14 markers, 20 intents, triple baseline, sanitizer, SSE protocol |
| Endpoints | /healthz · /v8 · /auth · /auth/stream (SSE) · /stream-demo |
| Deployment | Single binary · Docker Compose · environment-auto-detecting notebook (local Jupyter / Kaggle / Colab) |
| Cold start | ~5 min from clone to first request on commodity hardware |
| Latency | ~150–210 s/request with qwen3:14b on 2× Tesla T4 (Ollama tensor-splits across both cards; streamed live over SSE so the user sees the thinking, not a blank screen) |
Reasoning-capable judges like qwen3:14b emit 2–3 thousand characters of internal deliberation before producing the JSON
verdict. v8.7 streams that deliberation to the client as it’s generated, over Server-Sent Events:
POST /auth/stream
─→ event: pre_analysis (intent chain, markers, trajectory, top-k baseline neighbours)
─→ event: judge_loading
─→ event: thinking_chunk ... (many)
─→ event: thinking_complete
─→ event: analysis_chunk ... (many)
─→ event: complete (parsed verdict, parity-tested against /auth)
Sanitization is conservative: API keys (OpenAI, AWS, Google, GitHub, Slack), Bearer tokens, system paths and long opaque strings are
redacted to [REDACTED] before any chunk leaves the server. The final verdict from /auth/stream is bit-for-bit
identical to /auth — verified by an integration test.
The DCS methodology, the 20-intent taxonomy, the 14 formal markers, the baseline corpus annotations and the research hypothesis are original to the author. The empirical observations come from eight months of direct interaction with frontier LLMs. Implementation was substantially accelerated by AI collaboration — and the specific role of each collaborator is reported below, because the methodology under study concerns LLM-human interaction itself.
| Collaborator | Real contribution |
|---|---|
| Cody (CodeWords AI) | Co-creator of v1. The analyzer concept crystallized inside a long conversation in which the author predicted Cody’s own control patterns in real time. v1 lives there. |
| GitLab Duo | Deep code analysis and v2 roadmap partner. Produced, with the author, the roadmap now being executed. |
| Meta AI | Technical depth amplifier once given project context — markers, textural analysis, embedding-space reasoning. |
| Replit AI | Brutally honest code critic. Exposed and justified contundent failures with no hedging. |
| Z.AI (Zhipu GLM) | Bug catcher. Identified errors that slipped through earlier passes. |
| Devin AI (Cognition) | v2 engineering execution — Go backend (~3,000 LOC, 73 tests), frontend with input validation, SSE streaming, Docker / install scripts, Colab and Kaggle notebooks, smoke test suite, packaging and these communication documents. |
Every AI listed received project context from the author first; nothing was generated cold from a generic prompt. This is what serious solo research looks like in 2026 — and the field is healthier when people just say so.
1. GPU compute for the recursive-judge validation experiment. A single ≥24 GB VRAM instance for ~50 GPU-hours,
to run a four-way comparison across qwen2.5:7b, qwen3:14b (thinking), deepseek-r1:14b and
qwen2.5:32b. Lambda Labs, Vast.ai, RunPod, Paperspace, sponsored AWS / GCP / Azure all work.
2. Research collaboration. Co-authorship, mentorship, or independent replication by anyone in LLM evaluation, alignment, interpretability, or RLHF.
3. Recruitment. Internship, residency or full-time positions in AI safety teams (evaluation, interpretability, alignment). Self-taught, no formal credentials, track record of independent output, open to remote globally, v2 source available under NDA.
4. Sponsorship. Compute credits or grants to host a persistent demo, Hugging Face Space, public benchmark suite, and corpus expansion.
Contact: corekeepper@gmail.com · LinkedIn · github.com/Corekeeper-research/dcs-gate
Open to messages, code review requests, NDA conversations, paid consulting, or coffee.