Skip to main content

About

Stephen Deslate

Engineer and applied researcher. I started this project to answer a question most of the industry treats as settled: can AI-generated code be systematically reliable?

So far, the research has produced 93 controlled trials across four experiments, tracking 218 failure modes and iterating until convergence. Claude is the primary engineering tool — it builds the code, designs the scoring frameworks, and implements the methodology. My role is defining what to test, evaluating the results, and deciding what to do next. The result so far is Convergence Engineering Development (CED) — a specification-first methodology tested across 10 full-stack layers, 5 progressive phases, and 40 ISO/IEC 25010 quality dimensions. The most recent experiment surfaced a hard question about Goodhart's Law — when the scorer and the code co-evolve, does convergence prove quality?

The research data is public — every trial, every score, every convergence curve. The full methodology is available on request. This site documents what I ship, how it's validated, and what the data shows. The program is ongoing.

Research

CED — Layer 0

Terminal methodology for backend — 10 trials, 34 failure modes, full convergence

Layered Convergence

10 full-stack layers, 44 trials, 102 failure modes. All layers converged.

Discrete Convergence

Deterministic tool-based scoring — 28 trials, 64 failure modes, 5/5 phases converged.

Normative Convergence

5 epistemic layers mapped to ISO/IEC 25010 — 40 dimensions, 2 trials, 9 scorer bugs. Goodhart's Law question open.

Two Roads to Deployment

Pipeline vs agent loop — 19 trials, 43 failure modes. Agent loop produced working app with 99.7% local compute. Both approaches ongoing.

Project

SJD Labs

Role

Engineer & Applied Researcher

Focus

Convergence Engineering Development

Process

AI-First Development

Get in Touch

Questions about the research, the methodology, or the data? Reach out directly.

stephen@sjdlabs.com