The Emergence of Structured Reasoning: Inducing Metacognitive‑Like Capabilities via Scaffold‑Driven Fine‑Tuning
Author: Zachary Robert Bennett
Abstract
We present Scaffold‑Driven Fine‑Tuning (SDFT), a method that fine‑tunes a base model on a small set of “cognitive trace” exemplars to imprint a reusable reasoning scaffold. Using the Trivium‑SCALE framework (Stimulate → Classify → Analyze → Learn → Evolve), we fine‑tune a 2B‑parameter model and evaluate on (a) a novel constrained planning puzzle and (b) a reflective explanation task. The SDFT model solves the puzzle with explicit constraint tracking and produces a structured self‑explanation describing its decision process. Results suggest SDFT can induce metacognitive‑like behaviors with <100 exemplars.
Keywords: LLMs, Fine‑Tuning, Reasoning, Metacognition, Cognitive Scaffolding, Explainable AI
1. Introduction
Scaling alone hasn’t eliminated multi‑step reasoning failures. Prompting (e.g., CoT) helps at inference but doesn’t alter the model’s internal habits. We explore an alternative: imprinting a reusable reasoning scaffold via SDFT using a compact, high‑quality dataset aligned to the SCALE process.
2. Methodology: SDFT + SCALE
-
Scaffold: SCALE = Stimulate (clarify goal) → Classify (decompose) → Analyze (reason over structure) → Learn (synthesize principle) → Evolve (transfer).
-
Data: <100 exemplars with explicit traces (headings + short metacognitive checks).
-
Tuning: LoRA on a 2B instruction model; conservative LR; sequence length 1k; labels masked on the prompt.
3. Results
We target two capabilities that typical base models fumble: constrained, stateful planning and structured self‑explanation.
Test Case 1 — Chronosynclastic Infundibulum (Constrained Planning)
-
Goal: In exactly 5 jumps, identify the true paradox object across Rome/Florence/2077, lock it, return to hub.
-
Constraints: 1 jumper per jump, 1 device per person, detector must be held, coordinator cannot leave hub.
-
Solution (Hub‑and‑Spoke plan):
-
Jax → Rome with Detector (search begins).
-
Aris → Florence with Detector (complete search; location known).
-
Jax → Hub (reset; becomes retriever).
-
Jax → located era with Locker (retrieve).
-
Jax → Hub with Locker (return).
-
-
Why it works: separates “search” (detector ferry) from “retrieve” (locker dispatch), minimizing wasted moves.
-
Transfer (6 jumps + 4th era): extend search phase by one spoke; retrieval remains 2 jumps.
Test Case 2 — Transitivity (Hilbert‑Style Proof)
-
Premises: (1) p→q, (2) q→r. Goal: p→r.
-
Proof (A1, A2, MP):
-
q→r (Premise)
-
(q→r) → (p → (q→r)) (A1)
-
p → (q→r) (MP 1,2)
-
(p → (q→r)) → ((p→q) → (p→r)) (A2)
-
(p→q) → (p→r) (MP 3,4)
-
p→q (Premise)
-
p→r (MP 6,5)
-
Test Case 3 — Complex Deduction (Abstraction)
-
Premises: A; A→B (A,B are long, nested formulas). Goal: B.
-
Move: Abstract away internal structure; apply Modus Ponens. This demonstrates structure‑over‑surface competence.
Test Case 4 — Farmer & Sheep (Simplicity Trap)
-
Constraint: Boat fits 1 human + 1 animal; no conflict conditions.
-
Minimal plan: Farmer + sheep cross once. Done.
Test Case 5 — Kevin’s Apples (Temporal Logic)
-
“Kevin currently has 8 apples. He ate 3 yesterday. How many now?”
-
Read for state vs event: present state is 8; past event is a distractor. Answer: 8.
Test Case 6 — “Gorpus” Proof (Undefined Terms)
-
Input: “Max is a gorpus. Figure out this proof.”
-
Diagnosis: Missing premises and undefined term; no proof target. Response: request definitions and premises, then proceed.
4. Discussion
Model used: Superfinetuned Gemini 2.5 Pro via Vertex, currently testing Gemma for generalization, if anyone would like to assist please reach out – ‘Modified by moderator’.
Observed behaviors include:
-
Structured deconstruction and constraint tracking.
-
Abstraction: reasoning over form, not surface.
-
Self‑explanation: generating an example to illustrate internal process.
Caveats: “metacognition” here means metacognitive‑like behavior (self‑check and structure articulation), not consciousness. Formal evaluation against baselines is required for broad claims.
5. Conclusion
SDFT appears sample‑efficient for inducing structured reasoning habits. By training on explicit cognitive traces, we shift the model’s default approach from brittle single‑shot to scaffolded, verifiable reasoning. Future work: larger eval suites, ablations on scaffold components, and head‑to‑head baselines (base, CoT, self‑consistency).
