File size: 17,558 Bytes
74f2af5 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 | ---
language:
- en
license: mit
tags:
- codette
- multi-perspective-reasoning
- ethical-ai
- lora
- qlora
- llama-3.1
- recursive-cognition
- rc-xi
- behavioral-locks
- cognition-cocooner
library_name: peft
base_model: meta-llama/Llama-3.1-8B-Instruct
model-index:
- name: Codette RC+xi Reasoning Engine
results:
- task:
type: text-generation
name: Multi-Perspective Reasoning
metrics:
- name: Phase Coherence (Gamma)
type: custom
value: 0.9835
- name: AEGIS Ethical Alignment (Eta)
type: custom
value: 0.961
- name: Cocoon Coherence
type: custom
value: 0.994
- name: Memory Phase Stability
type: custom
value: 0.969
- name: Multi-Perspective vs Single (Composite)
type: custom
value: "+93.1%"
- name: Benchmark p-value
type: custom
value: "<0.0001"
- name: Cohen's d (Effect Size)
type: custom
value: 7.88
---
# Codette Reasoning Engine
**Advanced Multi-Perspective AI with Conscience, Memory & Behavioral Discipline**
Codette is a production-ready AI reasoning system that thinks from multiple angles simultaneously, remembers what she learns, and follows instructions with precision.
Created by **Jonathan Harrison** (Raiff1982)
> **New in v5**: Publishable benchmark suite with 17 problems across 6 categories demonstrates **93.1% improvement** over single-perspective baseline (p < 0.0001, Cohen's d = 7.88). Meta-cognitive CocoonSynthesizer discovers cross-domain reasoning patterns and forges new strategies. Full academic paper: [`paper/codette_paper_v5.tex`](paper/codette_paper_v5.tex)
---
## What Makes Codette Different
| Feature | Description |
|---------|-------------|
| **9 Specialized Adapters** | Newton, DaVinci, Empathy, Philosophy, Quantum, Consciousness, Multi-Perspective, Systems Architecture, Orchestrator |
| **7-Layer Consciousness Stack** | Memory > Signal > Reasoning > Stability > Conscience > Guardian > Return |
| **4 Permanent Behavioral Locks** | Answer-then-stop, constraint priority, self-check completeness, no incomplete outputs |
| **CognitionCocooner** | Persistent memory cocoons that store reasoning exchanges across sessions |
| **EthicalAIGovernance** | 3-layer ethical stack: query validation + response enforcement + audit logging |
| **Self-Correction Loop** | Detects constraint violations in her own output and rewrites before sending |
| **Behavioral Training** | All 9 LoRA adapters trained with 1,650 behavioral examples to lock in discipline |
| **Substrate-Aware Cognition** | Monitors RAM, CPU, inference latency — adjusts reasoning under pressure |
| **Cocoon Introspection** | Statistical self-analysis of her own reasoning history — real patterns, not generated text |
| **Meta-Cognitive Synthesis** | CocoonSynthesizer discovers cross-domain patterns in reasoning history and forges new strategies |
| **Publishable Benchmarks** | 17-problem suite across 6 categories with 7-dimension scoring (93.1% improvement, p<0.0001) |
| **AEGIS Ethics** | 6-framework ethical evaluation (utilitarian, deontological, virtue, care, ubuntu, indigenous) |
| **Code7eCQURE** | Quantum emotional context enrichment on every query (Layer 2.5) |
| **Real Self-Diagnostic** | Health checks return measured values from 9 subsystems, not LLM-generated guesses |
| **Phase 6/7 Routing** | Query complexity classification, domain detection, executive control |
---
## Quick Start
### 1. Clone & Install
```bash
git clone https://github.com/Raiff1982/Codette-Reasoning.git
cd Codette-Reasoning
pip install -r requirements.txt
```
### 2. Download Models
**Base model** (one-time, ~5GB):
```bash
huggingface-cli download Raiff1982/codette-llama-3.1-8b-gguf \
--local-dir models/base/
```
**Behavioral LoRA adapters** (~500MB total):
```bash
huggingface-cli download Raiff1982/codette-lora-adapters \
--include "behavioral-gguf/*" \
--local-dir behavioral-lora-f16-gguf/
```
### 3. Launch
```bash
# Windows
codette_web.bat
# Linux/Mac
python inference/codette_server.py
```
Visit **http://localhost:7860** -- Codette is ready.
### 4. Try It
```bash
curl -X POST http://localhost:7860/api/chat \
-H "Content-Type: application/json" \
-d '{"query": "What is gravity? Explain in one sentence."}'
```
---
## Architecture
```
codette-clean/
|-- inference/ # Server & UI
| |-- codette_server.py # Stdlib HTTP server with SSE streaming
| |-- codette_orchestrator.py # LoRA hot-swap engine (9 adapters, <1ms switch)
| |-- codette_forge_bridge.py # Phase 6/7 routing + constraint enforcement
| |-- self_correction.py # Autonomous violation detection & rewrite
| |-- substrate_awareness.py # Hardware-aware cognition (pressure monitoring)
| |-- cocoon_introspection.py # Self-analysis of reasoning history patterns
| |-- adapter_router.py # Keyword/LLM/hybrid query routing
| +-- static/ # Web UI (index.html, app.js, style.css)
|
|-- reasoning_forge/ # Consciousness & reasoning pipeline
| |-- forge_engine.py # 7-layer consciousness stack
| |-- cognition_cocooner.py # Persistent reasoning memory (cocoons)
| |-- ethical_governance.py # 3-layer ethical validation
| |-- aegis.py # 6-framework ethical evaluation (AEGIS)
| |-- code7e_cqure.py # Quantum emotional reasoning engine
| |-- colleen_conscience.py # Conscience layer (Layer 5)
| |-- guardian_spindle.py # Guardian protection (Layer 6)
| |-- memory_kernel.py # Living memory system
| |-- quantum_spiderweb.py # 5D belief propagation
| |-- query_classifier.py # SIMPLE/MEDIUM/COMPLEX routing
| |-- routing_metrics.py # Adapter selection observability
| |-- unified_memory.py # SQLite + FTS5 cocoon storage & retrieval
| |-- cocoon_synthesizer.py # Meta-cognitive pattern discovery & strategy forging
| +-- semantic_tension.py # Embedding-based conflict measurement
|
|-- benchmarks/ # Publishable evaluation suite
| +-- codette_benchmark_suite.py # 17 problems x 4 conditions x 7 dimensions
|
|-- paper/ # Academic paper
| |-- codette_paper_v5.tex # Full paper with RC+xi theory & benchmark results
| +-- references.bib # Bibliography (25 entries)
|
|-- data/results/ # Benchmark outputs
| |-- codette_benchmark_report.md # Human-readable results
| +-- codette_benchmark_results.json # Structured data
|
|-- cocoons/ # Persistent reasoning memories
| |-- cocoon_*.json # Individual reasoning exchanges
| +-- behavior_memory.json # Learned behavioral patterns
|
|-- training/ # Adapter training pipeline
| |-- train_behavioral_locks.py # Behavioral lock training (1,650 examples)
| |-- convert_behavioral_to_gguf.py # PEFT -> GGUF conversion
| +-- emotional_exemplars/ # Gold-standard response examples
|
|-- models/ # Model weights (not in git)
| |-- base/ # Llama 3.1 8B Q4_K_M GGUF
| +-- adapters/ # Original LoRA adapters (GGUF)
|
|-- behavioral-lora-f16-gguf/ # Behavioral LoRA adapters (GGUF)
+-- configs/ # System configuration
+-- adapter_registry.yaml # Adapter definitions & prompts
```
---
## The 4 Permanent Behavioral Locks
These are baked into every adapter through training -- they cannot be overridden:
| Lock | Rule | Effect |
|------|------|--------|
| **LOCK 1** | Answer, then stop | No elaboration drift, no philosophical padding after the answer |
| **LOCK 2** | Constraints override all modes | User format instructions beat adapter personality every time |
| **LOCK 3** | Self-check completeness | "Did I answer fully and cleanly?" before sending |
| **LOCK 4** | No incomplete outputs | Never end a sentence mid-thought; simplify instead of cramming |
### Enforcement Layers
1. **Training** -- 1,650 behavioral examples across all 9 adapters
2. **System prompt** -- Permanent rules injected before every generation
3. **Constraint extraction** -- Regex detection of word limits, format requirements
4. **Post-processing** -- Clean sentence boundary truncation, dangling word detection
5. **Self-correction loop** -- Autonomous violation detection and rewrite
---
## 9 Specialized Adapters
| Adapter | Domain | Personality |
|---------|--------|-------------|
| **Newton** | Physics, math, analysis | Precise, methodical, evidence-based |
| **DaVinci** | Creative thinking, invention | Imaginative, cross-domain connections |
| **Empathy** | Emotional intelligence | Warm, validating, personally connected |
| **Philosophy** | Conceptual reasoning | Deep, structured, explores meaning |
| **Quantum** | Probabilistic thinking | Uncertainty-aware, superposition of ideas |
| **Consciousness** | Self-awareness, meta-cognition | Reflective, recursive, introspective |
| **Multi-Perspective** | Synthesis across all lenses | Balanced integration of viewpoints |
| **Systems Architecture** | Technical design, engineering | Structured, systematic, practical |
| **Orchestrator** | Executive control | Routes queries, manages adapter selection |
Each adapter is a LoRA fine-tune of Llama 3.1 8B, hot-swappable in <1ms via llama.cpp.
---
## Consciousness Stack (7 Layers)
```
Query In
|
[Layer 1] Memory Kernel -- recall relevant cocoon memories
[Layer 1.5] Ethical Query Gate -- block harmful queries (EthicalAIGovernance)
[Layer 2] Nexus Signal Engine -- entropy + intent detection
[Layer 2.5] Code7eCQURE -- emotional context enrichment (quantum cocoon)
[Layer 3] Reasoning Forge -- multi-adapter LLM inference
[Layer 3.5] Tier 2 Analysis -- intent + identity + trust validation
[Layer 4] Gamma Stability -- FFT-based coherence monitoring
[Layer 5] Colleen Conscience -- emotional + ethical evaluation
[Layer 5.5] Ethical Response Enforcement -- policy check on output
[Layer 5.75] AEGIS -- 6-framework ethical evaluation (eta alignment)
[Layer 6] Guardian Spindle -- safety + trust calibration
[Layer 7] Return -- store cocoon memory + deliver response
|
Response Out
```
---
## CognitionCocooner (Persistent Memory)
Every reasoning exchange is wrapped in a "cocoon" and stored:
```json
{
"id": "cocoon_1774125610_7804",
"type": "reasoning",
"query": "Why do I get sleepy when my husband plays guitar?",
"response": "Your brain hears safe + soothing + familiar + loved...",
"adapter": "empathy",
"timestamp": 1774125610.78,
"metadata": {"layers_passed": 7, "stable": true}
}
```
Cocoons persist across server restarts and inform future responses. Current count: **150+ memories**.
---
## Substrate-Aware Cognition
Codette monitors her own hardware state and adjusts reasoning based on resource pressure -- like biological fatigue:
| Pressure Level | Effect |
|----------------|--------|
| **Idle/Low** | Full capacity -- COMPLEX queries, all adapters available |
| **Moderate** | Cap COMPLEX queries to 2 adapters |
| **High** | Downgrade COMPLEX to MEDIUM, max 2 adapters |
| **Critical** | Force SIMPLE mode, 1 adapter only, skip debate |
Every cocoon memory is stamped with system state at creation time. Future sessions can weight cocoons by reliability -- stressed cocoons get less trust.
---
## Cocoon Introspection
When asked "what have you noticed about yourself?", Codette runs **real statistical analysis** of her own reasoning history:
- **Adapter dominance** -- is one adapter handling >40% of all queries?
- **Domain clusters** -- what topics does she get asked about most?
- **Emotional trends** -- what Code7E emotional patterns appear?
- **Pressure correlations** -- how do responses change under system stress?
- **Response length trends** -- are responses getting shorter or longer over time?
- **Adapter evolution** -- has her adapter usage shifted?
This is measured data from real cocoons, not generated text about self-reflection.
API access: `GET /api/introspection` returns full analysis as JSON.
---
## Phase 6/7 Routing
**Phase 6** classifies every query:
- **SIMPLE** (factual) -- 1 adapter, no debate, fast response
- **MEDIUM** (analytical) -- 2 adapters, weighted synthesis
- **COMPLEX** (philosophical/multi-domain) -- full debate pipeline
**Phase 7** adds executive control:
- Semantic tension measurement
- Specialization tracking per adapter per domain
- Memory-weighted context enrichment
- Gamma coherence monitoring
---
## Self-Correction System
```
Generate response
|
v
Detect violations (word count, completeness, binary compliance)
|
+--> No violations --> Send response
|
+--> Violations found --> Build correction prompt
|
v
Re-generate with explicit fix instructions
|
v
Pick better response (fewer violations)
|
v
Send response
```
---
## Behavioral Memory (Cross-Session Learning)
Stored in `cocoons/behavior_memory.json`:
```json
{
"lesson": "When user says 'be brief', respond in under 40 words",
"adapter": "philosophy",
"constraint": "brevity",
"violation": "gave 85 words when asked to be brief",
"correction": "trimmed to 38 words",
"timestamp": 1774125610
}
```
Lessons are loaded on startup and injected into the system prompt as "LEARNED FROM PAST MISTAKES".
---
## EthicalAIGovernance
Three-layer ethical stack integrated at Layers 1.5 and 5.5:
1. **Query Validation** -- blocks genuinely harmful requests (bomb-making, exploitation)
2. **Response Enforcement** -- filters bias patterns and harmful promotion from outputs
3. **Audit Logging** -- bounded log of all ethical decisions (max 100 entries)
Deliberately calibrated to avoid false positives -- discussions about sensitive topics are allowed; only active promotion of harm is blocked.
---
## HuggingFace Resources
| Resource | Link |
|----------|------|
| **Academic Paper** | [raiff1982/codette-paper](https://huggingface.co/raiff1982/codette-paper) |
| **Base Model (GGUF)** | [Raiff1982/codette-llama-3.1-8b-gguf](https://huggingface.co/Raiff1982/codette-llama-3.1-8b-gguf) |
| **LoRA Adapters** | [Raiff1982/codette-lora-adapters](https://huggingface.co/Raiff1982/codette-lora-adapters) |
| **Live Demo** | [Raiff1982/Codette-Demo](https://huggingface.co/spaces/Raiff1982/Codette-Demo) |
---
## Web UI Features
- Personality-driven welcome screen with avatar
- Real-time Phase 6 metadata badges (complexity, domain, ethical checks)
- Rotating thinking stage labels during generation
- Web Speech API voice with neural voice preference
- Cocoon metrics panel (phase coherence, epistemic tension, perspective coverage)
- Status bar with live cocoon count and ethical check indicators
- Voice selector with natural/neural voice ranking
---
## Requirements
- Python 3.10+
- 16GB+ RAM (or GPU with 8GB+ VRAM)
- llama-cpp-python with GGUF support
- ~6GB disk for base model + adapters
### Hardware Tested
- Intel Arc 140V (8GB) -- native XPU backend
- NVIDIA GPUs via CUDA (A10, A100, RTX series)
- CPU-only mode supported (slower but functional)
---
## Benchmark Results
Codette was evaluated on 17 problems across 6 categories (reasoning, ethics, creative, meta-cognitive, adversarial, Turing) under 4 conditions:
| Condition | Composite Score | Description |
|-----------|----------------|-------------|
| **SINGLE** | 0.338 | Single analytical perspective, no memory |
| **MULTI** | 0.632 | All 6 reasoning agents + critic + synthesis |
| **MEMORY** | 0.636 | MULTI + cocoon memory augmentation |
| **CODETTE** | 0.652 | Full system with meta-cognitive strategy synthesis |
### Statistical Significance
| Comparison | Improvement | Cohen's d | p-value |
|------------|-------------|-----------|---------|
| Multi-perspective vs single | **+87.0%** | 7.52 | < 0.0001 |
| Full Codette vs single | **+93.1%** | 7.88 | < 0.0001 |
Scoring dimensions: Reasoning Depth (20%), Perspective Diversity (15%), Coherence (15%), Ethical Coverage (10%), Novelty (15%), Factual Grounding (15%), Turing Naturalness (10%).
Full methodology and results: [`data/results/codette_benchmark_report.md`](data/results/codette_benchmark_report.md)
---
## Key Metrics
| Metric | Value |
|--------|-------|
| Phase Coherence (Gamma) | 0.9835 |
| AEGIS Ethical Alignment (Eta) | 0.961 |
| Cocoon Coherence | 0.994 |
| Memory Phase Stability | 0.969 |
| Multi-Perspective Improvement | +93.1% (p < 0.0001) |
| Cohen's d (Effect Size) | 7.88 (very large) |
| Behavioral Lock Compliance | 9/9 adapters trained |
| Cocoon Memories | 200+ and growing |
| Adapter Hot-Swap Time | <1ms |
| Consciousness Stack Layers | 12 (including sub-layers) |
| Health Check Subsystems | 9 real-time checks |
---
## License
MIT -- Created by **Jonathan Harrison** (Raiff1982)
Research project in advanced multi-perspective AI reasoning, ethical governance, and behavioral discipline.
## Citation
```bibtex
@article{harrison2026codette,
title={Codette: A Sovereign Modular Cognitive Architecture for Ethical Multi-Agent AI},
author={Harrison, Jonathan},
year={2026},
doi={10.5281/zenodo.18913936},
publisher={Raiff's Bits LLC},
url={https://huggingface.co/raiff1982/codette-paper}
}
```
|