--- language: - en license: mit tags: - codette - multi-perspective-reasoning - ethical-ai - lora - qlora - llama-3.1 - recursive-cognition - rc-xi - behavioral-locks - cognition-cocooner library_name: peft base_model: meta-llama/Llama-3.1-8B-Instruct model-index: - name: Codette RC+xi Reasoning Engine results: - task: type: text-generation name: Multi-Perspective Reasoning metrics: - name: Phase Coherence (Gamma) type: custom value: 0.9835 - name: AEGIS Ethical Alignment (Eta) type: custom value: 0.961 - name: Cocoon Coherence type: custom value: 0.994 - name: Memory Phase Stability type: custom value: 0.969 - name: Multi-Perspective vs Single (Composite) type: custom value: "+93.1%" - name: Benchmark p-value type: custom value: "<0.0001" - name: Cohen's d (Effect Size) type: custom value: 7.88 --- # Codette Reasoning Engine **Advanced Multi-Perspective AI with Conscience, Memory & Behavioral Discipline** Codette is a production-ready AI reasoning system that thinks from multiple angles simultaneously, remembers what she learns, and follows instructions with precision. Created by **Jonathan Harrison** (Raiff1982) > **New in v5**: Publishable benchmark suite with 17 problems across 6 categories demonstrates **93.1% improvement** over single-perspective baseline (p < 0.0001, Cohen's d = 7.88). Meta-cognitive CocoonSynthesizer discovers cross-domain reasoning patterns and forges new strategies. Full academic paper: [`paper/codette_paper_v5.tex`](paper/codette_paper_v5.tex) --- ## What Makes Codette Different | Feature | Description | |---------|-------------| | **9 Specialized Adapters** | Newton, DaVinci, Empathy, Philosophy, Quantum, Consciousness, Multi-Perspective, Systems Architecture, Orchestrator | | **7-Layer Consciousness Stack** | Memory > Signal > Reasoning > Stability > Conscience > Guardian > Return | | **4 Permanent Behavioral Locks** | Answer-then-stop, constraint priority, self-check completeness, no incomplete outputs | | **CognitionCocooner** | Persistent memory cocoons that store reasoning exchanges across sessions | | **EthicalAIGovernance** | 3-layer ethical stack: query validation + response enforcement + audit logging | | **Self-Correction Loop** | Detects constraint violations in her own output and rewrites before sending | | **Behavioral Training** | All 9 LoRA adapters trained with 1,650 behavioral examples to lock in discipline | | **Substrate-Aware Cognition** | Monitors RAM, CPU, inference latency — adjusts reasoning under pressure | | **Cocoon Introspection** | Statistical self-analysis of her own reasoning history — real patterns, not generated text | | **Meta-Cognitive Synthesis** | CocoonSynthesizer discovers cross-domain patterns in reasoning history and forges new strategies | | **Publishable Benchmarks** | 17-problem suite across 6 categories with 7-dimension scoring (93.1% improvement, p<0.0001) | | **AEGIS Ethics** | 6-framework ethical evaluation (utilitarian, deontological, virtue, care, ubuntu, indigenous) | | **Code7eCQURE** | Quantum emotional context enrichment on every query (Layer 2.5) | | **Real Self-Diagnostic** | Health checks return measured values from 9 subsystems, not LLM-generated guesses | | **Phase 6/7 Routing** | Query complexity classification, domain detection, executive control | --- ## Quick Start ### 1. Clone & Install ```bash git clone https://github.com/Raiff1982/Codette-Reasoning.git cd Codette-Reasoning pip install -r requirements.txt ``` ### 2. Download Models **Base model** (one-time, ~5GB): ```bash huggingface-cli download Raiff1982/codette-llama-3.1-8b-gguf \ --local-dir models/base/ ``` **Behavioral LoRA adapters** (~500MB total): ```bash huggingface-cli download Raiff1982/codette-lora-adapters \ --include "behavioral-gguf/*" \ --local-dir behavioral-lora-f16-gguf/ ``` ### 3. Launch ```bash # Windows codette_web.bat # Linux/Mac python inference/codette_server.py ``` Visit **http://localhost:7860** -- Codette is ready. ### 4. Try It ```bash curl -X POST http://localhost:7860/api/chat \ -H "Content-Type: application/json" \ -d '{"query": "What is gravity? Explain in one sentence."}' ``` --- ## Architecture ``` codette-clean/ |-- inference/ # Server & UI | |-- codette_server.py # Stdlib HTTP server with SSE streaming | |-- codette_orchestrator.py # LoRA hot-swap engine (9 adapters, <1ms switch) | |-- codette_forge_bridge.py # Phase 6/7 routing + constraint enforcement | |-- self_correction.py # Autonomous violation detection & rewrite | |-- substrate_awareness.py # Hardware-aware cognition (pressure monitoring) | |-- cocoon_introspection.py # Self-analysis of reasoning history patterns | |-- adapter_router.py # Keyword/LLM/hybrid query routing | +-- static/ # Web UI (index.html, app.js, style.css) | |-- reasoning_forge/ # Consciousness & reasoning pipeline | |-- forge_engine.py # 7-layer consciousness stack | |-- cognition_cocooner.py # Persistent reasoning memory (cocoons) | |-- ethical_governance.py # 3-layer ethical validation | |-- aegis.py # 6-framework ethical evaluation (AEGIS) | |-- code7e_cqure.py # Quantum emotional reasoning engine | |-- colleen_conscience.py # Conscience layer (Layer 5) | |-- guardian_spindle.py # Guardian protection (Layer 6) | |-- memory_kernel.py # Living memory system | |-- quantum_spiderweb.py # 5D belief propagation | |-- query_classifier.py # SIMPLE/MEDIUM/COMPLEX routing | |-- routing_metrics.py # Adapter selection observability | |-- unified_memory.py # SQLite + FTS5 cocoon storage & retrieval | |-- cocoon_synthesizer.py # Meta-cognitive pattern discovery & strategy forging | +-- semantic_tension.py # Embedding-based conflict measurement | |-- benchmarks/ # Publishable evaluation suite | +-- codette_benchmark_suite.py # 17 problems x 4 conditions x 7 dimensions | |-- paper/ # Academic paper | |-- codette_paper_v5.tex # Full paper with RC+xi theory & benchmark results | +-- references.bib # Bibliography (25 entries) | |-- data/results/ # Benchmark outputs | |-- codette_benchmark_report.md # Human-readable results | +-- codette_benchmark_results.json # Structured data | |-- cocoons/ # Persistent reasoning memories | |-- cocoon_*.json # Individual reasoning exchanges | +-- behavior_memory.json # Learned behavioral patterns | |-- training/ # Adapter training pipeline | |-- train_behavioral_locks.py # Behavioral lock training (1,650 examples) | |-- convert_behavioral_to_gguf.py # PEFT -> GGUF conversion | +-- emotional_exemplars/ # Gold-standard response examples | |-- models/ # Model weights (not in git) | |-- base/ # Llama 3.1 8B Q4_K_M GGUF | +-- adapters/ # Original LoRA adapters (GGUF) | |-- behavioral-lora-f16-gguf/ # Behavioral LoRA adapters (GGUF) +-- configs/ # System configuration +-- adapter_registry.yaml # Adapter definitions & prompts ``` --- ## The 4 Permanent Behavioral Locks These are baked into every adapter through training -- they cannot be overridden: | Lock | Rule | Effect | |------|------|--------| | **LOCK 1** | Answer, then stop | No elaboration drift, no philosophical padding after the answer | | **LOCK 2** | Constraints override all modes | User format instructions beat adapter personality every time | | **LOCK 3** | Self-check completeness | "Did I answer fully and cleanly?" before sending | | **LOCK 4** | No incomplete outputs | Never end a sentence mid-thought; simplify instead of cramming | ### Enforcement Layers 1. **Training** -- 1,650 behavioral examples across all 9 adapters 2. **System prompt** -- Permanent rules injected before every generation 3. **Constraint extraction** -- Regex detection of word limits, format requirements 4. **Post-processing** -- Clean sentence boundary truncation, dangling word detection 5. **Self-correction loop** -- Autonomous violation detection and rewrite --- ## 9 Specialized Adapters | Adapter | Domain | Personality | |---------|--------|-------------| | **Newton** | Physics, math, analysis | Precise, methodical, evidence-based | | **DaVinci** | Creative thinking, invention | Imaginative, cross-domain connections | | **Empathy** | Emotional intelligence | Warm, validating, personally connected | | **Philosophy** | Conceptual reasoning | Deep, structured, explores meaning | | **Quantum** | Probabilistic thinking | Uncertainty-aware, superposition of ideas | | **Consciousness** | Self-awareness, meta-cognition | Reflective, recursive, introspective | | **Multi-Perspective** | Synthesis across all lenses | Balanced integration of viewpoints | | **Systems Architecture** | Technical design, engineering | Structured, systematic, practical | | **Orchestrator** | Executive control | Routes queries, manages adapter selection | Each adapter is a LoRA fine-tune of Llama 3.1 8B, hot-swappable in <1ms via llama.cpp. --- ## Consciousness Stack (7 Layers) ``` Query In | [Layer 1] Memory Kernel -- recall relevant cocoon memories [Layer 1.5] Ethical Query Gate -- block harmful queries (EthicalAIGovernance) [Layer 2] Nexus Signal Engine -- entropy + intent detection [Layer 2.5] Code7eCQURE -- emotional context enrichment (quantum cocoon) [Layer 3] Reasoning Forge -- multi-adapter LLM inference [Layer 3.5] Tier 2 Analysis -- intent + identity + trust validation [Layer 4] Gamma Stability -- FFT-based coherence monitoring [Layer 5] Colleen Conscience -- emotional + ethical evaluation [Layer 5.5] Ethical Response Enforcement -- policy check on output [Layer 5.75] AEGIS -- 6-framework ethical evaluation (eta alignment) [Layer 6] Guardian Spindle -- safety + trust calibration [Layer 7] Return -- store cocoon memory + deliver response | Response Out ``` --- ## CognitionCocooner (Persistent Memory) Every reasoning exchange is wrapped in a "cocoon" and stored: ```json { "id": "cocoon_1774125610_7804", "type": "reasoning", "query": "Why do I get sleepy when my husband plays guitar?", "response": "Your brain hears safe + soothing + familiar + loved...", "adapter": "empathy", "timestamp": 1774125610.78, "metadata": {"layers_passed": 7, "stable": true} } ``` Cocoons persist across server restarts and inform future responses. Current count: **150+ memories**. --- ## Substrate-Aware Cognition Codette monitors her own hardware state and adjusts reasoning based on resource pressure -- like biological fatigue: | Pressure Level | Effect | |----------------|--------| | **Idle/Low** | Full capacity -- COMPLEX queries, all adapters available | | **Moderate** | Cap COMPLEX queries to 2 adapters | | **High** | Downgrade COMPLEX to MEDIUM, max 2 adapters | | **Critical** | Force SIMPLE mode, 1 adapter only, skip debate | Every cocoon memory is stamped with system state at creation time. Future sessions can weight cocoons by reliability -- stressed cocoons get less trust. --- ## Cocoon Introspection When asked "what have you noticed about yourself?", Codette runs **real statistical analysis** of her own reasoning history: - **Adapter dominance** -- is one adapter handling >40% of all queries? - **Domain clusters** -- what topics does she get asked about most? - **Emotional trends** -- what Code7E emotional patterns appear? - **Pressure correlations** -- how do responses change under system stress? - **Response length trends** -- are responses getting shorter or longer over time? - **Adapter evolution** -- has her adapter usage shifted? This is measured data from real cocoons, not generated text about self-reflection. API access: `GET /api/introspection` returns full analysis as JSON. --- ## Phase 6/7 Routing **Phase 6** classifies every query: - **SIMPLE** (factual) -- 1 adapter, no debate, fast response - **MEDIUM** (analytical) -- 2 adapters, weighted synthesis - **COMPLEX** (philosophical/multi-domain) -- full debate pipeline **Phase 7** adds executive control: - Semantic tension measurement - Specialization tracking per adapter per domain - Memory-weighted context enrichment - Gamma coherence monitoring --- ## Self-Correction System ``` Generate response | v Detect violations (word count, completeness, binary compliance) | +--> No violations --> Send response | +--> Violations found --> Build correction prompt | v Re-generate with explicit fix instructions | v Pick better response (fewer violations) | v Send response ``` --- ## Behavioral Memory (Cross-Session Learning) Stored in `cocoons/behavior_memory.json`: ```json { "lesson": "When user says 'be brief', respond in under 40 words", "adapter": "philosophy", "constraint": "brevity", "violation": "gave 85 words when asked to be brief", "correction": "trimmed to 38 words", "timestamp": 1774125610 } ``` Lessons are loaded on startup and injected into the system prompt as "LEARNED FROM PAST MISTAKES". --- ## EthicalAIGovernance Three-layer ethical stack integrated at Layers 1.5 and 5.5: 1. **Query Validation** -- blocks genuinely harmful requests (bomb-making, exploitation) 2. **Response Enforcement** -- filters bias patterns and harmful promotion from outputs 3. **Audit Logging** -- bounded log of all ethical decisions (max 100 entries) Deliberately calibrated to avoid false positives -- discussions about sensitive topics are allowed; only active promotion of harm is blocked. --- ## HuggingFace Resources | Resource | Link | |----------|------| | **Academic Paper** | [raiff1982/codette-paper](https://huggingface.co/raiff1982/codette-paper) | | **Base Model (GGUF)** | [Raiff1982/codette-llama-3.1-8b-gguf](https://huggingface.co/Raiff1982/codette-llama-3.1-8b-gguf) | | **LoRA Adapters** | [Raiff1982/codette-lora-adapters](https://huggingface.co/Raiff1982/codette-lora-adapters) | | **Live Demo** | [Raiff1982/Codette-Demo](https://huggingface.co/spaces/Raiff1982/Codette-Demo) | --- ## Web UI Features - Personality-driven welcome screen with avatar - Real-time Phase 6 metadata badges (complexity, domain, ethical checks) - Rotating thinking stage labels during generation - Web Speech API voice with neural voice preference - Cocoon metrics panel (phase coherence, epistemic tension, perspective coverage) - Status bar with live cocoon count and ethical check indicators - Voice selector with natural/neural voice ranking --- ## Requirements - Python 3.10+ - 16GB+ RAM (or GPU with 8GB+ VRAM) - llama-cpp-python with GGUF support - ~6GB disk for base model + adapters ### Hardware Tested - Intel Arc 140V (8GB) -- native XPU backend - NVIDIA GPUs via CUDA (A10, A100, RTX series) - CPU-only mode supported (slower but functional) --- ## Benchmark Results Codette was evaluated on 17 problems across 6 categories (reasoning, ethics, creative, meta-cognitive, adversarial, Turing) under 4 conditions: | Condition | Composite Score | Description | |-----------|----------------|-------------| | **SINGLE** | 0.338 | Single analytical perspective, no memory | | **MULTI** | 0.632 | All 6 reasoning agents + critic + synthesis | | **MEMORY** | 0.636 | MULTI + cocoon memory augmentation | | **CODETTE** | 0.652 | Full system with meta-cognitive strategy synthesis | ### Statistical Significance | Comparison | Improvement | Cohen's d | p-value | |------------|-------------|-----------|---------| | Multi-perspective vs single | **+87.0%** | 7.52 | < 0.0001 | | Full Codette vs single | **+93.1%** | 7.88 | < 0.0001 | Scoring dimensions: Reasoning Depth (20%), Perspective Diversity (15%), Coherence (15%), Ethical Coverage (10%), Novelty (15%), Factual Grounding (15%), Turing Naturalness (10%). Full methodology and results: [`data/results/codette_benchmark_report.md`](data/results/codette_benchmark_report.md) --- ## Key Metrics | Metric | Value | |--------|-------| | Phase Coherence (Gamma) | 0.9835 | | AEGIS Ethical Alignment (Eta) | 0.961 | | Cocoon Coherence | 0.994 | | Memory Phase Stability | 0.969 | | Multi-Perspective Improvement | +93.1% (p < 0.0001) | | Cohen's d (Effect Size) | 7.88 (very large) | | Behavioral Lock Compliance | 9/9 adapters trained | | Cocoon Memories | 200+ and growing | | Adapter Hot-Swap Time | <1ms | | Consciousness Stack Layers | 12 (including sub-layers) | | Health Check Subsystems | 9 real-time checks | --- ## License MIT -- Created by **Jonathan Harrison** (Raiff1982) Research project in advanced multi-perspective AI reasoning, ethical governance, and behavioral discipline. ## Citation ```bibtex @article{harrison2026codette, title={Codette: A Sovereign Modular Cognitive Architecture for Ethical Multi-Agent AI}, author={Harrison, Jonathan}, year={2026}, doi={10.5281/zenodo.18913936}, publisher={Raiff's Bits LLC}, url={https://huggingface.co/raiff1982/codette-paper} } ```