--- license: other library_name: transformers base_model: - LucidityAI/Astral-4B-Coder - openfree/Darwin-Qwen3-4B - Qwen/Qwen3-4B tags: - qwen3 - mergekit - merge - text-generation-inference - code - coder - withinusai language: - en datasets: - LucidityAI/Astral-Post-Training-Dataset pipeline_tag: text-generation --- # Darwin-Astral-4B-Coder **Darwin-Astral-4B-Coder** is a merged 4B-class coding model release from **WithIn Us AI**, designed for code generation, instruction-following, and practical developer-assistant workflows. This repository is distributed as a standard **Transformers** checkpoint in **Safetensors** format and is positioned as a merge-based model that blends Darwin-style and Astral-style coding traits within a Qwen3-family 4B backbone. ## Model Summary This model is intended for: - code generation - code explanation - debugging assistance - implementation planning - instruction-following - developer assistant workflows - local or hosted coding inference As a 4B-class model, it aims to balance stronger coding capability than very small models with a lighter deployment footprint than larger coder checkpoints. ## Base Model Lineage The current repository metadata lists the following upstream model references: - `LucidityAI/Astral-4B-Coder` - `openfree/Darwin-Qwen3-4B` - `Qwen/Qwen3-4B` The visible merge configuration in the README also shows: - `Qwen/Qwen3-4B-Instruct-2507` as the base model in the YAML block - `Lucidity-AI-Astral-4B-Coder` as a merge source - `openfree-Darwin-Qwen3-4B` as a merge source These names are preserved here as shown on the repository page. ## Merge Details According to the current README: - this model is a merge of pre-trained language models - it was created using **mergekit** - the **SLERP** merge method was used The repository also includes a visible `mergekit_config.yml`, which supports the merge-based packaging of the release. ## Dataset Lineage The repository page currently shows the following dataset association: - `LucidityAI/Astral-Post-Training-Dataset` This suggests coding or post-training lineage connected to the Astral family used in the merge. ## Intended Use Recommended use cases include: - coding assistant experiments - generating utility functions and scripts - explaining code and technical concepts - debugging support - step-by-step implementation planning - local developer tools - hosted text-generation workflows for software tasks ## Suggested Use Cases This model can be useful for: - drafting Python, JavaScript, or general-purpose code - proposing refactors - generating boilerplate - answering developer questions - comparing implementation approaches - producing structured technical responses ## Out-of-Scope Use This model should not be relied on for: - legal advice - medical advice - financial advice - safety-critical automation - autonomous production engineering without review - security-critical code without expert validation All generated code should be reviewed, tested, and validated before real-world deployment. ## Repository Contents The repository currently includes standard Hugging Face model assets such as: - `README.md` - `.gitattributes` - `added_tokens.json` - `config.json` - `mergekit_config.yml` - `merges.txt` - `model-00001-of-00002.safetensors` - `model-00002-of-00002.safetensors` - `model.safetensors.index.json` - `special_tokens_map.json` - `tokenizer.json` - `tokenizer_config.json` ## Prompting Guidance This model will usually work best with prompts that are: - direct - scoped to a clear task - explicit about the language or framework - clear about whether code, explanation, or both are wanted - structured when step-by-step reasoning is useful ### Example prompt styles **Code generation** > Write a Python function that loads a JSON file, validates required keys, and returns cleaned records. **Debugging** > Explain why this code raises a KeyError and provide a safer corrected version. **Implementation planning** > Create a step-by-step plan for building a FastAPI service with authentication, logging, and tests. **Refactoring** > Refactor this function for readability and add basic error handling. ## Strengths This model may be especially useful for: - blended coding workflows - practical developer assistance - moderate-size local inference - structured software-task prompting - merge-model experimentation - compact coder deployments ## Limitations Like other merged 4B-class language models, this model may: - hallucinate APIs or implementation details - generate incomplete or incorrect code - produce insecure patterns - make reasoning mistakes on harder prompts - require prompt iteration for best results - need human validation before real-world use ## Attribution **WithIn Us AI** is the publisher of this merged model release. Credit for upstream assets remains with their original creators. The repository metadata and README specifically reference: - `LucidityAI/Astral-4B-Coder` - `openfree/Darwin-Qwen3-4B` - `Qwen/Qwen3-4B` - `Qwen/Qwen3-4B-Instruct-2507` and the dataset: - `LucidityAI/Astral-Post-Training-Dataset` ## License This draft uses: - `license: other` If you maintain this repo, replace this with the exact license terms you want displayed and make sure they align with any upstream obligations from the referenced source models and datasets. ## Acknowledgments Thanks to: - **WithIn Us AI** - **LucidityAI** - **openfree** - **Qwen** - the **mergekit** ecosystem - the Hugging Face platform - the broader open-source LLM community ## Disclaimer This model may produce inaccurate, insecure, biased, incomplete, or misleading outputs. All important generations, especially code and technical guidance, should be reviewed and tested before real-world use.