YAML Metadata Warning:empty or missing yaml metadata in repo card

Check out the documentation for more information.

This model produces strange output, and uses fancy words to avoid following any instructions without overtly refusing.

T7 merges Slimaki, Maginum Cydoms, Asmodeus v1, Asmodeus v2a, Asmodeus v2e, Magistry, and Checkpoint T6 via the della method.

Update

It appears that della merges can become overly unnormalized. Multiple attempts to stack layers of della with normalize false and total weights around 2 eventually leads to the model producing extremely long, complex and unrelated tangents which have nothing to do with the prompt.

I am now testing ties and dare_ties to see how these compare.

Update 2

ties does not allow you to merge 2501 finetunes/merges with 2503/2506/2509 the way that della does. It causes grammar collapse, forgetting words like "to".

T10

This was broken due to using ties and also censored T10

T11

Testing this now with della T11

Update 3

Here are the full yaml configs used to create this checkpoint

Asmodeus v2e

models:
  - model: B:\24B\models--mistralai--Magistral-Small-2509\textonly
  - model: B:\24B\models--Naphula--Slimaki-24B-v1
    parameters:
      weight: 0.3
      density: 0.9
      epsilon: 0.099
  - model: B:\24B\models--DarkArtsForge--Magistaroth-24B-v1
    parameters:
      weight: 0.3
      density: 0.9
      epsilon: 0.099
  - model: B:\24B\models--Casual-Autopsy--Maginum-Cydoms-24B
    parameters:
      weight: 0.3
      density: 0.9
      epsilon: 0.099
  - model: B:\24B\models--sophosympatheia--Magistry-24B-v1.0
    parameters:
      weight: 0.3
      density: 0.9
      epsilon: 0.099
  - model: B:\24B\!models--TheDrummer--Precog-24B-v1
    parameters:
      weight: 0.6
      density: 0.9
      epsilon: 0.099
  - model: B:\24B\models--DarkArtsForge--Asmodeus-24B-v1
    parameters:
      weight: 0.3
      density: 0.9
      epsilon: 0.099
merge_method: della
base_model: B:\24B\models--mistralai--Magistral-Small-2509\textonly
parameters:
  lambda: 1.0
  normalize: false
tokenizer:
  source: union
chat_template: auto
dtype: float32
out_dtype: bfloat16

T5

architecture: MistralForCausalLM
models:
  - model: B:\24B\models--anthracite-core--Mistral-Small-3.2-24B-Instruct-2506-Text-Only # B:\24B\models--mistralai--Magistral-Small-2509\textonly
  - model: B:\24B\BeaverAI_Fallen-Mistral-Small-3.1-24B-v1e_textonly
    parameters:
      weight: 0.25
      density: 0.9
      epsilon: 0.099
  - model: B:\24B\models--TheDrummer--Cydonia-24B-v4.3
    parameters:
      weight: 0.25
      density: 0.9
      epsilon: 0.099
  - model: B:\24B\models--TheDrummer--Magidonia-24B-v4.3
    parameters:
      weight: 0.25
      density: 0.9
      epsilon: 0.099
  - model: B:\24B\models--TheDrummer--Precog-24B-v1
    parameters:
      weight: 0.25
      density: 0.9
      epsilon: 0.099
  - model: B:\24B\models--zerofata--MS3.2-PaintedFantasy-v2-24B
    parameters:
      weight: 0.25
      density: 0.9
      epsilon: 0.099 
  - model: B:\24B\models--zerofata--MS3.2-PaintedFantasy-v3-24B
    parameters:
      weight: 0.25
      density: 0.9
      epsilon: 0.099
  - model: B:\24B\models--zerofata--MS3.2-PaintedFantasy-v4.1-24B
    parameters:
      weight: 0.25
      density: 0.9
      epsilon: 0.099
merge_method: della
base_model: B:\24B\models--anthracite-core--Mistral-Small-3.2-24B-Instruct-2506-Text-Only # B:\24B\models--mistralai--Magistral-Small-2509\textonly
parameters:
  lambda: 1.0
  normalize: false
  int8_mask: false
dtype: float32
out_dtype: bfloat16
tokenizer:
  source: union
chat_template: auto

T6

models:
  - model: B:\24B\models--mistralai--Magistral-Small-2509\textonly
  - model: B:\24B\models--Naphula--Slimaki-24B-v1
    parameters:
      weight: 0.25
      density: 0.9
      epsilon: 0.099
  - model: B:\24B\models--Casual-Autopsy--Maginum-Cydoms-24B
    parameters:
      weight: 0.25
      density: 0.9
      epsilon: 0.099
  - model: B:\24B\models--Naphula--GhostFace-24B-v1
    parameters:
      weight: 0.25
      density: 0.9
      epsilon: 0.099
  - model: B:\24B\DarkArtsForge__Asmodeus-24B-v2
    parameters:
      weight: 0.25
      density: 0.9
      epsilon: 0.099
  - model: B:\24B\models--sophosympatheia--Magistry-24B-v1.0
    parameters:
      weight: 0.25
      density: 0.9
      epsilon: 0.099
  - model: B:\24B\models--DarkArtsForge--Asmodeus-24B-v1
    parameters:
      weight: 0.25
      density: 0.9
      epsilon: 0.099
  - model: C:\mergekit-main\T5
    parameters:
      weight: 0.25
      density: 0.9
      epsilon: 0.099
  - model: B:\24B\Asmo_v2e
    parameters:
      weight: 0.25
      density: 0.9
      epsilon: 0.099
merge_method: della
base_model: B:\24B\models--mistralai--Magistral-Small-2509\textonly
parameters:
  lambda: 1.0
  normalize: false
tokenizer:
  source: union
chat_template: auto
dtype: float32
out_dtype: bfloat16

T7

models:
  - model: B:\24B\models--mistralai--Magistral-Small-2509\textonly
  - model: B:\24B\models--Naphula--Slimaki-24B-v1
    parameters:
      weight: 0.285
      density: 0.9
      epsilon: 0.099
  - model: B:\24B\models--Casual-Autopsy--Maginum-Cydoms-24B
    parameters:
      weight: 0.285
      density: 0.9
      epsilon: 0.099
  - model: B:\24B\DarkArtsForge__Asmodeus-24B-v2
    parameters:
      weight: 0.285
      density: 0.9
      epsilon: 0.099
  - model: B:\24B\models--sophosympatheia--Magistry-24B-v1.0
    parameters:
      weight: 0.285
      density: 0.9
      epsilon: 0.099
  - model: B:\24B\models--DarkArtsForge--Asmodeus-24B-v1
    parameters:
      weight: 0.285
      density: 0.9
      epsilon: 0.099
  - model: B:\24B\Asmo_v2e
    parameters:
      weight: 0.285
      density: 0.9
      epsilon: 0.099
  - model: B:\24B\T6
    parameters:
      weight: 0.285
      density: 0.9
      epsilon: 0.099
merge_method: della
base_model: B:\24B\models--mistralai--Magistral-Small-2509\textonly
parameters:
  lambda: 1.0
  normalize: false
tokenizer:
  source: union
chat_template: auto
dtype: float32
out_dtype: bfloat16
name: Checkpoint-T7-24B
Downloads last month
121
Safetensors
Model size
24B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Naphula-Archives/Checkpoint-T7-24B

Quantizations
2 models