Edit model card

AgoraMix is crafted using semi-automated proesses to merge top-performing models.

Note: This model was a response to seeing density parameters on model_stock merges on huggingface, an experiment to see their results. Meanwhile, checking the main branch of mergekit's merge_methods reveals that these parameters are unused. I don't have a fork that would allow that. So, this is a normal model_stock merge, but its performance is still promising.

Ancestor Models

Models Merged

The following YAML configuration was used to produce this model:

merge_method:        model_stock
base_model:          Qwen/Qwen2.5-14B
tokenizer_source:    base
parameters:
  int8_mask:         false
  normalize:         true
  rescale:           false
models:
  - model:           VAGOsolutions/SauerkrautLM-v2-14b-DPO
  - model:           arcee-ai/SuperNova-Medius
  - model:           CultriX/Qwen2.5-14B-Wernicke
  - model:           rombodawg/Rombos-LLM-V2.6-Qwen-14b
  - model:           underwoods/medius-erebus-magnum-14b
dtype:               bfloat16
out_dtype:           bfloat16
Downloads last month
9
Safetensors
Model size
14.8B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for sometimesanotion/AgoraMix-14B-model_stock-v0.1