Edit model card

merge

This is a merge of pre-trained language models created using mergekit.

Merge Details

Merge Method

This model was merged using the linear DARE merge method using Undi95/Meta-Llama-3-8B-hf as a base.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

merge_method: dare_linear  # Define the merging method (e.g., linear, dare_linear)
base_model: Undi95/Meta-Llama-3-8B-hf  # The base model for the merge
models:
  - model: Sao10K/L3-8B-Stheno-v3.2
    parameters:
      attention_heads: 0.6
      weight: 1.0  # Added missing weight parameter

  - model: ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.2
    parameters:
      feedforward: 0.4
      weight: 1.0  # Add weight for consistency

  - model: O1-OPEN/OpenO1-LLama-8B-v0.1
    parameters:
      all: 0.3
      weight: 1.0  # Add weight for consistency

  - model: huihui-ai/Llama-3.1-Tulu-3-8B-abliterated
    parameters:
      layer_norm: 0.2
      weight: 1.0  # Add weight for consistency

parameters:
  general: 0.6  # Default parameter for fallback
dtype: float16  # Data type for merging
tokenizer_source: union  # Tokenizer combining method
output_path: ./merged_model  # Output location for the merged model
Downloads last month
4
Safetensors
Model size
8.03B params
Tensor type
FP16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for mergekit-community/mergekit-dare_linear-mawqtuj