models: - model: princeton-nlp/Llama-3-8B-ProLong-512k-Instruct # Base model: no additional parameters necessary - model: Casual-Autopsy/L3-bluuwhale-SAO-MIX-8B-V1_fp32-merge-calc parameters: weight: 0.5 # Adjusts influence of roleplay features from L3-bluuwhale density: 0.6 # Preserves around 60% of significant parameters from the roleplay model merge_method: della base_model: princeton-nlp/Llama-3-8B-ProLong-512k-Instruct parameters: epsilon: 0.05 # Fine-tunes the granularity of pruning lambda: 1.0 # Scaling factor to harmonize parameter influence normalize: true # Ensures parameters align without large deviations int8_mask: true # Uses an efficient format to handle larger context dtype: float32 out_dtype: bfloat16 # Output type to balance precision and efficiency