--- base_model: - Steelskull/L3-MS-Astoria-70b - Sao10K/L3-70B-Euryale-v2.1 library_name: transformers tags: - mergekit - merge --- # Eurstoria-106B ![](Unholy.png) A Llama-3 Decoder only model by combining 2x finetuned Llama-3 70B models into 1. ## Merge Details ### Merge Method This model was merged using the passthrough merge method. ### Models Merged The following models were included in the merge: * [Steelskull/L3-MS-Astoria-70b](https://huggingface.co/Steelskull/L3-MS-Astoria-70b) * [Sao10K/L3-70B-Euryale-v2.1](https://huggingface.co/Sao10K/L3-70B-Euryale-v2.1) ### Configuration The following YAML configuration was used to produce this model: ```yaml slices: - sources: - model: Steelskull/L3-MS-Astoria-70b layer_range: [0, 16] - sources: - model: Sao10K/L3-70B-Euryale-v2.1 layer_range: [8, 24] - sources: - model: Steelskull/L3-MS-Astoria-70b layer_range: [25, 40] - sources: - model: Sao10K/L3-70B-Euryale-v2.1 layer_range: [33, 48] - sources: - model: Steelskull/L3-MS-Astoria-70b layer_range: [41, 56] - sources: - model: Sao10K/L3-70B-Euryale-v2.1 layer_range: [49, 64] - sources: - model: Steelskull/L3-MS-Astoria-70b layer_range: [57, 72] - sources: - model: Sao10K/L3-70B-Euryale-v2.1 layer_range: [65, 80] merge_method: passthrough dtype: float16 ``` # Prompting Format As both models are based off from instruct, Use the default [Llama 3 Instruct format](https://huggingface.co/Sao10K/L3-70B-Euryale-v2.1/blob/main/Euryale-v2.1-Llama-3-Instruct.json) would be fine. # Screenshots A Screenshot to confirm that: "It works" ![](8U0gwnhywS.png) # Benchmarks No one is willing to bench a 106B... right? # Acknowledgements The original ratios come from Goliath and are used as is. As such thanks to both [Alpindale](https://huggingface.co/alpindale) and [@Undi95](https://huggingface.co/Undi95) for their inital work. Credits goes to [@chargoddard](https://huggingface.co/chargoddard) for developing the framework used to merge the model - [mergekit](https://github.com/cg123/mergekit).