license: mit | |
language: | |
- en | |
pipeline_tag: text-generation | |
tags: | |
- moe | |
# FusionNet_7Bx2_MoE_v0.1 | |
Fine-tuned model on English language using MoE method. The improved version from FusionNet_7Bx2_MoE_14B. | |
## Model description | |
The FusionNet_7Bx2_MoE_v0.1 is a model to experiment with the MoE method, which could significantly increase the performance of the original model. The FusionNet has 12.9B parameters, and this model is fine-tuned. Enjoy! |