Locutusque/llama-3-neural-chat-v1-8b AWQ

image/png

Model Summary

I fine-tuned llama-3 8B on an approach similar to Intel's neural chat language model. I have slightly modified the data sources so it is stronger in coding, math, and writing. I use both SFT and DPO.

This model has great performance in writing and coding.

Training Data

  • Open-Orca/SlimOrca-Dedup
  • jondurbin/airoboros-3.2
  • microsoft/orca-math-word-problems-200k
  • m-a-p/Code-Feedback
  • MaziyarPanahi/WizardLM_evol_instruct_V2_196k
  • mlabonne/orpo-dpo-mix-40k
Downloads last month
7
Safetensors
Model size
1.98B params
Tensor type
I32
·
FP16
·
Inference Examples
Inference API (serverless) has been turned off for this model.

Model tree for solidrust/llama-3-neural-chat-v1-8b-AWQ

Quantized
(2)
this model

Datasets used to train solidrust/llama-3-neural-chat-v1-8b-AWQ

Collection including solidrust/llama-3-neural-chat-v1-8b-AWQ