mistral-nemo-gutenberg2-12B-test

mistralai/Mistral-Nemo-Instruct-2407 finetuned on nbeerbower/gutenberg2-dpo.

This model is a test for the sake of benchmarking my gutenberg2 dataset.

Method

Finetuned using an RTX 3090 for 3 epochs.

Fine-tune Llama 3 with ORPO

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 20.73
IFEval (0-Shot) 33.85
BBH (3-Shot) 32.04
MATH Lvl 5 (4-Shot) 10.20
GPQA (0-shot) 8.95
MuSR (0-shot) 10.97
MMLU-PRO (5-shot) 28.39
Downloads last month
20
Safetensors
Model size
12.2B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for nbeerbower/mistral-nemo-gutenberg2-12B-test

Finetuned
(32)
this model

Dataset used to train nbeerbower/mistral-nemo-gutenberg2-12B-test

Evaluation results