Edit model card

OpenCerebrum-2.0-7B

OpenCerebrum-2.0-7B is an open-source language model fine-tuned from the alpindale/Mistral-7B-v0.2-hf base model on a diverse dataset aimed at replicating capabilities of Aether Research's proprietary Cerebrum model.

The model was fine-tuned with SFT and DPO on approximately 7,000 examples across 15 data sources spanning coding, math, science, multi-turn conversation, RAG, reasoning, and general instruction-following. The goal was to assemble public datasets that could help the model achieve strong performance on benchmarks where Cerebrum excels.

Model Details

  • Base Model: alpindale/Mistral-7B-v0.2-hf
  • Parameters: 7 billion
  • Fine-Tuning Dataset Size: ~7,000 examples
  • Fine-Tuning Data: Advanced in-house curation techniques at Cognitive Computations, with 15 different data sources for DPO and SFT.
  • Language: English
  • License: Apache 2.0

Quants

EXL2 @bartowski

GGUF @bartowski

Intended Use

OpenCerebrum-2.0-7B is intended to be a powerful open-source model for coding, math, science, and general question-answering and text generation tasks. Its diverse fine-tuning data aims to equip it with broad knowledge and reasoning capabilities.

However, as an open-source replica trained on a subset of data compared to the original Cerebrum, it may not match Cerebrum's full performance. Additionally, biases and limitations of the fine-tuning data may be reflected in the model's outputs.

Limitations and Biases

  • The model may have biases and limitations inherited from its fine-tuning datasets. Thorough testing is needed to characterize these.
  • As the model is based on a 7B parameter model, it has computational and memory constraints compared to larger models.

Evaluations

Tasks Version Filter n-shot Metric Value Stderr
truthfulqa_mc2 2 none 0 acc 0.5182 ± 0.0152
ai2_arc N/A none 0 acc 0.7060 ± 0.0073
none 0 acc_norm 0.7049 ± 0.0074
- arc_challenge 1 none 0 acc 0.5000 ± 0.0146
none 0 acc_norm 0.5299 ± 0.0146
- arc_easy 1 none 0 acc 0.8077 ± 0.0081
none 0 acc_norm 0.7912 ± 0.0083
agieval_nous N/A none 0 acc 0.3778 ± 0.0093
none 0 acc_norm 0.3574 ± 0.0093
- agieval_aqua_rat 1 none 0 acc 0.2402 ± 0.0269
none 0 acc_norm 0.2205 ± 0.0261
- agieval_logiqa_en 1 none 0 acc 0.3164 ± 0.0182
none 0 acc_norm 0.3656 ± 0.0189
- agieval_lsat_ar 1 none 0 acc 0.2130 ± 0.0271
none 0 acc_norm 0.1913 ± 0.0260
- agieval_lsat_lr 1 none 0 acc 0.4078 ± 0.0218
none 0 acc_norm 0.3647 ± 0.0213
- agieval_lsat_rc 1 none 0 acc 0.4981 ± 0.0305
none 0 acc_norm 0.4498 ± 0.0304
- agieval_sat_en 1 none 0 acc 0.6650 ± 0.0330
none 0 acc_norm 0.5922 ± 0.0343
- agieval_sat_en_without_passage 1 none 0 acc 0.4612 ± 0.0348
none 0 acc_norm 0.3932 ± 0.0341
- agieval_sat_math 1 none 0 acc 0.3273 ± 0.0317
none 0 acc_norm 0.2818 ± 0.0304
Downloads last month
31
Safetensors
Model size
7.24B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for Locutusque/OpenCerebrum-2.0-7B

Merges
2 models
Quantizations
3 models

Spaces using Locutusque/OpenCerebrum-2.0-7B 5

Collection including Locutusque/OpenCerebrum-2.0-7B