Edit model card

Replete-LLM-V2.5-Qwen-72b

image/png

Replete-LLM-V2.5-Qwen-72b is a continues finetuned version of Qwen2.5-72B. I noticed recently that the Qwen team did not learn from my methods of continuous finetuning, the great benefits, and no downsides of it. So I took it upon myself to merge the instruct model with the base model myself using the Ties merge method

This version of the model shows higher performance than the original instruct and base models.

Quants:

GGUF: https://huggingface.co/bartowski/Replete-LLM-V2.5-Qwen-72b-GGUF

EXL2 6bpw: https://huggingface.co/Kotokin/Replete-LLM-V2.5-Qwen-72b-exl2-6bpw

Benchmarks: (Coming soon)

Downloads last month
383
Safetensors
Model size
72.7B params
Tensor type
BF16
ยท
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for Replete-AI/Replete-LLM-V2.5-Qwen-72b

Base model

Qwen/Qwen2.5-72B
Finetuned
(18)
this model
Quantizations
2 models

Space using Replete-AI/Replete-LLM-V2.5-Qwen-72b 1

Collection including Replete-AI/Replete-LLM-V2.5-Qwen-72b