Edit model card

Phi-3-mini-128k-instruct- abliterated-v3 -geminified

Credit to u/Anduin1357 on reddit for the name who wrote this comment

My Jupyter "cookbook" to replicate the methodology can be found here, refined library coming soon

What's this?

Well, after my abliterated models, I figured I should cover all the possible ground of such work and introduce a model that acts like the polar opposite of them. This is the result of that, and I feel it lines it up in performance to a certain search engine's AI model series.

Summary

This is microsoft/Phi-3-mini-128k-instruct with orthogonalized bfloat16 safetensor weights, generated with a refined methodology based on that which was described in the preview paper/blog post: 'Refusal in LLMs is mediated by a single direction' which I encourage you to read to understand more.

This model has been orthogonalized to act more like certain rhymes-with-Shmemini models.

Downloads last month
149
Safetensors
Model size
3.82B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for failspy/Phi-3-mini-4k-geminified

Merges
1 model
Quantizations
2 models

Space using failspy/Phi-3-mini-4k-geminified 1

Collection including failspy/Phi-3-mini-4k-geminified