Edit model card

30B-Epsilon

Epsilon is an instruct based general purpose model assembled from hand picked models and LoRAs. There is no censorship and it follows instructions in the Alpaca format. This means you can create your own rules in the context memory of your inference system of choice [mainly KoboldAI or Text Generation Webui and chat UIs like SillyTavern and so on].

Composition:

This model is the result of an experimental use of LoRAs on language models and model merges. [] = applied as LoRA to a composite model | () = combined as composite models 30B-Epsilon = [SuperCOT[SuperHOT-prototype13b-8192[(wizardlmuncensored+((hippogriff+manticore)+(StoryV2))]

Alpaca's instruct format can be used to do many things, including control of the terms of behavior between a user and a response from an agent in chat. Below is an example of a command injected into memory.

### Instruction:
Make Narrator function as a text based adventure game that responds with verbose, detailed, and creative descriptions of what happens next after Player's response.
Make Player function as the player input for Narrator's text based adventure game, controlling a character named (insert character name here, their short bio, and
whatever quest or other information to keep consistent in the interaction).

### Response:
{an empty new line here}

All datasets from all models and LoRAs used were documented and reviewed as model candidates for merging. Model candidates were based on five core principles: creativity, logic, inference, instruction following, and longevity of trained responses. SuperHOT-prototype30b-8192 was used in this mix, not the 8K version; the prototype LoRA seems to have been removed [from HF] as of this writing. The GPT4Alpaca LoRA from Chansung was removed from this amalgam following a thorough review of where censorship and railroading the user came from in 33B-Lazarus. This is not a reflection of ChanSung's excellent work - it merely did not fit the purpose of this model.

Language Models and LoRAs Used Credits:

manticore-30b-chat-pyg-alpha [Epoch0.4] by openaccess-ai-collective

https://huggingface.co/openaccess-ai-collective/manticore-30b-chat-pyg-alpha

hippogriff-30b-chat by openaccess-ai-collective

https://huggingface.co/openaccess-ai-collective/hippogriff-30b-chat

WizardLM-33B-V1.0-Uncensored by ehartford

https://huggingface.co/ehartford/WizardLM-33B-V1.0-Uncensored

Storytelling-LLaMa-LoRA [30B, Version 2] by GamerUnTouch

https://huggingface.co/GamerUntouch/Storytelling-LLaMa-LoRAs

SuperCOT-LoRA [30B] by kaiokendev

https://huggingface.co/kaiokendev/SuperCOT-LoRA

SuperHOT-LoRA-prototype30b-8192 [30b, not 8K version, but a removed prototype] by kaiokendev

https://huggingface.co/kaiokendev/superhot-30b-8k-no-rlhf-test [Similar LoRA to one since removed that was used in making this model.]

Also thanks to Meta for LLaMA and to each and every one of you who developed these fine-tunes and LoRAs.

Downloads last month
1,489
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for CalderaAI/30B-Epsilon

Quantizations
3 models

Spaces using CalderaAI/30B-Epsilon 21