Hubble-4B-v1-EXL2 / README.md
Ex_y
Update README.md
3fe73e3 verified
|
raw
history blame
1.42 kB
metadata
license: other
base_model:
  - TheDrummer/Hubble-4B-v1
library_name: transformers
quantized_by: Ex_y
base_model_relation: quantized

EXL2 quants of TheDrummer/Hubble-4B-v1

Default parameter. 6.5bpw and 8.0 bpw uses 8 bit lm_head layer, while 4.25bpw and 5.0bpw uses 6 bit lm_head layer.

Join our Discord! https://discord.gg/Nbv9pQ88Xb

Works on Kobold 1.74!

(Layla (iOS / Android) support is in progress)


BeaverAI proudly presents...

Hubble 4B v1

Equipped with his five senses, man explores the universe around him and calls the adventure 'Science'.

image/png

Description

This is a finetune of Nvidia's Llama 3.1 4B Minitron - a shrunk down model of Llama 3.1 8B 128K.

Usage

  • ChatML or Text Completion
  • Add <|im_end|> as a stop token

Links

Technical Note

Hubble was trained on ChatML with <|end_of_text|> as the EOS token. If you encounter any issues with the model, please let me know!