New discussion

output embeddings

#54 opened 4 months ago by pureve

output content

#53 opened 4 months ago by pureve

add template

#51 opened 9 months ago by philschmid

Issues with CUDA and exllama_kernels

9
#47 opened about 1 year ago by ditchtech

Understanding materials

1
#37 opened about 1 year ago by rishabh-gurbani

Temperature or top_p is not working

2
#35 opened about 1 year ago by chintan4560

Train model with webui

1
#34 opened about 1 year ago by Samitoo

HuggingFace's bitsandbytes vs AutoGPTQ?

2
#33 opened about 1 year ago by chongcy

Dataset used for quantisation

2
#31 opened about 1 year ago by CarlosAndrea

Necessary material for llama2

7
#27 opened over 1 year ago by Samitoo

Llama-2-13B-chat-GPTQ problem

2
#23 opened over 1 year ago by nigsdf

General Update Question for LLMs

2
#17 opened over 1 year ago by Acrious

CPU Inference

1
#13 opened over 1 year ago by Ange09

Slow Inference Speed

#12 opened over 1 year ago by asifahmed

Censorship is hilarious

6
#10 opened over 1 year ago by tea-lover-418

Llama v2 GPTQ context length

6
#7 opened over 1 year ago by andrewsameh

Prompt format

8
#5 opened over 1 year ago by mr96

Bravo! That was fast : )

2
#3 opened over 1 year ago by jacobgoldenart

Doesn't contain the files

3
#1 opened over 1 year ago by aminedjeghri