The generative output is strange
#11 opened 3 months ago
by
tangpeng
Speeds compared to llama_cpp_python?
2
#10 opened 9 months ago
by
SpaceCowboy850
Unable to start TGI service for TheBloke/Mixtral-8x7B-v0.1-GPTQ with num_shard as 4
#9 opened 10 months ago
by
swapnil3597
What would be the minimal Sagemaker instance to deploy this model ?
2
#7 opened 11 months ago
by
CarlosAndrea
ValueError: Unsupported model type mixtral
1
#6 opened 11 months ago
by
seabasshn
RuntimeError: shape '[32, 8]' is invalid for input of size 0
7
#5 opened 11 months ago
by
woldeM
Are you going to release mixtral-8x7B-v0.1-awq
#4 opened 11 months ago
by
HelloJiang
Running the model using "pip install auto-gptq" still results in "CUDA extension not installed"
#3 opened 11 months ago
by
mvetter
TypeError: mixtral isn't supported yet.
2
#2 opened 11 months ago
by
luv2261
Build AutoGPTQ from source
3
#1 opened 11 months ago
by
PeePants