--- license: apache-2.0 --- This is the LACIE fine-tuned version of Llama-3 70B, finetuned according to our paper [LACIE: Listener-Aware Finetuning for Confidence Calibration in Large Language Models](arxiv.org/abs/2405.21028) This model is a fine-tuned version of [Llama-3 70B](https://huggingface.co/meta-llama/Meta-Llama-3-70B) that has been finetuned using data from [TriviaQA](https://huggingface.co/datasets/mandarjoshi/trivia_qa). LACIE is pragmatic preference-based finetuning method that optimizes models to be better calibrated w.r.t. both implicit and explicit confidence statements. The preferences in the dataset are based on correctness and whether listener accepted to rejected the answer. For more details, please see our paper. ## Model Architecture The architecture is the same as Llama-3 70B; the weights in this repo are adapter weights for Llama.