File size: 2,683 Bytes
9805805 c7bc841 c7744d2 c7bc841 c7744d2 c7bc841 2cda7da ffd6ee4 2cda7da ffd6ee4 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 |
---
license: apache-2.0
language:
- en
base_model:
- meta-llama/Meta-Llama-3.1-8B
---
# Empathetic teacher model
## Overview
This is a LLM fine-tuned with real-life, ideally-empathetic teacher-student conversations.
This model processes the recent conversation history and provides guidance on how a teacher might respond to the student's utterance.
To fine-tune an open-weighted LLM to act as this generic teacher, we have used the following datasets:
the Teacher-Student Chatroom Corpus, TSCCv2 [Caines et al., 2022](https://aclanthology.org/2022.nlp4call-1.3),
CIMA [Stasaski et al., 2020](https://aclanthology.org/2020.bea-1.5),
the Multicultural Classroom Discourse Dataset [Rapanta et al., 2021](https://www.sciencedirect.com/science/article/pii/S2352340921007940),
MathDial [Macina et al., 2023](https://aclanthology.org/2023.findings-emnlp.372), and
Conversational Uptake [Demszky et al., 2021].
We are evaluating LLaMa-3 for this task.
Instead of using programmable fine-tuning libraries such as Axolotl ([link](https://github.com/OpenAccess-AI-Collective/axolotl))
or Huggingface TRL ([link](https://github.com/huggingface/trl)),
we are employing the more general command-line LLaMA-Factory ([link](https://github.com/hiyouga/LLaMA-Factory)) toolkit
that facilitates the fine-tuning of various well-known LLMs on custom data.
Parameter-efficient fine-tuning is achieved via the QLoRA method [Dettmers et al., 2023](https://proceedings.neurips.cc/paper_files/paper/2023/file/1feb87871436031bdc0f2beaa62a049b-Paper-Conference.pdf).
## Usage Guide
This project was executed on an Ubuntu 22.04.3 system running Linux kernel 6.8.0-40-generic.
### Installation
To get started, you first need to set up the environment using the **LLaMA-Factory** project. Please refer to the official [LLaMA-Factory repository](https://github.com/hiyouga/LLaMA-Factory) for more details.
You can install the project by running the following commands:
```bash
git clone --depth 1 https://github.com/hiyouga/LLaMA-Factory.git
cd LLaMA-Factory
pip install -e ".[torch,metrics]"
```
### Execution
In the DeMINT project, the model was utilized to create a REST API. Below is an example of how to configure and run it.
**Setting Server Configuration**
To specify the port and server address, use the following environment variables:
To set the port and the address of the server:
```bash
# Default 8000
export KIND_TEACHER_PORT=8000
# Default localhost
export KIND_TEACHER_HOST="localhost"
```
**Running the Program**
Once the environment is configured, you can execute the program by running the following command:
```bash
llamafactory-cli api run_api_inference_1.yaml
``` |