llm-qa-bench / README.md
dh-mc's picture
ms macro dataset
32a6937
|
raw
history blame
No virus
1.12 kB
---
title: Chat with LLMs
emoji: 🤖💬
colorFrom: purple
colorTo: blue
sdk: gradio
sdk_version: 4.26.0
app_file: app.py
pinned: true
short_description: 'Chat with LLMs'
---
## Running Locally
1. Check pre-conditions:
- [Git Large File Storage (LFS)](https://git-lfs.com/) must have been installed.
- Run `python --version` to make sure you're running Python version 3.10 or above.
- The latest PyTorch with GPU support must have been installed. Here is a sample `conda` command:
```
conda install -y pytorch torchvision torchaudio pytorch-cuda=12.1 -c pytorch -c nvidia
```
1. Clone the repo
```
git lfs install
git clone https://huggingface.co/spaces/inflaton-ai/llm-qa-bench
```
3. Install packages
```
pip install -r requirements.txt
4. Set up your environment variables
- By default, environment variables are loaded from `.env.example` file
- If you don't want to use the default settings, copy `.env.example` into `.env`. Your can then update it for your local runs.
5. Run automated test:
```
python qa_chain_test.py
```
6. Start the local server at `http://localhost:7860`:
```
python app.py
```