Spaces:
Sleeping
Sleeping
BASEDIR=$(dirname "$0") | |
cd $BASEDIR | |
echo Current Directory: | |
pwd | |
nvidia-smi | |
uname -a | |
cat /etc/os-release | |
lscpu | |
grep MemTotal /proc/meminfo | |
pip install -r requirements.txt | |
# Evaluating Models (from Smallest to Largest) for the MS MARCO Dataset | |
./eval_modules/eval-hf_v2.sh google gemma-1.1-2b-it true | |
./scripts/eval-hf_v2.sh microsoft Phi-3-mini-128k-instruct true | |
./eval_modules/eval-hf_v2.sh google gemma-1.1-7b-it true | |
./eval_modules/eval-hf_v2.sh mistralai Mistral-7B-Instruct-v0.2 true | |
./eval_modules/eval-hf_v2.sh meta-llama Llama-2-7b-chat-hf true | |
./eval_modules/eval-hf_v2.sh meta-llama Llama-2-13b-chat-hf true | |
./eval_modules/eval-hf_v2.sh meta-llama Llama-2-70b-chat-hf true | |
./eval_modules/eval-hf_v2.sh meta-llama Meta-Llama-3-8B-Instruct true | |
./eval_modules/eval-hf_v2.sh meta-llama Meta-Llama-3-70B-Instruct true | |
# Evaluating Models (from Smallest to Largest) for the WebQSP Dataset | |
./eval_modules/eval-hf_v2.sh google gemma-1.1-2b-it false | |
./scripts/eval-hf_v2.sh microsoft Phi-3-mini-128k-instruct false | |
./eval_modules/eval-hf_v2.sh google gemma-1.1-7b-it false | |
./eval_modules/eval-hf_v2.sh mistralai Mistral-7B-Instruct-v0.2 false | |
./eval_modules/eval-hf_v2.sh meta-llama Llama-2-7b-chat-hf false | |
./eval_modules/eval-hf_v2.sh meta-llama Llama-2-13b-chat-hf false | |
./eval_modules/eval-hf_v2.sh meta-llama Llama-2-70b-chat-hf false | |
./eval_modules/eval-hf_v2.sh meta-llama Meta-Llama-3-8B-Instruct false | |
./eval_modules/eval-hf_v2.sh meta-llama Meta-Llama-3-70B-Instruct false | |