nxphi47 commited on
Commit
e78b7fd
1 Parent(s): 0d971ad

Update src/about.py

Browse files
Files changed (1) hide show
  1. src/about.py +12 -12
src/about.py CHANGED
@@ -43,37 +43,39 @@ LLM_BENCHMARKS_TEXT = f"""
43
  How It Works
44
  To participate in the ContextualBench leaderboard, follow these steps to evaluate your Large Language Model (LLM) using the ContextualBench framework:
45
 
46
- Clone the Repository: Start by cloning the ContextualBench GitHub repository to your local machine using the following command:
47
-
48
- bash
49
- git clone https://github.com/ShreyPandit/ContextualBench
50
 
 
 
 
51
 
52
  Navigate to the Directory: Move into the cloned repository's directory:
53
 
54
 
55
- bash
56
  cd ContextualBench
57
-
58
 
59
  Install Dependencies: Install all necessary dependencies by executing:
60
 
61
- bash
62
  pip install -r requirements.txt
 
63
 
64
  Prepare Your Model and Dataset: Set up your model and dataset according to the guidelines provided in the repository's documentation.
65
  Run the Evaluation Script: Execute the evaluation script to generate outputs for your model on the specified dataset:
66
 
67
 
68
- bash
69
  python run.py [dataset_name]
70
-
71
 
72
  Collect and Format Outputs: Gather the outputs generated for each dataset and format them according to the leaderboard submission guidelines.
73
  Submit Your Results: Email the formatted outputs to the author's email address for evaluation. Our team will assess the performance and update the leaderboard accordingly.
74
 
75
  Reproducibility
76
- Ensuring reproducibility is a key aspect of the ContextualBench leaderboard. By following the standardized steps outlined above, participants can consistently reproduce evaluation results. This process not only facilitates fair comparisons across different models but also encourages transparency and reliability in model assessments. Participants are encouraged to adhere strictly to the submission guidelines to ensure their results are accurately reflected on the leaderboard.
 
77
  """
78
 
79
  EVALUATION_QUEUE_TEXT = """
@@ -82,13 +84,11 @@ EVALUATION_QUEUE_TEXT = """
82
 
83
  CITATION_BUTTON_LABEL = "Copy the following snippet to cite these results"
84
  CITATION_BUTTON_TEXT = r"""
85
-
86
  ```
87
  @article{nguyen2024sfrrag,
88
  title={SFR-RAG: Towards Contextually Faithful LLMs},
89
  author={Nguyen, Xuan-Phi and Pandit, Shrey and Purushwalkam, Senthil and Xu, Austin and Chen, Hailin and Ming, Yifei and Ke, Zixuan and Savarese, Silvio and Xong, Caiming and Joty, Shafiq},
90
  year={2024}
91
  }
92
-
93
  ```
94
  """
 
43
  How It Works
44
  To participate in the ContextualBench leaderboard, follow these steps to evaluate your Large Language Model (LLM) using the ContextualBench framework:
45
 
46
+ Clone the Repository: Start by cloning the ContextualBench GitHub repository to your local machine using the following command:
 
 
 
47
 
48
+ ```bash
49
+ git clone https://github.com/SalesforceAIResearch/SFR-RAG
50
+ ```
51
 
52
  Navigate to the Directory: Move into the cloned repository's directory:
53
 
54
 
55
+ ``bash
56
  cd ContextualBench
57
+ ```
58
 
59
  Install Dependencies: Install all necessary dependencies by executing:
60
 
61
+ ```bash
62
  pip install -r requirements.txt
63
+ ```
64
 
65
  Prepare Your Model and Dataset: Set up your model and dataset according to the guidelines provided in the repository's documentation.
66
  Run the Evaluation Script: Execute the evaluation script to generate outputs for your model on the specified dataset:
67
 
68
 
69
+ ```bash
70
  python run.py [dataset_name]
71
+ ```
72
 
73
  Collect and Format Outputs: Gather the outputs generated for each dataset and format them according to the leaderboard submission guidelines.
74
  Submit Your Results: Email the formatted outputs to the author's email address for evaluation. Our team will assess the performance and update the leaderboard accordingly.
75
 
76
  Reproducibility
77
+ Ensuring reproducibility is a key aspect of the ContextualBench leaderboard.
78
+ By following the standardized steps outlined above, participants can consistently reproduce evaluation results. This process not only facilitates fair comparisons across different models but also encourages transparency and reliability in model assessments. Participants are encouraged to adhere strictly to the submission guidelines to ensure their results are accurately reflected on the leaderboard.
79
  """
80
 
81
  EVALUATION_QUEUE_TEXT = """
 
84
 
85
  CITATION_BUTTON_LABEL = "Copy the following snippet to cite these results"
86
  CITATION_BUTTON_TEXT = r"""
 
87
  ```
88
  @article{nguyen2024sfrrag,
89
  title={SFR-RAG: Towards Contextually Faithful LLMs},
90
  author={Nguyen, Xuan-Phi and Pandit, Shrey and Purushwalkam, Senthil and Xu, Austin and Chen, Hailin and Ming, Yifei and Ke, Zixuan and Savarese, Silvio and Xong, Caiming and Joty, Shafiq},
91
  year={2024}
92
  }
 
93
  ```
94
  """