File size: 2,296 Bytes
7df5f68 c2b1691 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 |
---
configs:
- config_name: multi-turn_chat
data_files:
- split: test
path: multi-turn_chat.parquet
- config_name: code_completion
data_files:
- split: test
path: code_completion.parquet
- config_name: instruction_tuning
data_files:
- split: test
path: instruction_tuning.parquet
- config_name: code_fixing
data_files:
- split: test
path: code_fixing.parquet
- config_name: rag
data_files:
- split: test
path: rag.parquet
- config_name: large_summarization
data_files:
- split: test
path: large_summarization.parquet
- config_name: docstring
data_files:
- split: test
path: docstring.parquet
---
This dataset contains inference performance benchmarking obtained with vllm version 0.6.1.post2 on different use-case scenarios. The scenarios are defined as bellow:
<table>
<tr>
<td style="text-align: center"><strong>Use case</strong></td>
<td style="text-align: center"><strong>Prompt tokens</strong></td>
<td style="text-align: center"><strong>Generated tokens</strong></td>
</tr>
<tr>
<td>Code Completion</td>
<td style="text-align: center">256</td>
<td style="text-align: center">1024</td>
</tr>
<tr>
<td>Docstring Generation</td>
<td style="text-align: center">768</td>
<td style="text-align: center">128</td>
</tr>
<tr>
<td>Code Fixing</td>
<td style="text-align: center">1024</td>
<td style="text-align: center">1024</td>
</tr>
<tr>
<td>RAG</td>
<td style="text-align: center">1024</td>
<td style="text-align: center">128</td>
</tr>
<tr>
<td>Instruction Following</td>
<td style="text-align: center">256</td>
<td style="text-align: center">128</td>
</tr>
<tr>
<td>Multi-turn chat</td>
<td style="text-align: center">512</td>
<td style="text-align: center">256</td>
</tr>
<tr>
<td>Large Summarization</td>
<td style="text-align: center">4096</td>
<td style="text-align: center">512</td>
</tr>
</table>
Benchmarking was conducted with [GuideLLM](https://github.com/neuralmagic/guidellm) using the following syntax:
```
guidellm --model <model name> --data-type emulated --data "prompt_tokens=<prompt tokens>,generated_tokens=<generated tokens>"
```
|