mattshumer commited on
Commit
41417f6
1 Parent(s): 35c1e1e

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +63 -0
README.md ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: llama3.1
3
+ base_model: meta-llama/Meta-Llama-3.1-70B-Instruct
4
+ pipeline_tag: text-generation
5
+ library_name: transformers
6
+ ---
7
+ # Reflection Llama-3.1 70B
8
+
9
+ | IMPORTANT — This is the new, working version of the Reflection Llama 3.1 70B model. Please use this version.
10
+
11
+ **Reflection Llama-3.1 70B is (currently) the world's top open-source LLM, trained with a new technique called Reflection-Tuning that teaches a LLM to detect mistakes in its reasoning and correct course.**
12
+
13
+ The model was trained on synthetic data generated by [Glaive](https://glaive.ai). If you're training a model, Glaive is incredible — use them.
14
+
15
+ ## Benchmarks
16
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/60518f3731c5be7f3dd5ebc3/zNs-ZFs0SbnomH7mikiOU.png)
17
+
18
+ All benchmarks tested have been checked for contamination by running [LMSys's LLM Decontaminator](https://github.com/lm-sys/llm-decontaminator). When benchmarking, we isolate the `<output>` and benchmark on solely that section.
19
+
20
+ Trained from Llama 3.1 70B Instruct, you can sample from Reflection Llama-3.1 70B using the same code, pipelines, etc. as any other Llama model. It even uses the stock Llama 3.1 chat template format (though, we've trained in a few new special tokens to aid in reasoning and reflection).
21
+
22
+ During sampling, the model will start by outputting reasoning inside `<thinking>` and `</thinking>` tags, and then once it is satisfied with its reasoning, it will output the final answer inside `<output>` and `</output>` tags. Each of these tags are special tokens, trained into the model.
23
+
24
+ This enables the model to separate its internal thoughts and reasoning from its final answer, improving the experience for the user.
25
+
26
+ Inside the `<thinking>` section, the model may output one or more `<reflection>` tags, which signals the model has caught an error in its reasoning and will attempt to correct it before providing a final answer.
27
+
28
+ ## System Prompt
29
+
30
+ The system prompt used for training this model is:
31
+
32
+ ```
33
+ You are a world-class AI system, capable of complex reasoning and reflection. Reason through the query inside <thinking> tags, and then provide your final response inside <output> tags. If you detect that you made a mistake in your reasoning at any point, correct yourself inside <reflection> tags.
34
+ ```
35
+
36
+ We recommend using this exact system prompt to get the best results from Reflection Llama-3.1 70B. You may also want to experiment combining this system prompt with your own custom instructions to customize the behavior of the model.
37
+
38
+ ## Chat Format
39
+
40
+ As mentioned above, the model uses the standard Llama 3.1 chat format. Here’s an example:
41
+
42
+ ```
43
+ <|begin_of_text|><|start_header_id|>system<|end_header_id|>
44
+
45
+ You are a world-class AI system, capable of complex reasoning and reflection. Reason through the query inside <thinking> tags, and then provide your final response inside <output> tags. If you detect that you made a mistake in your reasoning at any point, correct yourself inside <reflection> tags.<|eot_id|><|start_header_id|>user<|end_header_id|>
46
+
47
+ what is 2+2?<|eot_id|><|start_header_id|>assistant<|end_header_id|>
48
+ ```
49
+
50
+ ## Tips for Performance
51
+
52
+ - We are initially recommending a `temperature` of `.7` and a `top_p` of `.95`.
53
+ - For increased accuracy, append `Think carefully.` at the end of your messages.
54
+
55
+ ## Dataset / Report
56
+
57
+ Both the dataset and a brief report detailing how we trained this model will be released next week, alongside our Reflection 405B model that we expect will be the top-performing LLM in the world, including closed-source models.
58
+
59
+ ---
60
+
61
+ Thanks to Jason Kuperberg and Josh Bickett from the [HyperWrite](https://hyperwriteai.com) team for reviewing drafts of the report we'll be releasing next week.
62
+
63
+ Also, we know right now the model is split into a ton of files. We'll condense this soon to make the model easier to download and work with!