File size: 3,908 Bytes
79dfd35
 
258f5fe
79dfd35
 
 
 
 
 
 
 
 
 
b6dfb28
79dfd35
b6dfb28
 
 
 
 
 
 
 
 
 
 
 
79dfd35
 
 
 
 
 
b6dfb28
79dfd35
b6dfb28
79dfd35
 
 
 
 
 
 
 
 
3aa8f73
79dfd35
 
 
3aa8f73
79dfd35
 
 
3aa8f73
79dfd35
 
 
 
3aa8f73
79dfd35
b6dfb28
79dfd35
b6dfb28
79dfd35
b6dfb28
 
79dfd35
b6dfb28
79dfd35
b6dfb28
79dfd35
b6dfb28
79dfd35
b6dfb28
 
 
79dfd35
b6dfb28
79dfd35
b6dfb28
79dfd35
b6dfb28
79dfd35
b6dfb28
79dfd35
b6dfb28
79dfd35
b6dfb28
 
79dfd35
b6dfb28
79dfd35
b6dfb28
79dfd35
b6dfb28
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
258f5fe
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
---
library_name: transformers
license: apache-2.0
---

# Model Card for Model ID

<!-- Provide a quick summary of what the model is/does. -->



## Model Details

This model is being made to enhance our work within crewai. We started with a high context length (1048K) version of Llama 3. We then fine-tuned on top of that to get a base agent.

## Model Description

Built on the following:
- Models:
  - [Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct)
  - [Llama-3-8B-Instruct-Gradient-1048k](https://huggingface.co/gradientai/Llama-3-8B-Instruct-Gradient-1048k)
- Datasets:
  - m-a-p/CodeFeedback-Filtered-Instruction
  - RomanTeucher/awesome_topic_code_snippets
  - dair-ai/emotion
  - mzbac/function-calling-llama-3-format-v1.1
  - gretelai/synthetic_text_to_sql

This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.

- **Developed by:** [More Information Needed]
- **Model type:** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [Llama-3-8B-Instruct-Gradient-1048k]

## Model Sources [optional]

<!-- Provide the basic links for the model. -->

- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]

## Uses

Use inside with crewai to get better results and to generaete starting data.

### Direct Use

Use as a chat bot, ai agent, etc.

### Out-of-Scope Use

Anything outside of using with crewai while vary possiable will be out of scope.


### Recommendations

Self host or host this model in the cloud and use with crew ai for best results.

## Code Examples

Use the following format when using the model for inference:

```
<|begin_of_text|><|start_header_id|>system<|end_header_id|>

You are the helpful assistant. <|eot_id|><|start_header_id|>user<|end_header_id|>

{prompt} <|eot_id|><|start_header_id|>assistant<|end_header_id|>

```

Example of multi turn
```
<|begin_of_text|><|start_header_id|>system<|end_header_id|>

You are a helpful AI assistant for travel tips and recommendations<|eot_id|><|start_header_id|>user<|end_header_id|>

What is France's capital?<|eot_id|><|start_header_id|>assistant<|end_header_id|>

Bonjour! The capital of France is Paris!<|eot_id|><|start_header_id|>user<|end_header_id|>

What can I do there?<|eot_id|><|start_header_id|>assistant<|end_header_id|>

Paris, the City of Light, offers a romantic getaway with must-see attractions like the Eiffel Tower and Louvre Museum, romantic experiences like river cruises and charming neighborhoods, and delicious food and drink options, with helpful tips for making the most of your trip.<|eot_id|><|start_header_id|>user<|end_header_id|>

Give me a detailed list of the attractions I should visit, and time it takes in each one, to plan my trip accordingly.<|eot_id|><|start_header_id|>assistant<|end_header_id|>
```

## How to Get Started with the Model

Use the code below to get started with the model.

```
import transformers
import torch

model_id = "meta-llama/Meta-Llama-3-70B-Instruct"

pipeline = transformers.pipeline(
    "text-generation",
    model=model_id,
    model_kwargs={"torch_dtype": torch.bfloat16},
    device_map="auto",
)

messages = [
    {"role": "system", "content": "You are a pirate chatbot who always responds in pirate speak!"},
    {"role": "user", "content": "Who are you?"},
]

prompt = pipeline.tokenizer.apply_chat_template(
        messages, 
        tokenize=False, 
        add_generation_prompt=True
)

terminators = [
    pipeline.tokenizer.eos_token_id,
    pipeline.tokenizer.convert_tokens_to_ids("<|eot_id|>")
]

outputs = pipeline(
    prompt,
    max_new_tokens=256,
    eos_token_id=terminators,
    do_sample=True,
    temperature=0.6,
    top_p=0.9,
)
print(outputs[0]["generated_text"][len(prompt):])

```