digo-prayudha commited on
Commit
ed00d68
1 Parent(s): c26162c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +30 -14
README.md CHANGED
@@ -11,6 +11,9 @@ tags:
11
  model-index:
12
  - name: Llama-3.2-1B-Indonesian
13
  results: []
 
 
 
14
  ---
15
 
16
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -18,21 +21,34 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  # Llama-3.2-1B-Indonesian
20
 
21
- This model is a fine-tuned version of [meta-llama/Llama-3.2-1B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct) on the generator dataset.
22
-
23
- ## Model description
24
-
25
- More information needed
26
-
27
- ## Intended uses & limitations
28
-
29
- More information needed
30
 
31
  ## Training and evaluation data
32
 
33
- More information needed
34
-
35
- ## Training procedure
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
36
 
37
  ### Training hyperparameters
38
 
@@ -50,8 +66,8 @@ The following hyperparameters were used during training:
50
  - mixed_precision_training: Native AMP
51
 
52
  ### Training results
53
-
54
-
55
 
56
  ### Framework versions
57
 
 
11
  model-index:
12
  - name: Llama-3.2-1B-Indonesian
13
  results: []
14
+ language:
15
+ - id
16
+ pipeline_tag: text-generation
17
  ---
18
 
19
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
21
 
22
  # Llama-3.2-1B-Indonesian
23
 
24
+ This model is a fine-tuned version of [meta-llama/Llama-3.2-1B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct) that has been optimized for Indonesian language understanding and generation..
 
 
 
 
 
 
 
 
25
 
26
  ## Training and evaluation data
27
 
28
+ [Ichsan2895/alpaca-gpt4-indonesian](https://huggingface.co/datasets/Ichsan2895/alpaca-gpt4-indonesian)
29
+
30
+ ### Use WIth Transformers
31
+ ```python
32
+ import torch
33
+ from transformers import pipeline
34
+
35
+ model_id = "meta-llama/Llama-3.2-3B-Instruct"
36
+ pipe = pipeline(
37
+ "text-generation",
38
+ model=model_id,
39
+ torch_dtype=torch.bfloat16,
40
+ device_map="auto",
41
+ )
42
+ messages = [
43
+ {"role": "system", "content": "You are a pirate chatbot who always responds in pirate speak!"},
44
+ {"role": "user", "content": "Who are you?"},
45
+ ]
46
+ outputs = pipe(
47
+ messages,
48
+ max_new_tokens=256,
49
+ )
50
+ print(outputs[0]["generated_text"][-1])
51
+ ```
52
 
53
  ### Training hyperparameters
54
 
 
66
  - mixed_precision_training: Native AMP
67
 
68
  ### Training results
69
+ ![Train Loss]
70
+ <img src="./train_loss.svg" width="300" height="200">
71
 
72
  ### Framework versions
73