File size: 888 Bytes
81dc001
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
from huggingface_hub import upload_folder

upload_folder(
    folder_path="/home/yiren/new_ssd2/chunhui/yaning/project/mission-impossible-language-models",
    repo_id="Yaning1001/impossible_llm",
    path_in_repo="mission-impossible-language-models"
)

# import torch
# from transformers import AutoModelForCausalLM, AutoTokenizer

# model_id_1 = "meta-llama/Llama-3.2-3B"
# model_id_2 = "Qwen/Qwen2.5-7B"

# # Check your authentication - this line should succeed without errors!
# tokenizer_0 = AutoTokenizer.from_pretrained('gpt2')
# tokenizer_1 = AutoTokenizer.from_pretrained(model_id_1)
# tokenizer_2 = AutoTokenizer.from_pretrained(model_id_2)
# tokenizer_2.pad_token = tokenizer_1.pad_token

# print("tokenizer_0.pad_token:", type(tokenizer_0.pad_token))
# print("tokenizer_1.pad_token:", type(tokenizer_1.pad_token))
# print("tokenizer_2.pad_token:", type(tokenizer_2.pad_token))