from huggingface_hub import upload_folder upload_folder( folder_path="/home/yiren/new_ssd2/chunhui/yaning/project/mission-impossible-language-models", repo_id="Yaning1001/impossible_llm", path_in_repo="mission-impossible-language-models" ) # import torch # from transformers import AutoModelForCausalLM, AutoTokenizer # model_id_1 = "meta-llama/Llama-3.2-3B" # model_id_2 = "Qwen/Qwen2.5-7B" # # Check your authentication - this line should succeed without errors! # tokenizer_0 = AutoTokenizer.from_pretrained('gpt2') # tokenizer_1 = AutoTokenizer.from_pretrained(model_id_1) # tokenizer_2 = AutoTokenizer.from_pretrained(model_id_2) # tokenizer_2.pad_token = tokenizer_1.pad_token # print("tokenizer_0.pad_token:", type(tokenizer_0.pad_token)) # print("tokenizer_1.pad_token:", type(tokenizer_1.pad_token)) # print("tokenizer_2.pad_token:", type(tokenizer_2.pad_token))