From 40e00bf90f8dd92a6ff86c0ab3fa8627489bee2b Mon Sep 17 00:00:00 2001 From: zhtjtcz <18377221@buaa.edu.cn> Date: Thu, 8 Aug 2024 11:06:09 +0800 Subject: [PATCH 1/2] fix path --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index 37fbff8..991b667 100644 --- a/README.md +++ b/README.md @@ -55,7 +55,7 @@ python test_osediff.py \ -i preset/datasets/test_dataset/input \ -o preset/datasets/test_dataset/output \ ---osediff_path preset\models\osediff.pkl \ +--osediff_path preset/models/osediff.pkl \ --pretrained_model_name_or_path SD21BASE_PATH \ --ram_ft_path DAPE_PATH \ --ram_path RAM_PATH From 63b8343e003e0d0f847e5e47f8dcd98d598cbc1d Mon Sep 17 00:00:00 2001 From: zhtjtcz <18377221@buaa.edu.cn> Date: Thu, 8 Aug 2024 13:58:01 +0800 Subject: [PATCH 2/2] fix: BertTokenizer path --- ram/models/utils.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/ram/models/utils.py b/ram/models/utils.py index 292d161..446a4f8 100644 --- a/ram/models/utils.py +++ b/ram/models/utils.py @@ -129,7 +129,7 @@ def forward(self, x): def init_tokenizer(): # tokenizer = BertTokenizer.from_pretrained('bert-base-uncased') - tokenizer = BertTokenizer.from_pretrained('/home/notebook/data/group/LowLevelLLM/LLM/bert-base-uncased', local_files_only=True) + tokenizer = BertTokenizer.from_pretrained('google-bert/bert-base-uncased') tokenizer.add_special_tokens({'bos_token': '[DEC]'}) tokenizer.add_special_tokens({'additional_special_tokens': ['[ENC]']}) tokenizer.enc_token_id = tokenizer.additional_special_tokens_ids[0] @@ -362,4 +362,4 @@ def forward(self, x, y): torch.set_grad_enabled(True) loss *= one_sided_w - return -loss.sum() \ No newline at end of file + return -loss.sum()