From a1d411d38b36a88a8eb7d0f300dabe6766771270 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=D0=9C=D0=B0=D0=B7=D1=83=D1=80=20=D0=93=D1=80=D0=B5=D1=82?= =?UTF-8?q?=D0=B0=20=D0=95=D0=B2=D0=B3=D0=B5=D0=BD=D1=8C=D0=B5=D0=B2=D0=BD?= =?UTF-8?q?=D0=B0?= <gemazur_1@edu.hse.ru> Date: Tue, 25 Mar 2025 22:13:35 +0300 Subject: [PATCH] micro zapusk no cross --- .ipynb_checkpoints/checkLora-checkpoint.py | 8 ++++---- checkLora.py | 8 ++++---- 2 files changed, 8 insertions(+), 8 deletions(-) diff --git a/.ipynb_checkpoints/checkLora-checkpoint.py b/.ipynb_checkpoints/checkLora-checkpoint.py index 9f4d793..4777d1d 100644 --- a/.ipynb_checkpoints/checkLora-checkpoint.py +++ b/.ipynb_checkpoints/checkLora-checkpoint.py @@ -312,11 +312,11 @@ tokenizer = BertTokenizer.from_pretrained("./micro_no_cross_fine_tuned2") +# base_model = MultiTaskBert.from_pretrained('bert-base-uncased').to(device) +# print("Загружена базовая") -base_model = MultiTaskBert.from_pretrained('bert-base-uncased').to(device) -print("Загружена базовая") - -model = PeftModel.from_pretrained(base_model, "./micro_no_cross_fine_tuned/lora2", strict=False) +# model = PeftModel.from_pretrained(base_model, "./micro_no_cross_fine_tuned/lora2", strict=False) +model = PeftModel.from_pretrained("./micro_no_cross_fine_tuned/lora2") print("Загружена с лора") diff --git a/checkLora.py b/checkLora.py index 9f4d793..4777d1d 100644 --- a/checkLora.py +++ b/checkLora.py @@ -312,11 +312,11 @@ tokenizer = BertTokenizer.from_pretrained("./micro_no_cross_fine_tuned2") +# base_model = MultiTaskBert.from_pretrained('bert-base-uncased').to(device) +# print("Загружена базовая") -base_model = MultiTaskBert.from_pretrained('bert-base-uncased').to(device) -print("Загружена базовая") - -model = PeftModel.from_pretrained(base_model, "./micro_no_cross_fine_tuned/lora2", strict=False) +# model = PeftModel.from_pretrained(base_model, "./micro_no_cross_fine_tuned/lora2", strict=False) +model = PeftModel.from_pretrained("./micro_no_cross_fine_tuned/lora2") print("Загружена с лора") -- GitLab