hf roberta base regression layer on top

This commit is contained in:
Jakub Pokrywka 2022-07-04 18:31:40 +00:00
parent 33e11dad3d
commit 9af4dd453e
4 changed files with 38218 additions and 38218 deletions

File diff suppressed because it is too large Load Diff

View File

@ -9,7 +9,7 @@ from transformers import get_scheduler
import torch import torch
from tqdm.auto import tqdm from tqdm.auto import tqdm
BATCH_SIZE = 1 BATCH_SIZE = 24
EARLY_STOPPING = 3 EARLY_STOPPING = 3
WARMUP_STEPS = 10_000 WARMUP_STEPS = 10_000
@ -38,7 +38,7 @@ model = AutoModelForSequenceClassification.from_pretrained(MODEL, num_labels=1)
optimizer = AdamW(model.parameters(), lr=1e-6) optimizer = AdamW(model.parameters(), lr=1e-6)
num_epochs = 1 num_epochs = 15
num_training_steps = num_epochs * len(train_dataloader) num_training_steps = num_epochs * len(train_dataloader)
lr_scheduler = get_scheduler( lr_scheduler = get_scheduler(
"linear", "linear",

View File

@ -1,4 +1,4 @@
#MODEL = '/home/wmi/RoBERTa/without_date/checkpoint-1325000' #MODEL = '/home/wmi/RoBERTa/without_date/checkpoint-1325000'
MODEL = 'roberta-base' MODEL = 'roberta-base'
TEST=True TEST=False

File diff suppressed because it is too large Load Diff