roberta_year_as_token_everywhere_better_finetunning
This commit is contained in:
parent
443895822e
commit
cede76239d
@ -5,7 +5,7 @@ from config import MODEL
|
||||
from tqdm import tqdm
|
||||
|
||||
dataset = load_dataset('csv', sep='\t', data_files={'train': ['../train/huggingface_format_year_clear.csv'], 'test': ['../dev-0/huggingface_format_year_clear.csv']})
|
||||
test_dataset = load_dataset('csv', sep='\t', data_files='../test-A/huggingface_format_year.csv')
|
||||
test_dataset = load_dataset('csv', sep='\t', data_files='../test-A/huggingface_format_year_clear.csv')
|
||||
|
||||
tokenizer = AutoTokenizer.from_pretrained(MODEL)
|
||||
|
||||
|
93262
test-A/out.tsv
93262
test-A/out.tsv
File diff suppressed because it is too large
Load Diff
Loading…
Reference in New Issue
Block a user