roberta_year_as_token_everywhere_better_finetunning
This commit is contained in:
parent
443895822e
commit
cede76239d
@ -5,7 +5,7 @@ from config import MODEL
|
|||||||
from tqdm import tqdm
|
from tqdm import tqdm
|
||||||
|
|
||||||
dataset = load_dataset('csv', sep='\t', data_files={'train': ['../train/huggingface_format_year_clear.csv'], 'test': ['../dev-0/huggingface_format_year_clear.csv']})
|
dataset = load_dataset('csv', sep='\t', data_files={'train': ['../train/huggingface_format_year_clear.csv'], 'test': ['../dev-0/huggingface_format_year_clear.csv']})
|
||||||
test_dataset = load_dataset('csv', sep='\t', data_files='../test-A/huggingface_format_year.csv')
|
test_dataset = load_dataset('csv', sep='\t', data_files='../test-A/huggingface_format_year_clear.csv')
|
||||||
|
|
||||||
tokenizer = AutoTokenizer.from_pretrained(MODEL)
|
tokenizer = AutoTokenizer.from_pretrained(MODEL)
|
||||||
|
|
||||||
|
93262
test-A/out.tsv
93262
test-A/out.tsv
File diff suppressed because it is too large
Load Diff
Loading…
Reference in New Issue
Block a user