readme
This commit is contained in:
parent
97754985fc
commit
99b93a2579
14
README.md
Normal file
14
README.md
Normal file
@ -0,0 +1,14 @@
|
|||||||
|
# Transformer Encoder - RoBERTa
|
||||||
|
## Modyfikacje
|
||||||
|
1. ????
|
||||||
|
|
||||||
|
# Transformer Decoder - GPT-2
|
||||||
|
## Modyfikacje
|
||||||
|
1. ????
|
||||||
|
|
||||||
|
|
||||||
|
# Transformer Encoder-Decoder - T5
|
||||||
|
## Modyfikacje
|
||||||
|
1. Zamrożenie pierwszych 20 warstw
|
||||||
|
|
||||||
|
# Transformer w trybie few-shot/zero-shot learning - ?????
|
@ -260,8 +260,12 @@ class DataTrainingArguments:
|
|||||||
def freeze_model_weights(model: torch.nn.Module) -> None:
|
def freeze_model_weights(model: torch.nn.Module) -> None:
|
||||||
count = 0
|
count = 0
|
||||||
for param in model.parameters():
|
for param in model.parameters():
|
||||||
logger.info(count)
|
count += 1
|
||||||
|
if count < 20:
|
||||||
|
logger.info(f'Freezing layer {count}')
|
||||||
param.requires_grad = False
|
param.requires_grad = False
|
||||||
|
else:
|
||||||
|
logger.info(f'Ignoring layer {count}')
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
|
Loading…
Reference in New Issue
Block a user