generated from ArneBinder/pytorch-ie-hydra-template-1
-
Notifications
You must be signed in to change notification settings - Fork 0
Open
Description
general
all the experiments should be executed with the multi-model variant of the individual task models
- target-model only
- frozen pre-trained target-model + bert-base-cased
- frozen pre-trained target-model + frozen other-task-model
- pre-trained target-model + frozen other-task-model
Hyperparameters:
- learning rate
- training time: start with a lot of epochs (50?); early stopping (see
patienceparameter)? - warming: for later (seems interesting, but requires effort)
Co-ref
- target-model only:
- frozen pre-trained target-model + bert-base-cased: config, W&B run, results
- frozen pre-trained target-model + frozen other-task-model:
- pre-trained target-model + frozen other-task-model:
EDIT:
- PR that adds respective configs and updates the
log.md: Coreference experiments #64 - experimental results: log.md#2023-09-28
- wandb report with the val/f1 and val/loss graphs (experiments from 2023-09-28 and 2023-09-29): https://wandb.ai/tanikina/conll2012-multi_model_coref_hoi-training/reports/Coreference-Experiments--Vmlldzo1NjAwNTMy
NER
- target-model only
- frozen pre-trained target-model + bert-base-cased
- frozen pre-trained target-model + frozen other-task-model
- pre-trained target-model + frozen other-task-model
PR that adds the NER configs and update the experiment log: #71
RE
- target-model only: config, W&B run, results(TODO @ArneBinder)
- frozen pre-trained target-model + bert-base-cased: config, W&B run, results(TODO @ArneBinder)
- frozen pre-trained target-model + frozen other-task-model: config W&B run results
- pre-trained target-model + frozen other-task-model: config W&B run results
Sanity Checks
- frozen pre-trained target-model with mean "aggregation" config + "model.aggregate=mean" W&B run results
- frozen pre-trained target-model with attention-based "aggregation" config W&B run results
extractive QA
- target-model only
- frozen pre-trained target-model + bert-base-cased
- frozen pre-trained target-model + frozen other-task-model
- pre-trained target-model + frozen other-task-model
Weights & Biases project for EQA: https://wandb.ai/stalvars/dataset+squadv2-task+extractive_question_answering-training
Metadata
Metadata
Assignees
Labels
No labels