https://www.kaggle.com/c/jigsaw-toxic-severity-rating/submissions
Private Leaderboard: 0.80831(23/2302)
-
Modified Training Pairs
- 1:10 (1:5 hard neg samples)
-
10 folds
-
3 epochs -> best
-
64 batch_size
-
Adafactor
- eps=(1e-30, 1e-3)
- clip_threshold=1.0
- decay_rate=-0.8
- weight_decay=1e-6
- relative_step=False
- scale_parameter=True
- warmup_init=False
-
lr 3e-4 -> 1e-5
- T_max 3000
- CosineAnnealingLR
-
eval & save interval 1000
Input files:
- ../input/jigsaw-toxic-severity-rating - current competition data
- ../input/jigsaw-toxic-comment-classification-challenge - 2017 competition data "The problem of classification of toxic comments"
- ../input/roberta-base - model data roberta base [for tokenizer]
- ../input/ruddit-jigsaw-dataset - Norms of Offensiveness for English Reddit Comments is a dataset of English language Reddit comments
- Roberta checkpoint
-
- Data1: jigsaw-toxic-comment-classification-challenge
- Validation(2021 data)
0.664
- LB
0.768
-
- Data1: jigsaw-toxic-comment-classification-challenge
- Validation(2021 data) :
0.684
- Validation(2021 data) :
- Data2: jigsaw-unintended-bias-in-toxicity-classification
- Validation(2021 data):
0.684
- Validation(2021 data):
- Data3: ruddit-jigsaw-dataset
- Validation(2021 data):
0.647
- Validation(2021 data):
- Ensemble
- Validation(2021 data)
0.692
- LB
0.830
- Validation(2021 data)
- Data1: jigsaw-toxic-comment-classification-challenge
notes:
why not save sklearn model? can save about 1 hour for submission--> sklearn joblib
-
Train only on 2021 data
-
Others' pretrained checkpoint
notes:
about 30min for 1fold inference
[Best Score 0.856] Jigsaw for beginners, Jigsaw Ensemble [TFIDF+BERT]
- use mutli loss: rank & score -> toxic
- filter non-english