bert finetune:When the training loss is unstable(decreasing), the validation loss has already started to rise
I am currently fine-tuning BERT using the bert-base-uncased model from Hugging Face. I have set the learning rate for the classification head to 5e-3 and for the other parameters to 5e-5. However, my training loss looks very strange, as shown in the attached figure. At the same time, my validation loss has started to increase. Why is this happening? I have tried using warm-up and various schedulers, but they don’t seem to have much effect. I urgently need help.[enter image description here](https://i.sstatic.net/2FSu3ZM6.png)