This model is training on a HPC platform using 5 GeForce RTX 3090. Resource usage summary:

CPU time :                                   150541.41 sec.
Max Memory :                                 11148 MB
Average Memory :                             8411.42 MB
Total Requested Memory :                     -
Delta Memory :                               -
Max Swap :                                   -
Max Processes :                              4
Max Threads :                                195
Run time :                                   67735 sec.
Turnaround time :                            67774 sec.

Early stopping due to no improvement in validation loss after 7 epoch. Accuracy: 0.9709595558622203


<aside> 📢 More about ‘Chinese BERT with Whole Word Masking’

</aside>

The finetune model is based on hfl/chinese-bert-wwm:

https://github.com/ymcui/Chinese-BERT-wwm

hfl/chinese-bert-wwm · Hugging Face


We explain the steps we take to finetune the pre-trained chinese-bert-wwm:

To classify job postings to the correct SOC occupation, we finetune a multi-class flat classification model based on a BERT model. We spotlight some of the key elements below. The full training code:

Job-Posting/flat_classification.ipynb at main · lzxlll/Job-Posting

1. Split the final estimation data into train, validation, and test sets as follows: