/Library/Frameworks/Python.framework/Versions/3.11/lib/python3.11/site-packages/transformers/tokenization_utils_base.py:1601: FutureWarning: `clean_up_tokenization_spaces` was not set. It will be set to `True` by default. This behavior will be depracted in transformers v4.45, and will be then set to `False` by default. For more details check this issue: https://github.com/huggingface/transformers/issues/31884 | |
warnings.warn( | |
Some weights of DistilBertForSequenceClassification were not initialized from the model checkpoint at distilbert/distilbert-base-uncased and are newly initialized: ['classifier.bias', 'classifier.weight', 'pre_classifier.bias', 'pre_classifier.weight'] | |
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference. | |
[34m[1mwandb[39m[22m: [33mWARNING[39m The `run_name` is currently set to the same value as `TrainingArguments.output_dir`. If this was not intended, please specify a different run name by setting the `TrainingArguments.run_name` parameter. | |
{'loss': 0.3343, 'grad_norm': 13.296141624450684, 'learning_rate': 1.872040946896993e-05, 'epoch': 0.32} | |
{'loss': 0.2518, 'grad_norm': 9.50092887878418, 'learning_rate': 1.744081893793986e-05, 'epoch': 0.64} | |
{'loss': 0.2252, 'grad_norm': 15.53085994720459, 'learning_rate': 1.616122840690979e-05, 'epoch': 0.96} | |