|
LANG=java |
|
DATADIR=../dataset/javaCorpus/token_completion |
|
LITFILE=../dataset/javaCorpus/literals.json |
|
OUTPUTDIR=../model/javaCorpus |
|
PRETRAINDIR=microsoft/CodeGPT-small-java |
|
LOGFILE=train_javaCorpus.log |
|
PER_NODE_GPU=4 |
|
|
|
CUDA_VISIBLE_DEVICES=0,1,2,3 python run_lm.py \ |
|
--data_dir=$DATADIR \ |
|
--lit_file=$LITFILE \ |
|
--langs=$LANG \ |
|
--output_dir=$OUTPUTDIR \ |
|
--pretrain_dir=$PRETRAINDIR \ |
|
--log_file=$LOGFILE \ |
|
--model_type=gpt2 \ |
|
--block_size=512 \ |
|
--do_train \ |
|
--gpu_per_node $PER_NODE_GPU \ |
|
--learning_rate=8e-5 \ |
|
--weight_decay=0.01 \ |
|
--evaluate_during_training \ |
|
--per_gpu_train_batch_size=1 \ |
|
--per_gpu_eval_batch_size=4 \ |
|
--gradient_accumulation_steps=4 \ |
|
--num_train_epochs=5 \ |
|
--logging_steps=100 \ |
|
--save_steps=1000 \ |
|
--seed=42 \ |
|
--overwrite_output_dir \ |
|
--not_pretrain |