02/17/2024 15:12:04 - INFO - __main__ - device: cuda, n_gpu: 2
02/17/2024 15:12:08 - INFO - __main__ - new token {'additional_special_tokens': ['global_variable', 'heredoc_end', 'decimal_floating_point_literal', 'rune_literal', 'int_literal', 'ERROR', 'class', 'heredoc_content', 'field_identifier', 'name', 'string', 'hash_key_symbol', 'hex_integer_literal', 'statement_identifier', 'boolean', 'separators', 'escape_sequence', 'boolean_type', 'regex_flags', 'string_fragment', 'identifier', 'instance_variable', 'regex_pattern', 'decimal_integer_literal', 'raw_string_literal', 'property_identifier', 'operator', 'label_name', 'namespace', 'string_literal', 'package_identifier', 'float_literal', 'integer', 'php_tag', 'shorthand_property_identifier', 'shorthand_property_identifier_pattern', 'extends', 'none', 'text', 'void_type', 'null_literal', 'heredoc_beginning', 'keyword', 'simple_symbol', 'type_identifier', 'character_literal', 'string_content', 'comment', 'number', '"', 'constant', 'class_variable']}
02/17/2024 15:12:09 - INFO - __main__ - +-------------------------------------------------------------------+--------------+----------+
| Layer Name | Output Shape | Param # |
+-------------------------------------------------------------------+--------------+----------+
| code_encoder_q.embeddings.word_embeddings.weight | [51451, 768] | 39514368 |
| code_encoder_q.embeddings.position_embeddings.weight | [1026, 768] | 787968 |
| code_encoder_q.embeddings.token_type_embeddings.weight | [10, 768] | 7680 |
| code_encoder_q.embeddings.LayerNorm.weight | [768] | 768 |
| code_encoder_q.embeddings.LayerNorm.bias | [768] | 768 |
| code_encoder_q.encoder.layer.0.attention.self.query.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.0.attention.self.query.bias | [768] | 768 |
| code_encoder_q.encoder.layer.0.attention.self.key.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.0.attention.self.key.bias | [768] | 768 |
| code_encoder_q.encoder.layer.0.attention.self.value.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.0.attention.self.value.bias | [768] | 768 |
| code_encoder_q.encoder.layer.0.attention.output.dense.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.0.attention.output.dense.bias | [768] | 768 |
| code_encoder_q.encoder.layer.0.attention.output.LayerNorm.weight | [768] | 768 |
| code_encoder_q.encoder.layer.0.attention.output.LayerNorm.bias | [768] | 768 |
| code_encoder_q.encoder.layer.0.intermediate.dense.weight | [3072, 768] | 2359296 |
| code_encoder_q.encoder.layer.0.intermediate.dense.bias | [3072] | 3072 |
| code_encoder_q.encoder.layer.0.output.dense.weight | [768, 3072] | 2359296 |
| code_encoder_q.encoder.layer.0.output.dense.bias | [768] | 768 |
| code_encoder_q.encoder.layer.0.output.LayerNorm.weight | [768] | 768 |
| code_encoder_q.encoder.layer.0.output.LayerNorm.bias | [768] | 768 |
| code_encoder_q.encoder.layer.1.attention.self.query.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.1.attention.self.query.bias | [768] | 768 |
| code_encoder_q.encoder.layer.1.attention.self.key.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.1.attention.self.key.bias | [768] | 768 |
| code_encoder_q.encoder.layer.1.attention.self.value.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.1.attention.self.value.bias | [768] | 768 |
| code_encoder_q.encoder.layer.1.attention.output.dense.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.1.attention.output.dense.bias | [768] | 768 |
| code_encoder_q.encoder.layer.1.attention.output.LayerNorm.weight | [768] | 768 |
| code_encoder_q.encoder.layer.1.attention.output.LayerNorm.bias | [768] | 768 |
| code_encoder_q.encoder.layer.1.intermediate.dense.weight | [3072, 768] | 2359296 |
| code_encoder_q.encoder.layer.1.intermediate.dense.bias | [3072] | 3072 |
| code_encoder_q.encoder.layer.1.output.dense.weight | [768, 3072] | 2359296 |
| code_encoder_q.encoder.layer.1.output.dense.bias | [768] | 768 |
| code_encoder_q.encoder.layer.1.output.LayerNorm.weight | [768] | 768 |
| code_encoder_q.encoder.layer.1.output.LayerNorm.bias | [768] | 768 |
| code_encoder_q.encoder.layer.2.attention.self.query.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.2.attention.self.query.bias | [768] | 768 |
| code_encoder_q.encoder.layer.2.attention.self.key.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.2.attention.self.key.bias | [768] | 768 |
| code_encoder_q.encoder.layer.2.attention.self.value.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.2.attention.self.value.bias | [768] | 768 |
| code_encoder_q.encoder.layer.2.attention.output.dense.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.2.attention.output.dense.bias | [768] | 768 |
| code_encoder_q.encoder.layer.2.attention.output.LayerNorm.weight | [768] | 768 |
| code_encoder_q.encoder.layer.2.attention.output.LayerNorm.bias | [768] | 768 |
| code_encoder_q.encoder.layer.2.intermediate.dense.weight | [3072, 768] | 2359296 |
| code_encoder_q.encoder.layer.2.intermediate.dense.bias | [3072] | 3072 |
| code_encoder_q.encoder.layer.2.output.dense.weight | [768, 3072] | 2359296 |
| code_encoder_q.encoder.layer.2.output.dense.bias | [768] | 768 |
| code_encoder_q.encoder.layer.2.output.LayerNorm.weight | [768] | 768 |
| code_encoder_q.encoder.layer.2.output.LayerNorm.bias | [768] | 768 |
| code_encoder_q.encoder.layer.3.attention.self.query.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.3.attention.self.query.bias | [768] | 768 |
| code_encoder_q.encoder.layer.3.attention.self.key.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.3.attention.self.key.bias | [768] | 768 |
| code_encoder_q.encoder.layer.3.attention.self.value.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.3.attention.self.value.bias | [768] | 768 |
| code_encoder_q.encoder.layer.3.attention.output.dense.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.3.attention.output.dense.bias | [768] | 768 |
| code_encoder_q.encoder.layer.3.attention.output.LayerNorm.weight | [768] | 768 |
| code_encoder_q.encoder.layer.3.attention.output.LayerNorm.bias | [768] | 768 |
| code_encoder_q.encoder.layer.3.intermediate.dense.weight | [3072, 768] | 2359296 |
| code_encoder_q.encoder.layer.3.intermediate.dense.bias | [3072] | 3072 |
| code_encoder_q.encoder.layer.3.output.dense.weight | [768, 3072] | 2359296 |
| code_encoder_q.encoder.layer.3.output.dense.bias | [768] | 768 |
| code_encoder_q.encoder.layer.3.output.LayerNorm.weight | [768] | 768 |
| code_encoder_q.encoder.layer.3.output.LayerNorm.bias | [768] | 768 |
| code_encoder_q.encoder.layer.4.attention.self.query.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.4.attention.self.query.bias | [768] | 768 |
| code_encoder_q.encoder.layer.4.attention.self.key.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.4.attention.self.key.bias | [768] | 768 |
| code_encoder_q.encoder.layer.4.attention.self.value.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.4.attention.self.value.bias | [768] | 768 |
| code_encoder_q.encoder.layer.4.attention.output.dense.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.4.attention.output.dense.bias | [768] | 768 |
| code_encoder_q.encoder.layer.4.attention.output.LayerNorm.weight | [768] | 768 |
| code_encoder_q.encoder.layer.4.attention.output.LayerNorm.bias | [768] | 768 |
| code_encoder_q.encoder.layer.4.intermediate.dense.weight | [3072, 768] | 2359296 |
| code_encoder_q.encoder.layer.4.intermediate.dense.bias | [3072] | 3072 |
| code_encoder_q.encoder.layer.4.output.dense.weight | [768, 3072] | 2359296 |
| code_encoder_q.encoder.layer.4.output.dense.bias | [768] | 768 |
| code_encoder_q.encoder.layer.4.output.LayerNorm.weight | [768] | 768 |
| code_encoder_q.encoder.layer.4.output.LayerNorm.bias | [768] | 768 |
| code_encoder_q.encoder.layer.5.attention.self.query.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.5.attention.self.query.bias | [768] | 768 |
| code_encoder_q.encoder.layer.5.attention.self.key.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.5.attention.self.key.bias | [768] | 768 |
| code_encoder_q.encoder.layer.5.attention.self.value.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.5.attention.self.value.bias | [768] | 768 |
| code_encoder_q.encoder.layer.5.attention.output.dense.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.5.attention.output.dense.bias | [768] | 768 |
| code_encoder_q.encoder.layer.5.attention.output.LayerNorm.weight | [768] | 768 |
| code_encoder_q.encoder.layer.5.attention.output.LayerNorm.bias | [768] | 768 |
| code_encoder_q.encoder.layer.5.intermediate.dense.weight | [3072, 768] | 2359296 |
| code_encoder_q.encoder.layer.5.intermediate.dense.bias | [3072] | 3072 |
| code_encoder_q.encoder.layer.5.output.dense.weight | [768, 3072] | 2359296 |
| code_encoder_q.encoder.layer.5.output.dense.bias | [768] | 768 |
| code_encoder_q.encoder.layer.5.output.LayerNorm.weight | [768] | 768 |
| code_encoder_q.encoder.layer.5.output.LayerNorm.bias | [768] | 768 |
| code_encoder_q.encoder.layer.6.attention.self.query.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.6.attention.self.query.bias | [768] | 768 |
| code_encoder_q.encoder.layer.6.attention.self.key.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.6.attention.self.key.bias | [768] | 768 |
| code_encoder_q.encoder.layer.6.attention.self.value.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.6.attention.self.value.bias | [768] | 768 |
| code_encoder_q.encoder.layer.6.attention.output.dense.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.6.attention.output.dense.bias | [768] | 768 |
| code_encoder_q.encoder.layer.6.attention.output.LayerNorm.weight | [768] | 768 |
| code_encoder_q.encoder.layer.6.attention.output.LayerNorm.bias | [768] | 768 |
| code_encoder_q.encoder.layer.6.intermediate.dense.weight | [3072, 768] | 2359296 |
| code_encoder_q.encoder.layer.6.intermediate.dense.bias | [3072] | 3072 |
| code_encoder_q.encoder.layer.6.output.dense.weight | [768, 3072] | 2359296 |
| code_encoder_q.encoder.layer.6.output.dense.bias | [768] | 768 |
| code_encoder_q.encoder.layer.6.output.LayerNorm.weight | [768] | 768 |
| code_encoder_q.encoder.layer.6.output.LayerNorm.bias | [768] | 768 |
| code_encoder_q.encoder.layer.7.attention.self.query.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.7.attention.self.query.bias | [768] | 768 |
| code_encoder_q.encoder.layer.7.attention.self.key.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.7.attention.self.key.bias | [768] | 768 |
| code_encoder_q.encoder.layer.7.attention.self.value.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.7.attention.self.value.bias | [768] | 768 |
| code_encoder_q.encoder.layer.7.attention.output.dense.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.7.attention.output.dense.bias | [768] | 768 |
| code_encoder_q.encoder.layer.7.attention.output.LayerNorm.weight | [768] | 768 |
| code_encoder_q.encoder.layer.7.attention.output.LayerNorm.bias | [768] | 768 |
| code_encoder_q.encoder.layer.7.intermediate.dense.weight | [3072, 768] | 2359296 |
| code_encoder_q.encoder.layer.7.intermediate.dense.bias | [3072] | 3072 |
| code_encoder_q.encoder.layer.7.output.dense.weight | [768, 3072] | 2359296 |
| code_encoder_q.encoder.layer.7.output.dense.bias | [768] | 768 |
| code_encoder_q.encoder.layer.7.output.LayerNorm.weight | [768] | 768 |
| code_encoder_q.encoder.layer.7.output.LayerNorm.bias | [768] | 768 |
| code_encoder_q.encoder.layer.8.attention.self.query.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.8.attention.self.query.bias | [768] | 768 |
| code_encoder_q.encoder.layer.8.attention.self.key.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.8.attention.self.key.bias | [768] | 768 |
| code_encoder_q.encoder.layer.8.attention.self.value.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.8.attention.self.value.bias | [768] | 768 |
| code_encoder_q.encoder.layer.8.attention.output.dense.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.8.attention.output.dense.bias | [768] | 768 |
| code_encoder_q.encoder.layer.8.attention.output.LayerNorm.weight | [768] | 768 |
| code_encoder_q.encoder.layer.8.attention.output.LayerNorm.bias | [768] | 768 |
| code_encoder_q.encoder.layer.8.intermediate.dense.weight | [3072, 768] | 2359296 |
| code_encoder_q.encoder.layer.8.intermediate.dense.bias | [3072] | 3072 |
| code_encoder_q.encoder.layer.8.output.dense.weight | [768, 3072] | 2359296 |
| code_encoder_q.encoder.layer.8.output.dense.bias | [768] | 768 |
| code_encoder_q.encoder.layer.8.output.LayerNorm.weight | [768] | 768 |
| code_encoder_q.encoder.layer.8.output.LayerNorm.bias | [768] | 768 |
| code_encoder_q.encoder.layer.9.attention.self.query.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.9.attention.self.query.bias | [768] | 768 |
| code_encoder_q.encoder.layer.9.attention.self.key.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.9.attention.self.key.bias | [768] | 768 |
| code_encoder_q.encoder.layer.9.attention.self.value.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.9.attention.self.value.bias | [768] | 768 |
| code_encoder_q.encoder.layer.9.attention.output.dense.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.9.attention.output.dense.bias | [768] | 768 |
| code_encoder_q.encoder.layer.9.attention.output.LayerNorm.weight | [768] | 768 |
| code_encoder_q.encoder.layer.9.attention.output.LayerNorm.bias | [768] | 768 |
| code_encoder_q.encoder.layer.9.intermediate.dense.weight | [3072, 768] | 2359296 |
| code_encoder_q.encoder.layer.9.intermediate.dense.bias | [3072] | 3072 |
| code_encoder_q.encoder.layer.9.output.dense.weight | [768, 3072] | 2359296 |
| code_encoder_q.encoder.layer.9.output.dense.bias | [768] | 768 |
| code_encoder_q.encoder.layer.9.output.LayerNorm.weight | [768] | 768 |
| code_encoder_q.encoder.layer.9.output.LayerNorm.bias | [768] | 768 |
| code_encoder_q.encoder.layer.10.attention.self.query.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.10.attention.self.query.bias | [768] | 768 |
| code_encoder_q.encoder.layer.10.attention.self.key.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.10.attention.self.key.bias | [768] | 768 |
| code_encoder_q.encoder.layer.10.attention.self.value.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.10.attention.self.value.bias | [768] | 768 |
| code_encoder_q.encoder.layer.10.attention.output.dense.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.10.attention.output.dense.bias | [768] | 768 |
| code_encoder_q.encoder.layer.10.attention.output.LayerNorm.weight | [768] | 768 |
| code_encoder_q.encoder.layer.10.attention.output.LayerNorm.bias | [768] | 768 |
| code_encoder_q.encoder.layer.10.intermediate.dense.weight | [3072, 768] | 2359296 |
| code_encoder_q.encoder.layer.10.intermediate.dense.bias | [3072] | 3072 |
| code_encoder_q.encoder.layer.10.output.dense.weight | [768, 3072] | 2359296 |
| code_encoder_q.encoder.layer.10.output.dense.bias | [768] | 768 |
| code_encoder_q.encoder.layer.10.output.LayerNorm.weight | [768] | 768 |
| code_encoder_q.encoder.layer.10.output.LayerNorm.bias | [768] | 768 |
| code_encoder_q.encoder.layer.11.attention.self.query.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.11.attention.self.query.bias | [768] | 768 |
| code_encoder_q.encoder.layer.11.attention.self.key.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.11.attention.self.key.bias | [768] | 768 |
| code_encoder_q.encoder.layer.11.attention.self.value.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.11.attention.self.value.bias | [768] | 768 |
| code_encoder_q.encoder.layer.11.attention.output.dense.weight | [768, 768] | 589824 |
| code_encoder_q.encoder.layer.11.attention.output.dense.bias | [768] | 768 |
| code_encoder_q.encoder.layer.11.attention.output.LayerNorm.weight | [768] | 768 |
| code_encoder_q.encoder.layer.11.attention.output.LayerNorm.bias | [768] | 768 |
| code_encoder_q.encoder.layer.11.intermediate.dense.weight | [3072, 768] | 2359296 |
| code_encoder_q.encoder.layer.11.intermediate.dense.bias | [3072] | 3072 |
| code_encoder_q.encoder.layer.11.output.dense.weight | [768, 3072] | 2359296 |
| code_encoder_q.encoder.layer.11.output.dense.bias | [768] | 768 |
| code_encoder_q.encoder.layer.11.output.LayerNorm.weight | [768] | 768 |
| code_encoder_q.encoder.layer.11.output.LayerNorm.bias | [768] | 768 |
| code_encoder_q.pooler.dense.weight | [768, 768] | 589824 |
| code_encoder_q.pooler.dense.bias | [768] | 768 |
+-------------------------------------------------------------------+--------------+----------+
02/17/2024 15:12:09 - INFO - __main__ - Training/evaluation parameters Namespace(agg_way='cls_pooler', aug_type_way='random_replace_type', code_length=64, codebase_file='dataset/ruby/codebase.jsonl', config_name='microsoft/unixcoder-base', couninue_pre_train_data_files=['dataset/java/train.jsonl', 'dataset/javascript/train.jsonl', 'dataset/python/train.jsonl', 'dataset/php/train.jsonl', 'dataset/go/train.jsonl', 'dataset/ruby/train.jsonl'], data_aug_type='other', data_flow_length=0, debug=False, device=device(type='cuda'), do_avg=False, do_continue_pre_trained=False, do_eval=False, do_fine_tune=False, do_ineer_loss=False, do_multi_lang_continue_pre_train=True, do_single_lang_continue_pre_train=False, do_test=True, do_train=False, do_whitening=False, do_zero_short=False, epoch=50, eval_batch_size=64, eval_data_file='dataset/ruby/valid.jsonl', eval_frequency=100, fp16=False, gradient_accumulation_steps=1, hidden_size=768, lang='ruby', learning_rate=2e-05, loaded_codebert_model_filename=None, loaded_model_filename=None, local_rank=-1, logging_steps=50, max_codeblock_num=10, max_grad_norm=1.0, max_steps=100000, mlm_probability=0.1, mlp=False, moco_dim=768, moco_k=1024, moco_m=0.999, moco_t=0.07, moco_type='encoder_queue', model_name_or_path='microsoft/unixcoder-base', model_type='multi-loss-cocosoda', n_debug_samples=100, n_gpu=2, nl_length=64, num_train_epochs=10, num_warmup_steps=0, only_save_the_nl_code_vec=False, output_dir='./saved_models/cocosoda/', print_align_unif_loss=False, save_evaluation_reuslt=False, save_evaluation_reuslt_dir=None, save_steps=1000, seed=123456, test_data_file='dataset/ruby/test.jsonl', time_score=1, tokenizer_name='microsoft/unixcoder-base', train_batch_size=128, train_data_file='dataset/ruby/train.jsonl', use_best_mrr_model=False, weight_decay=0.01)
02/17/2024 15:15:03 - INFO - __main__ - *** Example ***
02/17/2024 15:15:03 - INFO - __main__ - idx: 0
02/17/2024 15:15:03 - INFO - __main__ - code_tokens: ['', '', '', '@', '_Override', '_public', '_Image', 'Source', '_apply', '_(', '_Image', 'Source', '_input', '_)', '_{', '_final', '_int', '_[', '_]', '_[', '_]', '_pixel', 'Matrix', '_=', '_new', '_int', '_[', '_3', '_]', '_[', '_3', '_]', '_;', '_int', '_w', '_=', '_input', '_.', '_getWidth', '_(', '_)', '_;', '_int', '_h', '_=', '_input', '_.', '_getHeight', '_(', '_)', '_;', '_int', '_[', '_]', '_[', '_]', '_output', '_=', '_new', '_int', '_[', '_h', '_]', '']
02/17/2024 15:15:03 - INFO - __main__ - code_ids: 0 6 2 150 19505 1240 6085 1768 5230 400 6085 1768 1586 743 399 1920 554 626 2406 626 2406 5578 3679 385 579 554 626 995 2406 626 995 2406 2476 554 477 385 1586 746 32671 400 743 2476 554 566 385 1586 746 32720 400 743 2476 554 626 2406 626 2406 1721 385 579 554 626 566 2406 2
02/17/2024 15:15:03 - INFO - __main__ - nl_tokens: ['', '', '', 'Expect', 's', '_a', '_height', '_mat', '_as', '_input', '']
02/17/2024 15:15:03 - INFO - __main__ - nl_ids: 0 6 2 7871 201 434 3082 5772 880 1586 2 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
02/17/2024 15:15:03 - INFO - __main__ - *** Example ***
02/17/2024 15:15:03 - INFO - __main__ - idx: 1
02/17/2024 15:15:03 - INFO - __main__ - code_tokens: ['', '', '', 'public', '_<', '_L', 'extends', 'Listener', '_>', '_void', '_pop', 'Event', '_(', '_Event', '_<', '_?', '_,', '_L', '_>', '_expected', '_)', '_{', '_synchronized', '_(', '_this', '_.', '_stack', '_)', '_{', '_final', '_Event', '_<', '_?', '_,', '_?', '_>', '_actual', '_=', '_this', '_.', '_stack', '_.', '_pop', '_(', '_)', '_;', '_if', '_(', '_actual', '_!=', '_expected', '_)', '_{', '_throw', '_new', '_IllegalStateException', '_(', '_String', '_.', '_format', '_(', '"', 'Un', '']
02/17/2024 15:15:03 - INFO - __main__ - code_ids: 0 6 2 653 517 747 13125 2486 711 723 5012 1089 400 3916 517 999 2019 747 711 2048 743 399 9401 400 547 746 3325 743 399 1920 3916 517 999 2019 999 711 3780 385 547 746 3325 746 5012 400 743 2476 462 400 3780 620 2048 743 399 1185 579 16219 400 1167 746 2021 400 120 965 2
02/17/2024 15:15:03 - INFO - __main__ - nl_tokens: ['', '', '', 'P', 'ops', '_the', '_top', '_event', '_off', '_the', '_current', '_event', '_stack', '_.', '_This', '_action', '_has', '_to', '_be', '_performed', '_immediately', '_after', '_the', '_event', '_has', '_been', '_dispatched', '_to', '_all', '_listeners', '_.', '']
02/17/2024 15:15:03 - INFO - __main__ - nl_ids: 0 6 2 166 2489 448 3194 1488 3413 448 1434 1488 3325 746 1600 2657 1559 508 661 13181 10086 2493 448 1488 1559 3022 43340 508 1345 11839 746 2 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
02/17/2024 15:15:03 - INFO - __main__ - *** Example ***
02/17/2024 15:15:03 - INFO - __main__ - idx: 2
02/17/2024 15:15:03 - INFO - __main__ - code_tokens: ['', '', '', 'protected', '_void', '_modify', '_(', '_Transaction', '_t', '_)', '_{', '_try', '_{', '_this', '_.', '_lock', '_.', '_write', 'Lock', '_(', '_)', '_.', '_lock', '_(', '_)', '_;', '_t', '_.', '_perform', '_(', '_)', '_;', '_}', '_finally', '_{', '_this', '_.', '_lock', '_.', '_write', 'Lock', '_(', '_)', '_.', '_unlock', '_(', '_)', '_;', '_}', '_}', '']
02/17/2024 15:15:03 - INFO - __main__ - code_ids: 0 6 2 1933 723 8660 400 13081 422 743 399 1568 399 547 746 3505 746 2250 2896 400 743 746 3505 400 743 2476 422 746 4729 400 743 2476 425 6110 399 547 746 3505 746 2250 2896 400 743 746 14552 400 743 2476 425 425 2 1 1 1 1 1 1 1 1 1 1 1 1 1
02/17/2024 15:15:03 - INFO - __main__ - nl_tokens: ['', '', '', 'Executes', '_the', '_given', '_transaction', '_within', '_the', '_con', 'text', 'of', '_a', '_write', '_lock', '_.', '']
02/17/2024 15:15:03 - INFO - __main__ - nl_ids: 0 6 2 40551 448 2076 4993 5289 448 549 625 757 434 2250 3505 746 2 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
02/17/2024 15:15:59 - INFO - __main__ - *** Example ***
02/17/2024 15:15:59 - INFO - __main__ - idx: 0
02/17/2024 15:15:59 - INFO - __main__ - code_tokens: ['', '', '', 'function', '_(', '_state', '_,', '_action', '_)', '_{', '_return', '__', '_.', '_defaults', '_(', '_{', '_isValid', 'ating', '_:', '_action', '_.', '_isValid', 'ating', '_,', '_last', 'Action', '_:', '_IS', '_', 'VALID', 'ATING', '_}', '_,', '_state', '_)', '_}', '']
02/17/2024 15:15:59 - INFO - __main__ - code_ids: 0 6 2 618 400 1404 2019 2657 743 399 483 623 746 7470 400 399 17002 2335 545 2657 746 17002 2335 2019 2023 1888 545 1947 181 7477 40173 425 2019 1404 743 425 2 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
02/17/2024 15:15:59 - INFO - __main__ - nl_tokens: ['', '', '', 'Update', '_is', '_validating', '_result', '']
02/17/2024 15:15:59 - INFO - __main__ - nl_ids: 0 6 2 2056 555 38924 1046 2 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
02/17/2024 15:15:59 - INFO - __main__ - *** Example ***
02/17/2024 15:15:59 - INFO - __main__ - idx: 1
02/17/2024 15:15:59 - INFO - __main__ - code_tokens: ['', '', '', 'function', '_add', 'Widget', 'For', 'Filter', '_(', '_view', '_,', '_filter', '_,', '_edit', 'Mode', 'Hint', '_)', '_{', '_var', '_grid', 'ster', '_=', '_view', '_.', '__', 'widgets', 'Grid', 'ster', '_;', '_var', '_row', '_=', '_filter', '_.', '_row', '_||', '_1', '_;', '_var', '_col', '_=', '_filter', '_.', '_col', '_||', '_1', '_;', '_var', '_size', 'X', '_=', '_filter', '_.', '_size', '_', 'x', '_||', '_3', '_;', '_var', '_size', 'Y', '_=', '']
02/17/2024 15:15:59 - INFO - __main__ - code_ids: 0 6 2 618 1103 3104 1459 2274 400 2859 2019 2866 2019 7277 1649 7641 743 399 660 6335 7400 385 2859 746 623 14718 3981 7400 2476 660 2562 385 2866 746 2562 853 524 2476 660 1253 385 2866 746 1253 853 524 2476 660 1014 174 385 2866 746 1014 181 206 853 995 2476 660 1014 175 385 2
02/17/2024 15:15:59 - INFO - __main__ - nl_tokens: ['', '', '', 'Add', '_a', '_widget', '_to', '_the', '_analyze', '_page', '_for', '_the', '_given', '_filter', '']
02/17/2024 15:15:59 - INFO - __main__ - nl_ids: 0 6 2 972 434 6949 508 448 25087 2303 563 448 2076 2866 2 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
02/17/2024 15:15:59 - INFO - __main__ - *** Example ***
02/17/2024 15:15:59 - INFO - __main__ - idx: 2
02/17/2024 15:15:59 - INFO - __main__ - code_tokens: ['', '', '', 'function', '_in', 'Range', '_(', '_value', '_,', '_min', '_,', '_max', '_)', '_{', '_const', '_int', '_=', '_parseInt', '_(', '_value', '_,', '_10', '_)', '_return', '_(', '_`', '_${', '_int', '_}', '_`', '_===', '_`', '_${', '_value', '_.', '_replace', '_(', '_/', '_^', '0', '_/', '_,', "_''", '_)', '_}', '_`', '_&&', '_int', '_>=', '_min', '_&&', '_int', '_<=', '_max', '_)', '_}', '']
02/17/2024 15:15:59 - INFO - __main__ - code_ids: 0 6 2 618 488 2228 400 767 2019 2069 2019 1621 743 399 925 554 385 9998 400 767 2019 1865 743 483 400 1222 5593 554 425 1222 1246 1222 5593 767 746 4126 400 1017 3855 134 1017 2019 3606 743 425 1222 698 554 1451 2069 698 554 1826 1621 743 425 2 1 1 1 1 1 1 1
02/17/2024 15:15:59 - INFO - __main__ - nl_tokens: ['', '', '', 'Determine', '_if', '_value', '_is', '_within', '_a', '_numeric', '_range', '']
02/17/2024 15:15:59 - INFO - __main__ - nl_ids: 0 6 2 17591 462 767 555 5289 434 10397 1780 2 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
02/17/2024 15:19:33 - INFO - __main__ - *** Example ***
02/17/2024 15:19:33 - INFO - __main__ - idx: 0
02/17/2024 15:19:33 - INFO - __main__ - code_tokens: ['', '', '', 'def', '_split', '_', 'phy', 'log', 'en', 'y', '_(', '_p', '_,', '_level', '_=', '"', 's', '"', ')', '_:', '_level', '_=', '_level', '_+', '"', '__', '"', 'result', '_=', '_p', '_.', '_split', '_(', '_level', '_)', '_return', '_result', '_[', '_0', '_]', '_+', '_level', '_+', '_result', '_[', '_1', '_]', '_.', '_split', '_(', '"', ';', '"', ')', '_[', '_0', '_]', '']
02/17/2024 15:19:33 - INFO - __main__ - code_ids: 0 6 2 729 5192 181 3258 896 386 207 400 428 2019 3144 385 120 201 120 127 545 3144 385 3144 513 120 876 120 1125 385 428 746 5192 400 3144 743 483 1046 626 461 2406 513 3144 513 1046 626 524 2406 746 5192 400 120 145 120 127 626 461 2406 2 1 1 1 1 1 1
02/17/2024 15:19:33 - INFO - __main__ - nl_tokens: ['', '', '', 'Return', '_either', '_the', '_full', '_or', '_truncated', '_version', '_of', '_a', '_Q', 'II', 'ME', '_-', '_formatted', '_taxonomy', 'string', '.', '']
02/17/2024 15:19:33 - INFO - __main__ - nl_ids: 0 6 2 1675 4759 448 3662 872 19307 2229 595 434 1152 4300 1098 581 10440 29021 571 132 2 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
02/17/2024 15:19:33 - INFO - __main__ - *** Example ***
02/17/2024 15:19:33 - INFO - __main__ - idx: 1
02/17/2024 15:19:33 - INFO - __main__ - code_tokens: ['', '', '', 'def', '_ensure', '_', 'dir', '_(', '_d', '_)', '_:', '_if', '_not', '_os', '_.', '_path', '_.', '_exists', '_(', '_d', '_)', '_:', '_try', '_:', '_os', '_.', '_m', 'akedirs', '_(', '_d', '_)', '_except', '_OSError', '_as', '_oe', '_:', '_#', '_should', '_not', '_happen', '_with', '_os', '.', 'makedirs', '_#', '_ENOENT', ':', '_No', '_such', '_file', '_or', '_directory', '_if', '_os', '_.', '_errno', '_==', '_errno', '_.', '_ENOENT', '_:', '_msg', '_=', '']
02/17/2024 15:19:33 - INFO - __main__ - code_ids: 0 6 2 729 6229 181 1282 400 480 743 545 462 800 2215 746 1391 746 4534 400 480 743 545 1568 545 2215 746 446 23328 400 480 743 3552 22934 880 44902 545 830 1570 800 7564 918 2215 132 24429 830 41059 144 4038 5632 1012 872 3456 462 2215 746 2341 550 2341 746 41059 545 2345 385 2
02/17/2024 15:19:33 - INFO - __main__ - nl_tokens: ['', '', '', 'Check', '_to', '_make', '_sure', '_the', '_supplied', '_directory', '_path', '_does', '_not', '_exist', '_if', '_so', '_create', '_it', '_.', '_The', '_method', '_catch', 'es', '_OSError', '_exceptions', '_and', '_returns', '_a', '_desc', 'riptive', '_message', '_instead', '_of', '_re', '_-', '_raising', '_the', '_error', '_.', '']
02/17/2024 15:19:33 - INFO - __main__ - nl_ids: 0 6 2 1749 508 2002 3984 448 8813 3456 1391 2129 800 3040 462 1769 1738 835 746 1044 1454 2092 482 22934 12300 706 2060 434 2162 44105 1841 4488 595 479 581 47183 448 843 746 2 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
02/17/2024 15:19:33 - INFO - __main__ - *** Example ***
02/17/2024 15:19:33 - INFO - __main__ - idx: 2
02/17/2024 15:19:33 - INFO - __main__ - code_tokens: ['', '', '', 'def', '_file', '_', 'handle', '_(', '_fn', 'h', '_,', '_mode', '_=', '"', 'r', 'U', '"', ')', '_:', '_handle', '_=', '_None', '_if', '_isinstance', '_(', '_fn', 'h', '_,', '_file', '_)', '_:', '_if', '_fn', 'h', '_.', '_closed', '_:', '_raise', '_ValueError', '_(', '"', 'Input', '_file', '_is', '_closed', '.', '"', ')', '_handle', '_=', '_fn', 'h', '_elif', '_isinstance', '_(', '_fn', 'h', '_,', '_str', '_)', '_:', '_handle', '_=', '']
02/17/2024 15:19:33 - INFO - __main__ - code_ids: 0 6 2 729 1012 181 2133 400 4065 190 2019 2119 385 120 200 171 120 127 545 2384 385 1938 462 5408 400 4065 190 2019 1012 743 545 462 4065 190 746 8264 545 3085 6052 400 120 1834 1012 555 8264 132 120 127 2384 385 4065 190 3625 5408 400 4065 190 2019 1113 743 545 2384 385 2
02/17/2024 15:19:33 - INFO - __main__ - nl_tokens: ['', '', '', 'Takes', '_either', '_a', '_file', '_path', '_or', '_an', '_open', '_file', '_handle', '_checks', '_validity', '_and', '_returns', '_an', '_open', '_file', '_handle', '_or', '_raises', '_an', '_appropriate', '_Exception', '_.', '']
02/17/2024 15:19:33 - INFO - __main__ - nl_ids: 0 6 2 27408 4759 434 1012 1391 872 817 2717 1012 2384 7825 25911 706 2060 817 2717 1012 2384 872 23154 817 7900 2654 746 2 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
02/17/2024 15:23:02 - INFO - __main__ - *** Example ***
02/17/2024 15:23:02 - INFO - __main__ - idx: 0
02/17/2024 15:23:02 - INFO - __main__ - code_tokens: ['', '', '', 'public', '_function', '_on', 'Channel', 'Pre', 'Delete', '_(', '_Resource', 'Controller', 'Event', '_$', '_event', '_)', '_:', '_void', '_{', '_$', '_channel', '_=', '_$', '_event', '_->', '_get', 'Subject', '_(', '_)', '_;', '_if', '_(', '_!', '_$', '_channel', '_instanceof', '_Channel', 'Interface', '_)', '_{', '_throw', '_new', '_Unexpected', 'TypeException', '_(', '_$', '_channel', '_,', '_Channel', 'Interface', '_::', 'class', ')', '_;', '_}', '_$', '_results', '_=', '_$', '_this', '_->', '_channel', 'Repository', '']
02/17/2024 15:23:02 - INFO - __main__ - code_ids: 0 6 2 653 603 854 3267 1782 2843 400 7606 3357 1089 440 1488 743 545 723 399 440 3225 385 440 1488 1703 744 7562 400 743 2476 462 400 552 440 3225 3052 11322 2285 743 399 1185 579 23297 48098 400 440 3225 2019 11322 2285 5431 1149 127 2476 425 440 3286 385 440 547 1703 3225 5674 2
02/17/2024 15:23:02 - INFO - __main__ - nl_tokens: ['', '', '', 'Prevent', '_channel', '_deletion', '_if', '_no', '_more', '_channels', '_enabled', '_.', '']
02/17/2024 15:23:02 - INFO - __main__ - nl_ids: 0 6 2 42669 3225 19744 462 1375 2726 8630 5334 746 2 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
02/17/2024 15:23:02 - INFO - __main__ - *** Example ***
02/17/2024 15:23:02 - INFO - __main__ - idx: 1
02/17/2024 15:23:02 - INFO - __main__ - code_tokens: ['', '', '', 'public', '_function', '_get', 'Tax', 'Total', '_(', '_)', '_:', '_int', '_{', '_$', '_tax', 'Total', '_=', '_0', '_;', '_foreach', '_(', '_$', '_this', '_->', '_get', 'Adjustments', '_(', '_Adjust', 'ment', 'Interface', '_::', '_T', 'AX', '_', 'ADJUST', 'MENT', '_)', '_as', '_$', '_tax', 'Adjustment', '_)', '_{', '_$', '_tax', 'Total', '_+=', '_$', '_tax', 'Adjustment', '_->', '_get', 'Amount', '_(', '_)', '_;', '_}', '_foreach', '_(', '_$', '_this', '_->', '_units', '']
02/17/2024 15:23:02 - INFO - __main__ - code_ids: 0 6 2 653 603 744 11266 4703 400 743 545 554 399 440 14990 4703 385 461 2476 2315 400 440 547 1703 744 39930 400 16203 564 2285 5431 515 3383 181 44094 4332 743 880 440 14990 21585 743 399 440 14990 4703 1054 440 14990 21585 1703 744 6933 400 743 2476 425 2315 400 440 547 1703 10931 2
02/17/2024 15:23:02 - INFO - __main__ - nl_tokens: ['', '', '', 'Returns', '_sum', '_of', '_ne', 'utral', '_and', '_non', '_ne', 'utral', '_tax', '_adjust', 'ments', '_on', '_order', '_item', '_and', '_total', '_tax', '_of', '_units', '_.', '']
02/17/2024 15:23:02 - INFO - __main__ - nl_ids: 0 6 2 2853 3863 595 1472 22943 706 2514 1472 22943 14990 7780 2067 854 2991 1573 706 3704 14990 595 10931 746 2 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
02/17/2024 15:23:02 - INFO - __main__ - *** Example ***
02/17/2024 15:23:02 - INFO - __main__ - idx: 2
02/17/2024 15:23:02 - INFO - __main__ - code_tokens: ['', '', '', 'private', '_function', '_is', 'Last', 'Enabled', 'Entity', '_(', '_$', '_result', '_,', '_$', '_entity', '_)', '_:', '_bool', '_{', '_return', '_!', '_$', '_result', '_||', '_0', '_===', '_count', '_(', '_$', '_result', '_)', '_||', '_(', '_1', '_===', '_count', '_(', '_$', '_result', '_)', '_&&', '_$', '_entity', '_===', '_(', '_$', '_result', '_instanceof', '_\\', '_Iterator', '_?', '_$', '_result', '_->', '_current', '_(', '_)', '_:', '_current', '_(', '_$', '_result', '_)', '']
02/17/2024 15:23:02 - INFO - __main__ - code_ids: 0 6 2 1335 603 555 2954 3060 2268 400 440 1046 2019 440 4498 743 545 1223 399 483 552 440 1046 853 461 1246 1752 400 440 1046 743 853 400 524 1246 1752 400 440 1046 743 698 440 4498 1246 400 440 1046 3052 1216 13119 999 440 1046 1703 1434 400 743 545 1434 400 440 1046 743 2
02/17/2024 15:23:02 - INFO - __main__ - nl_tokens: ['', '', '', 'If', '_no', '_entity', '_matched', '_the', '_query', '_criteria', '_or', '_a', '_single', '_entity', '_matched', '_which', '_is', '_the', '_same', '_as', '_the', '_entity', '_being', '_validated', '_the', '_entity', '_is', '_the', '_last', '_enabled', '_entity', '_available', '_.', '']
02/17/2024 15:23:02 - INFO - __main__ - nl_ids: 0 6 2 2815 1375 4498 5865 448 2616 14677 872 434 3501 4498 5865 1839 555 448 2641 880 448 4498 4251 20709 448 4498 555 448 2023 5334 4498 3777 746 2 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
02/17/2024 15:25:32 - INFO - __main__ - *** Example ***
02/17/2024 15:25:32 - INFO - __main__ - idx: 0
02/17/2024 15:25:32 - INFO - __main__ - code_tokens: ['', '', '', 'func', '_getAll', 'Dep', 'Types', '_(', '_)', '_[', '_]', 'string', '{', '_dep', 'Types', '_:=', '_make', '_(', '_[', '_]', 'string', ',', '_0', '_,', '_len', '_(', '_cmds', '_)', '_)', '_Ċ', '_for', '_dep', 'Type', '_:=', '_range', '_cmds', '_{', '_dep', 'Types', '_=', '_append', '_(', '_dep', 'Types', '_,', '_dep', 'Type', '_)', '_Ċ', '_}', '_Ċ', '_sort', '_.', '_Strings', '_(', '_dep', 'Types', '_)', '_Ċ', '_return', '_dep', 'Types', '_Ċ', '']
02/17/2024 15:25:32 - INFO - __main__ - code_ids: 0 6 2 763 21556 15010 2531 400 743 626 2406 571 209 13994 2531 716 2002 400 626 2406 571 130 461 2019 1015 400 22803 743 743 1022 563 13994 641 716 1780 22803 399 13994 2531 385 2746 400 13994 2531 2019 13994 641 743 1022 425 1022 4821 746 23012 400 13994 2531 743 1022 483 13994 2531 1022 2
02/17/2024 15:25:32 - INFO - __main__ - nl_tokens: ['', '', '', 'getAll', 'Dep', 'Types', '_returns', '_a', '_sorted', '_list', '_of', 'name', 's', '_of', '_all', '_dep', '_type', '_commands', '_.', '']
02/17/2024 15:25:32 - INFO - __main__ - nl_ids: 0 6 2 12199 15010 2531 2060 434 6977 1182 595 616 201 595 1345 13994 889 7997 746 2 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
02/17/2024 15:25:32 - INFO - __main__ - *** Example ***
02/17/2024 15:25:32 - INFO - __main__ - idx: 1
02/17/2024 15:25:32 - INFO - __main__ - code_tokens: ['', '', '', 'func', '_get', 'Io', 'Progress', 'Reader', '_(', '_label', 'string', ',', '_res', '_*', '_http', '_.', '_Response', '_)', '_io', '_.', '_Reader', '_{', '_prefix', '_:=', '"', '"', '+', '_label', '_Ċ', '_fmt', 'Bytes', 'Size', '_:=', '_18', '_Ċ', '_bar', 'Size', '_:=', '_int', '64', '_(', '_80', '_-', '_len', '_(', '_prefix', '_)', '_-', '_fmt', 'Bytes', 'Size', '_)', '_Ċ', '_bar', '_:=', '_i', 'opro', 'gress', '_.', '_Draw', 'Text', 'Format', 'Bar', '']
02/17/2024 15:25:32 - INFO - __main__ - code_ids: 0 6 2 763 744 8499 4909 2692 400 2649 571 130 705 426 2014 746 6397 743 3095 746 15471 399 3603 716 120 120 129 2649 1022 2771 2240 939 716 7837 1022 5252 939 716 554 848 400 8967 581 1015 400 3603 743 581 2771 2240 939 743 1022 5252 716 548 31375 2639 746 8548 1072 1660 3238 2
02/17/2024 15:25:32 - INFO - __main__ - nl_tokens: ['', '', '', 'get', 'Io', 'Progress', 'Reader', '_returns', '_a', '_reader', '_that', '_wraps', '_the', '_HTTP', '_response', '_body', '_so', '_it', '_prints', '_a', '_pretty', '_progress', '_bar', '_when', '_reading', '_data', '_from', '_it', '_.', '']
02/17/2024 15:25:32 - INFO - __main__ - nl_ids: 0 6 2 459 8499 4909 2692 2060 434 4636 922 28232 448 4383 1925 3444 1769 835 22199 434 15344 6687 5252 1672 8267 869 1029 835 746 2 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
02/17/2024 15:25:32 - INFO - __main__ - *** Example ***
02/17/2024 15:25:32 - INFO - __main__ - idx: 2
02/17/2024 15:25:32 - INFO - __main__ - code_tokens: ['', '', '', 'func', '_(', '_f', '_*', '_remove', 'OnClose', '_)', '_Close', '_(', '_)', '_error', '_{', '_if', '_f', '_==', '_nil', '_||', '_f', '_.', '_File', '_==', '_nil', '_{', '_return', '_nil', '_Ċ', '_}', 'name', ':', '=', '_f', '_.', '_File', '_.', '_Name', '_(', '_)', '_Ċ', '_if', '_err', '_:=', '_f', '_.', '_File', '_.', '_Close', '_(', '_)', '_;', '_err', '_!=', '_nil', '_{', '_return', '_err', '_Ċ', '_}', '_Ċ', '_if', '_err', '']
02/17/2024 15:25:32 - INFO - __main__ - code_ids: 0 6 2 763 400 412 426 3033 45359 743 5832 400 743 843 399 462 412 550 845 853 412 746 2536 550 845 399 483 845 1022 425 616 144 147 412 746 2536 746 3725 400 743 1022 462 573 716 412 746 2536 746 5832 400 743 2476 573 620 845 399 483 573 1022 425 1022 462 573 2
02/17/2024 15:25:32 - INFO - __main__ - nl_tokens: ['', '', '', 'Close', '_closes', '_the', '_file', '_and', '_then', '_removes', '_it', '_from', '_disk', '_.', '_No', '_error', '_is', '_returned', '_if', '_the', '_file', '_did', '_not', '_exist', '_at', '_the', '_point', '_of', '_removal', '_.', '']
02/17/2024 15:25:32 - INFO - __main__ - nl_ids: 0 6 2 3108 19735 448 1012 706 2270 15719 835 1029 8236 746 4038 843 555 2862 462 448 1012 6088 800 3040 1035 448 1704 595 23066 746 2 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
02/17/2024 15:25:48 - INFO - __main__ - *** Example ***
02/17/2024 15:25:48 - INFO - __main__ - idx: 0
02/17/2024 15:25:48 - INFO - __main__ - code_tokens: ['', '', '', 'def', '_render', '_', 'body', '_(', '_con', 'text', ',', '_options', '_)', '_if', '_options', '_.', '_key', '?', '_(', '_:', 'partial', '_)', '_[', '_render', '_', 'partial', '_(', '_con', 'text', ',', '_options', '_)', '_]', '_else', '_Streaming', 'Template', 'Renderer', '_.', '_new', '_(', '_@', 'lookup', '_', 'con', 'text', ')', '_.', '_render', '_(', '_con', 'text', ',', '_options', '_)', '_end', '_end', '']
02/17/2024 15:25:48 - INFO - __main__ - code_ids: 0 6 2 729 4342 181 1995 400 549 625 130 1466 743 462 1466 746 1129 149 400 545 7609 743 626 4342 181 7609 400 549 625 130 1466 743 2406 669 47128 3057 6412 746 579 400 890 4961 181 525 625 127 746 4342 400 549 625 130 1466 743 1013 1013 2 1 1 1 1 1 1 1
02/17/2024 15:25:48 - INFO - __main__ - nl_tokens: ['', '', '', 'Render', '_but', '_returns', '_a', '_valid', '_R', 'ack', '_body', '_.', '_If', '_fib', 'ers', '_are', '_defined', '_we', '_return', '_a', '_streaming', '_body', '_that', '_renders', '_the', '_template', '_piece', '_by', '_piece', '_.', '']
02/17/2024 15:25:48 - INFO - __main__ - nl_ids: 0 6 2 3726 2107 2060 434 1976 821 598 3444 746 1359 24766 560 1147 3474 937 483 434 22676 3444 922 40840 448 3636 18781 1243 18781 746 2 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
02/17/2024 15:25:48 - INFO - __main__ - *** Example ***
02/17/2024 15:25:48 - INFO - __main__ - idx: 1
02/17/2024 15:25:48 - INFO - __main__ - code_tokens: ['', '', '', 'def', '_attribute', '_', 'missing', '_(', '_match', '_,', '_*', '_args', '_,', '_&', '_block', '_)', '___', 'send', '__', '_(', '_match', '_.', '_target', '_,', '_match', '_.', '_attr', '_', 'name', ',', '_args', '_,', '_block', '_)', '_end', '']
02/17/2024 15:25:48 - INFO - __main__ - code_ids: 0 6 2 729 2416 181 8487 400 1655 2019 426 1822 2019 519 1818 743 1267 2414 876 400 1655 746 1744 2019 1655 746 3526 181 616 130 1822 2019 1818 743 1013 2 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
02/17/2024 15:25:48 - INFO - __main__ - nl_tokens: ['', '', '', '+', '_attribute', '_', 'missing', '_+', '_is', '_like', '_+', '_method', '_', 'missing', '_+', '_but', '_for', '_attributes', '_.', '_When', '_+', '_method', '_', 'missing', '_+', '_is', '_called', '_we', '_check', '_to', '_see', '_if', '_there', '_is', '_a', '_matching', '_attribute', '_method', '_.', '_If', '_so', '_we', '_tell', '_+', '_attribute', '_', 'missing', '_+', '_to', '_dispatch', '_the', '_attribute', '_.', '_This', '_method', '_can', '_be', '_overloaded', '_to', '_customize', '_the', '_behavior', '_.', '']
02/17/2024 15:25:48 - INFO - __main__ - nl_ids: 0 6 2 129 2416 181 8487 513 555 4401 513 1454 181 8487 513 2107 563 4402 746 5919 513 1454 181 8487 513 555 2953 937 1382 508 3986 462 2550 555 434 6506 2416 1454 746 1359 1769 937 11931 513 2416 181 8487 513 508 9363 448 2416 746 1600 1454 1347 661 45869 508 36145 448 9050 746 2
02/17/2024 15:25:48 - INFO - __main__ - *** Example ***
02/17/2024 15:25:48 - INFO - __main__ - idx: 2
02/17/2024 15:25:48 - INFO - __main__ - code_tokens: ['', '', '', 'def', '_matched', '_', 'attribute', '_', 'method', '_(', '_method', '_', 'name', ')', '_matches', '_=', '_self', '_.', 'class', '.', '_send', '_(', '_:', 'attribute', '_', 'method', '_', 'matchers', '_', 'matching', '_,', '_method', '_', 'name', ')', '_matches', '_.', '_detect', '_{', '_|', '_match', '_|', '_attribute', '_', 'method', '?', '_(', '_match', '_.', '_attr', '_', 'name', ')', '_}', '_end', '']
02/17/2024 15:25:48 - INFO - __main__ - code_ids: 0 6 2 729 5865 181 2163 181 1521 400 1454 181 616 127 5288 385 1358 746 1149 132 2904 400 545 2163 181 1521 181 38734 181 13575 2019 1454 181 616 127 5288 746 10241 399 649 1655 649 2416 181 1521 149 400 1655 746 3526 181 616 127 425 1013 2 1 1 1 1 1 1 1 1
02/17/2024 15:25:48 - INFO - __main__ - nl_tokens: ['', '', '', 'Returns', '_a', '_struct', '_representing', '_the', '_matching', '_attribute', '_method', '_.', '_The', '_struct', '_s', '_attributes', '_are', '_prefix', '_base', '_and', '_suffix', '_.', '']
02/17/2024 15:25:48 - INFO - __main__ - nl_ids: 0 6 2 2853 434 1277 8466 448 6506 2416 1454 746 1044 1277 431 4402 1147 3603 1712 706 8436 746 2 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
02/17/2024 15:25:48 - INFO - __main__ - ***** Running training *****
02/17/2024 15:25:48 - INFO - __main__ - Num examples = 908224
02/17/2024 15:25:48 - INFO - __main__ - Num Epochs = 10
02/17/2024 15:25:48 - INFO - __main__ - Num quene = 1024
02/17/2024 15:25:48 - INFO - __main__ - Instantaneous batch size per GPU = 64
02/17/2024 15:25:48 - INFO - __main__ - Total train batch size = 128
Traceback (most recent call last):
File "run.py", line 1200, in
main()
File "run.py", line 1160, in main
multi_lang_continue_pre_train(args, model, tokenizer, pool)
File "run.py", line 756, in multi_lang_continue_pre_train
nl_q=nl_inputs , nl_k=nl_transformations_ids )
File "/home/yiming/anaconda3/envs/CoCoSoDa/lib/python3.6/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
return forward_call(*input, **kwargs)
File "/home/yiming/anaconda3/envs/CoCoSoDa/lib/python3.6/site-packages/torch/nn/parallel/data_parallel.py", line 168, in forward
outputs = self.parallel_apply(replicas, inputs, kwargs)
File "/home/yiming/anaconda3/envs/CoCoSoDa/lib/python3.6/site-packages/torch/nn/parallel/data_parallel.py", line 178, in parallel_apply
return parallel_apply(replicas, inputs, kwargs, self.device_ids[:len(replicas)])
File "/home/yiming/anaconda3/envs/CoCoSoDa/lib/python3.6/site-packages/torch/nn/parallel/parallel_apply.py", line 86, in parallel_apply
output.reraise()
File "/home/yiming/anaconda3/envs/CoCoSoDa/lib/python3.6/site-packages/torch/_utils.py", line 434, in reraise
raise exception
UnboundLocalError: Caught UnboundLocalError in replica 0 on device 0.
Original Traceback (most recent call last):
File "/home/yiming/anaconda3/envs/CoCoSoDa/lib/python3.6/site-packages/torch/nn/parallel/parallel_apply.py", line 61, in _worker
output = module(*input, **kwargs)
File "/home/yiming/anaconda3/envs/CoCoSoDa/lib/python3.6/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
return forward_call(*input, **kwargs)
File "/home/yiming/cocosoda/CoCoSoDa/model.py", line 235, in forward
code_q = torch.nn.functional.normalize(code_q, p=2, dim=1)
UnboundLocalError: local variable 'code_q' referenced before assignment