+ echo Logging output to /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743//log_node1.txt Logging output to /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743//log_node1.txt + export ASCEND_PROCESS_LOG_PATH=/data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743//ascend/1 + ASCEND_PROCESS_LOG_PATH=/data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743//ascend/1 + mkdir -p /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743//ascend/1 + DATA_PATH=/local_disk/cognitron_vl//configs/lcvlm_finetune_stage4.yaml + TOKENIZER_PATH=/data_4/models/Qwen/Qwen2.5-14B-Instruct/ + CKPT_LOAD_DIR=/data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/ + VIT_CKPT_LOAD_DIR=/ + CKPT_SAVE_DIR=/data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743// + rsync -avh /local_disk/cognitron_vl//configs/lcvlm_finetune_stage4.yaml /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743/ sending incremental file list sent 71 bytes received 12 bytes 166.00 bytes/sec total size is 23.84K speedup is 287.17 + cd /local_disk/cognitron_vl/ + rm -fr datasets + mkdir -p datasets + ln -s /data/data/ datasets/CV + ln -s /data/data/LLM datasets/LLM + ln -s /data/data/LMM datasets/LMM + source /local_disk/cognitron_vl//scripts/set_env_mg_npu.sh ++ source /usr/local/Ascend/driver/bin/setenv.bash +++ DEP_INFO_FILE=/etc/ascend_install.info +++ [[ -f /etc/ascend_install.info ]] +++ . /etc/ascend_install.info +++ DRV_LIB64_COMMON_LDPATH=/driver/lib64/common +++ DRV_LIB64_DRV_LDPATH=/driver/lib64/driver +++ DRV_LIB64_LDPATH=/driver/lib64 +++ export LD_LIBRARY_PATH=/driver/lib64/common:/driver/lib64/driver:/driver/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/lib/x86_64-linux-gnu/hdf5/serial: +++ LD_LIBRARY_PATH=/driver/lib64/common:/driver/lib64/driver:/driver/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/lib/x86_64-linux-gnu/hdf5/serial: +++ export PATH=/usr/local/Ascend/ascend-toolkit/latest/bin:/usr/local/Ascend/ascend-toolkit/latest/compiler/ccec_compiler/bin:/usr/local/Ascend/ascend-toolkit/latest/tools/ccec_compiler/bin:/root/miniconda3/envs/py38/bin:/root/miniconda3/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/local/sbin:/usr/local/bin +++ PATH=/usr/local/Ascend/ascend-toolkit/latest/bin:/usr/local/Ascend/ascend-toolkit/latest/compiler/ccec_compiler/bin:/usr/local/Ascend/ascend-toolkit/latest/tools/ccec_compiler/bin:/root/miniconda3/envs/py38/bin:/root/miniconda3/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/local/sbin:/usr/local/bin ++ source /usr/local/Ascend/ascend-toolkit/set_env.sh +++ export LD_LIBRARY_PATH=/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/driver/lib64/common:/driver/lib64/driver:/driver/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/lib/x86_64-linux-gnu/hdf5/serial: +++ LD_LIBRARY_PATH=/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/driver/lib64/common:/driver/lib64/driver:/driver/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/lib/x86_64-linux-gnu/hdf5/serial: +++ export ASCEND_TOOLKIT_HOME=/usr/local/Ascend/ascend-toolkit/latest +++ ASCEND_TOOLKIT_HOME=/usr/local/Ascend/ascend-toolkit/latest ++++ arch +++ export LD_LIBRARY_PATH=/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/driver/lib64/common:/driver/lib64/driver:/driver/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/lib/x86_64-linux-gnu/hdf5/serial: +++ LD_LIBRARY_PATH=/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/driver/lib64/common:/driver/lib64/driver:/driver/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/lib/x86_64-linux-gnu/hdf5/serial: +++ export LD_LIBRARY_PATH=/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/driver/lib64/common:/driver/lib64/driver:/driver/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/lib/x86_64-linux-gnu/hdf5/serial: +++ LD_LIBRARY_PATH=/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/driver/lib64/common:/driver/lib64/driver:/driver/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/lib/x86_64-linux-gnu/hdf5/serial: +++ export PYTHONPATH=/usr/local/Ascend/ascend-toolkit/latest/python/site-packages:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe:/usr/local/Ascend/ascend-toolkit/latest/python/site-packages:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe: +++ PYTHONPATH=/usr/local/Ascend/ascend-toolkit/latest/python/site-packages:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe:/usr/local/Ascend/ascend-toolkit/latest/python/site-packages:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe: +++ export PATH=/usr/local/Ascend/ascend-toolkit/latest/bin:/usr/local/Ascend/ascend-toolkit/latest/compiler/ccec_compiler/bin:/usr/local/Ascend/ascend-toolkit/latest/tools/ccec_compiler/bin:/usr/local/Ascend/ascend-toolkit/latest/bin:/usr/local/Ascend/ascend-toolkit/latest/compiler/ccec_compiler/bin:/usr/local/Ascend/ascend-toolkit/latest/tools/ccec_compiler/bin:/root/miniconda3/envs/py38/bin:/root/miniconda3/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/local/sbin:/usr/local/bin +++ PATH=/usr/local/Ascend/ascend-toolkit/latest/bin:/usr/local/Ascend/ascend-toolkit/latest/compiler/ccec_compiler/bin:/usr/local/Ascend/ascend-toolkit/latest/tools/ccec_compiler/bin:/usr/local/Ascend/ascend-toolkit/latest/bin:/usr/local/Ascend/ascend-toolkit/latest/compiler/ccec_compiler/bin:/usr/local/Ascend/ascend-toolkit/latest/tools/ccec_compiler/bin:/root/miniconda3/envs/py38/bin:/root/miniconda3/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/local/sbin:/usr/local/bin +++ export ASCEND_AICPU_PATH=/usr/local/Ascend/ascend-toolkit/latest +++ ASCEND_AICPU_PATH=/usr/local/Ascend/ascend-toolkit/latest +++ export ASCEND_OPP_PATH=/usr/local/Ascend/ascend-toolkit/latest/opp +++ ASCEND_OPP_PATH=/usr/local/Ascend/ascend-toolkit/latest/opp +++ export TOOLCHAIN_HOME=/usr/local/Ascend/ascend-toolkit/latest/toolkit +++ TOOLCHAIN_HOME=/usr/local/Ascend/ascend-toolkit/latest/toolkit +++ export ASCEND_HOME_PATH=/usr/local/Ascend/ascend-toolkit/latest +++ ASCEND_HOME_PATH=/usr/local/Ascend/ascend-toolkit/latest ++ export HCCL_CONNECT_TIMEOUT=7200 ++ HCCL_CONNECT_TIMEOUT=7200 ++ export HCCL_EXEC_TIMEOUT=7200 ++ HCCL_EXEC_TIMEOUT=7200 ++ export COMBINED_ENABLE=1 ++ COMBINED_ENABLE=1 ++ export MULTI_STREAM_MEMORY_REUSE=1 ++ MULTI_STREAM_MEMORY_REUSE=1 ++ export HCCL_RDMA_TC=160 ++ HCCL_RDMA_TC=160 ++ export HCCL_RDMA_SL=5 ++ HCCL_RDMA_SL=5 ++ export HCCL_INTRA_PCIE_ENABLE=0 ++ HCCL_INTRA_PCIE_ENABLE=0 ++ export HCCL_INTRA_ROCE_ENABLE=1 ++ HCCL_INTRA_ROCE_ENABLE=1 ++ export HCCL_RDMA_TIMEOUT=20 ++ HCCL_RDMA_TIMEOUT=20 ++ export INF_NAN_MODE_ENABLE=1 ++ INF_NAN_MODE_ENABLE=1 ++ export DISTRIBUTED_BACKEND=hccl ++ DISTRIBUTED_BACKEND=hccl ++ export ASCEND_LAUNCH_BLOCKING=0 ++ ASCEND_LAUNCH_BLOCKING=0 ++ export ASCEND_SLOG_PRINT_TO_STDOUT=0 ++ ASCEND_SLOG_PRINT_TO_STDOUT=0 ++ export ASCEND_GLOBAL_LOG_LEVEL=3 ++ ASCEND_GLOBAL_LOG_LEVEL=3 ++ export ASCEND_GLOBAL_EVENT_ENABLE=0 ++ ASCEND_GLOBAL_EVENT_ENABLE=0 ++ export TASK_QUEUE_ENABLE=1 ++ TASK_QUEUE_ENABLE=1 ++ export PTCOPY_ENABLE=1 ++ PTCOPY_ENABLE=1 ++ export COMBINED_ENABLE=1 ++ COMBINED_ENABLE=1 ++ export DYNAMIC_OP=ADD#MUL ++ DYNAMIC_OP=ADD#MUL ++ export HCCL_WHITELIST_DISABLE=1 ++ HCCL_WHITELIST_DISABLE=1 ++ export HCCL_CONNECT_TIMEOUT=7200 ++ HCCL_CONNECT_TIMEOUT=7200 ++ export HCCL_WHITELIST_DISABLE=1 ++ HCCL_WHITELIST_DISABLE=1 ++ export CUDA_DEVICE_MAX_CONNECTIONS=1 ++ CUDA_DEVICE_MAX_CONNECTIONS=1 ++ pip3 install --no-index --find-links=/data/software/ -r requirements_npu.txt Looking in links: /data/software/ Processing data/software/expecttest-0.2.1-py3-none-any.whl (from -r requirements_npu.txt (line 1)) Requirement already satisfied: peft in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 2)) (0.7.0) Processing data/software/XlsxWriter-3.2.0-py3-none-any.whl (from -r requirements_npu.txt (line 3)) Requirement already satisfied: termcolor in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 4)) (2.4.0) Requirement already satisfied: tabulate in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 5)) (0.9.0) Processing data/software/tiktoken-0.7.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (from -r requirements_npu.txt (line 6)) Requirement already satisfied: matplotlib in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 7)) (3.7.5) Processing data/software/datasets-3.0.0-py3-none-any.whl (from -r requirements_npu.txt (line 8)) Requirement already satisfied: einops in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 9)) (0.7.0) Processing data/software/pybind11-2.13.6-py3-none-any.whl (from -r requirements_npu.txt (line 10)) Requirement already satisfied: tensorboardX in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 11)) (2.6.2.2) Processing data/software/pyarrow-17.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (from -r requirements_npu.txt (line 12)) Requirement already satisfied: transformers>=4.40.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 13)) (4.40.1) Requirement already satisfied: deepspeed>=0.14.2 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 14)) (0.14.5) Processing data/software/accelerate-0.34.2-py3-none-any.whl (from -r requirements_npu.txt (line 15)) Requirement already satisfied: timm in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 16)) (0.9.16) Processing data/software/flask-3.0.3-py3-none-any.whl (from -r requirements_npu.txt (line 17)) Processing data/software/Flask_RESTful-0.3.10-py2.py3-none-any.whl (from -r requirements_npu.txt (line 18)) Processing data/software/decord-0.6.0-py3-none-manylinux2010_x86_64.whl (from -r requirements_npu.txt (line 19)) Processing data/software/natsort-8.4.0-py3-none-any.whl (from -r requirements_npu.txt (line 20)) Requirement already satisfied: numpy>=1.17 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft->-r requirements_npu.txt (line 2)) (1.24.4) Requirement already satisfied: packaging>=20.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft->-r requirements_npu.txt (line 2)) (23.2) Requirement already satisfied: psutil in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft->-r requirements_npu.txt (line 2)) (5.9.8) Requirement already satisfied: pyyaml in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft->-r requirements_npu.txt (line 2)) (5.4.1) Requirement already satisfied: torch>=1.13.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft->-r requirements_npu.txt (line 2)) (2.1.0+cpu) Requirement already satisfied: tqdm in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft->-r requirements_npu.txt (line 2)) (4.66.2) Requirement already satisfied: safetensors in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft->-r requirements_npu.txt (line 2)) (0.4.2) Requirement already satisfied: huggingface-hub>=0.17.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft->-r requirements_npu.txt (line 2)) (0.20.3) Requirement already satisfied: regex>=2022.1.18 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from tiktoken->-r requirements_npu.txt (line 6)) (2023.12.25) Requirement already satisfied: requests>=2.26.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from tiktoken->-r requirements_npu.txt (line 6)) (2.31.0) Requirement already satisfied: contourpy>=1.0.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from matplotlib->-r requirements_npu.txt (line 7)) (1.1.1) Requirement already satisfied: cycler>=0.10 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from matplotlib->-r requirements_npu.txt (line 7)) (0.12.1) Requirement already satisfied: fonttools>=4.22.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from matplotlib->-r requirements_npu.txt (line 7)) (4.49.0) Requirement already satisfied: kiwisolver>=1.0.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from matplotlib->-r requirements_npu.txt (line 7)) (1.4.5) Requirement already satisfied: pillow>=6.2.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from matplotlib->-r requirements_npu.txt (line 7)) (10.2.0) Requirement already satisfied: pyparsing>=2.3.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from matplotlib->-r requirements_npu.txt (line 7)) (3.1.1) Requirement already satisfied: python-dateutil>=2.7 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from matplotlib->-r requirements_npu.txt (line 7)) (2.8.2) Requirement already satisfied: importlib-resources>=3.2.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from matplotlib->-r requirements_npu.txt (line 7)) (6.1.2) Requirement already satisfied: filelock in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets>=2.21.0->-r requirements_npu.txt (line 8)) (3.13.1) Requirement already satisfied: dill<0.3.9,>=0.3.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets>=2.21.0->-r requirements_npu.txt (line 8)) (0.3.7) Requirement already satisfied: pandas in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets>=2.21.0->-r requirements_npu.txt (line 8)) (2.0.3) Processing data/software/requests-2.32.3-py3-none-any.whl (from tiktoken->-r requirements_npu.txt (line 6)) Processing data/software/tqdm-4.67.1-py3-none-any.whl (from peft->-r requirements_npu.txt (line 2)) Requirement already satisfied: xxhash in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets>=2.21.0->-r requirements_npu.txt (line 8)) (3.4.1) Requirement already satisfied: multiprocess in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets>=2.21.0->-r requirements_npu.txt (line 8)) (0.70.15) Requirement already satisfied: fsspec<=2024.6.1,>=2023.1.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from fsspec[http]<=2024.6.1,>=2023.1.0->datasets>=2.21.0->-r requirements_npu.txt (line 8)) (2023.10.0) Requirement already satisfied: aiohttp in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets>=2.21.0->-r requirements_npu.txt (line 8)) (3.9.3) Processing data/software/huggingface_hub-0.26.2-py3-none-any.whl (from peft->-r requirements_npu.txt (line 2)) Requirement already satisfied: protobuf>=3.20 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from tensorboardX->-r requirements_npu.txt (line 11)) (4.25.3) Requirement already satisfied: tokenizers<0.20,>=0.19 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from transformers>=4.40.1->-r requirements_npu.txt (line 13)) (0.19.1) Requirement already satisfied: hjson in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from deepspeed>=0.14.2->-r requirements_npu.txt (line 14)) (3.1.0) Requirement already satisfied: ninja in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from deepspeed>=0.14.2->-r requirements_npu.txt (line 14)) (1.11.1.1) Requirement already satisfied: nvidia-ml-py in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from deepspeed>=0.14.2->-r requirements_npu.txt (line 14)) (12.560.30) Requirement already satisfied: py-cpuinfo in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from deepspeed>=0.14.2->-r requirements_npu.txt (line 14)) (9.0.0) Requirement already satisfied: pydantic in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from deepspeed>=0.14.2->-r requirements_npu.txt (line 14)) (1.10.15) Processing data/software/safetensors-0.4.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (from peft->-r requirements_npu.txt (line 2)) Requirement already satisfied: torchvision in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from timm->-r requirements_npu.txt (line 16)) (0.16.0) Requirement already satisfied: Werkzeug>=3.0.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from flask->-r requirements_npu.txt (line 17)) (3.0.1) Requirement already satisfied: Jinja2>=3.1.2 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from flask->-r requirements_npu.txt (line 17)) (3.1.3) Processing data/software/itsdangerous-2.2.0-py3-none-any.whl (from flask->-r requirements_npu.txt (line 17)) Requirement already satisfied: click>=8.1.3 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from flask->-r requirements_npu.txt (line 17)) (8.1.7) Processing data/software/blinker-1.8.2-py3-none-any.whl (from flask->-r requirements_npu.txt (line 17)) Requirement already satisfied: importlib-metadata>=3.6.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from flask->-r requirements_npu.txt (line 17)) (7.0.1) Processing data/software/aniso8601-9.0.1-py2.py3-none-any.whl (from flask_restful->-r requirements_npu.txt (line 18)) Requirement already satisfied: six>=1.3.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from flask_restful->-r requirements_npu.txt (line 18)) (1.16.0) Requirement already satisfied: pytz in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from flask_restful->-r requirements_npu.txt (line 18)) (2024.1) Requirement already satisfied: aiosignal>=1.1.2 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets>=2.21.0->-r requirements_npu.txt (line 8)) (1.3.1) Requirement already satisfied: attrs>=17.3.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets>=2.21.0->-r requirements_npu.txt (line 8)) (23.2.0) Requirement already satisfied: frozenlist>=1.1.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets>=2.21.0->-r requirements_npu.txt (line 8)) (1.4.1) Requirement already satisfied: multidict<7.0,>=4.5 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets>=2.21.0->-r requirements_npu.txt (line 8)) (6.0.5) Requirement already satisfied: yarl<2.0,>=1.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets>=2.21.0->-r requirements_npu.txt (line 8)) (1.9.4) Requirement already satisfied: async-timeout<5.0,>=4.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets>=2.21.0->-r requirements_npu.txt (line 8)) (4.0.3) Requirement already satisfied: typing-extensions>=3.7.4.3 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from huggingface-hub>=0.17.0->peft->-r requirements_npu.txt (line 2)) (4.10.0) Requirement already satisfied: zipp>=0.5 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from importlib-metadata>=3.6.0->flask->-r requirements_npu.txt (line 17)) (3.17.0) Requirement already satisfied: MarkupSafe>=2.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from Jinja2>=3.1.2->flask->-r requirements_npu.txt (line 17)) (2.1.5) Requirement already satisfied: charset-normalizer<4,>=2 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from requests>=2.26.0->tiktoken->-r requirements_npu.txt (line 6)) (3.3.2) Requirement already satisfied: idna<4,>=2.5 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from requests>=2.26.0->tiktoken->-r requirements_npu.txt (line 6)) (3.6) Requirement already satisfied: urllib3<3,>=1.21.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from requests>=2.26.0->tiktoken->-r requirements_npu.txt (line 6)) (1.26.18) Requirement already satisfied: certifi>=2017.4.17 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from requests>=2.26.0->tiktoken->-r requirements_npu.txt (line 6)) (2024.2.2) Requirement already satisfied: sympy in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from torch>=1.13.0->peft->-r requirements_npu.txt (line 2)) (1.4) Requirement already satisfied: networkx in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from torch>=1.13.0->peft->-r requirements_npu.txt (line 2)) (3.1) Requirement already satisfied: tzdata>=2022.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from pandas->datasets>=2.21.0->-r requirements_npu.txt (line 8)) (2024.1) Requirement already satisfied: mpmath>=0.19 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from sympy->torch>=1.13.0->peft->-r requirements_npu.txt (line 2)) (1.3.0) DEPRECATION: apex 0.1-ascend-20240523 has a non-standard version number. pip 24.1 will enforce this behaviour change. A possible replacement is to upgrade to a newer version of apex or contact the author to suggest that they release a version with a conforming version number. Discussion can be found at https://github.com/pypa/pip/issues/12063 Installing collected packages: aniso8601, xlsxwriter, tqdm, safetensors, requests, pybind11, pyarrow, natsort, itsdangerous, expecttest, decord, blinker, tiktoken, huggingface-hub, flask, flask_restful, accelerate, datasets Attempting uninstall: tqdm Found existing installation: tqdm 4.66.2 Uninstalling tqdm-4.66.2: Successfully uninstalled tqdm-4.66.2 Attempting uninstall: safetensors Found existing installation: safetensors 0.4.2 Uninstalling safetensors-0.4.2: Successfully uninstalled safetensors-0.4.2 Attempting uninstall: requests Found existing installation: requests 2.31.0 Uninstalling requests-2.31.0: Successfully uninstalled requests-2.31.0 Attempting uninstall: pyarrow Found existing installation: pyarrow 15.0.0 Uninstalling pyarrow-15.0.0: Successfully uninstalled pyarrow-15.0.0 Attempting uninstall: huggingface-hub Found existing installation: huggingface-hub 0.20.3 Uninstalling huggingface-hub-0.20.3: Successfully uninstalled huggingface-hub-0.20.3 Attempting uninstall: accelerate Found existing installation: accelerate 0.25.0 Uninstalling accelerate-0.25.0: Successfully uninstalled accelerate-0.25.0 Attempting uninstall: datasets Found existing installation: datasets 2.16.0 Uninstalling datasets-2.16.0: Successfully uninstalled datasets-2.16.0 ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts. tikit 1.8.2.240926 requires dicttoxml==1.7.4, which is not installed. tikit 1.8.2.240926 requires docopt==0.6.2, which is not installed. tikit 1.8.2.240926 requires future==0.18.2, which is not installed. tikit 1.8.2.240926 requires hdfs==2.6.0, which is not installed. tikit 1.8.2.240926 requires pure-sasl==0.6.2, which is not installed. tikit 1.8.2.240926 requires py4j==0.10.7, which is not installed. tikit 1.8.2.240926 requires PyHive[hive]==0.6.4, which is not installed. tikit 1.8.2.240926 requires pyjwt>=2.4.0, which is not installed. tikit 1.8.2.240926 requires requests-kerberos>=0.14.0, which is not installed. tikit 1.8.2.240926 requires sasl==0.3.1, which is not installed. tikit 1.8.2.240926 requires thrift==0.15.0, which is not installed. tikit 1.8.2.240926 requires thrift-sasl>=0.1.0, which is not installed. tikit 1.8.2.240926 requires certifi==2021.10.8, but you have certifi 2024.2.2 which is incompatible. tikit 1.8.2.240926 requires cos-python-sdk-v5==1.9.29, but you have cos-python-sdk-v5 1.9.26 which is incompatible. tikit 1.8.2.240926 requires idna==3.3, but you have idna 3.6 which is incompatible. tikit 1.8.2.240926 requires prettytable==2.5.0, but you have prettytable 3.11.0 which is incompatible. tikit 1.8.2.240926 requires urllib3==1.26.7, but you have urllib3 1.26.18 which is incompatible. tikit 1.8.2.240926 requires wcwidth==0.2.5, but you have wcwidth 0.2.13 which is incompatible. Successfully installed accelerate-0.34.2 aniso8601-9.0.1 blinker-1.8.2 datasets-3.0.0 decord-0.6.0 expecttest-0.2.1 flask-3.0.3 flask_restful-0.3.10 huggingface-hub-0.26.2 itsdangerous-2.2.0 natsort-8.4.0 pyarrow-17.0.0 pybind11-2.13.6 requests-2.32.3 safetensors-0.4.5 tiktoken-0.7.0 tqdm-4.67.1 xlsxwriter-3.2.0 WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv ++ return 0 + MEGATRON_DIR=/local_disk/cognitron_vl//third_party/Megatron-LM_core_r0.6.0/ + MINDSPEED_DIR=/local_disk/cognitron_vl//third_party/MindSpeed_core_r0.6.0/ + MODELLINK_DIR=/local_disk/cognitron_vl//third_party/ModelLink/ + pip3 install --no-index --find-links=/data/software/ -e /local_disk/cognitron_vl//third_party/Megatron-LM_core_r0.6.0/ Looking in links: /data/software/ Obtaining file://local_disk/cognitron_vl/third_party/Megatron-LM_core_r0.6.0 Installing build dependencies: started Installing build dependencies: finished with status 'done' Checking if build backend supports build_editable: started Checking if build backend supports build_editable: finished with status 'done' Getting requirements to build editable: started Getting requirements to build editable: finished with status 'done' Installing backend dependencies: started Installing backend dependencies: finished with status 'done' Preparing editable metadata (pyproject.toml): started Preparing editable metadata (pyproject.toml): finished with status 'done' Building wheels for collected packages: megatron_core Building editable for megatron_core (pyproject.toml): started Building editable for megatron_core (pyproject.toml): finished with status 'done' Created wheel for megatron_core: filename=megatron_core-0.6.0-0.editable-cp38-cp38-linux_x86_64.whl size=8791 sha256=45264bd466d16535ad2060a555b895f6f1ef1411bdb5aac2bc508966f4578e95 Stored in directory: /tmp/pip-ephem-wheel-cache-ibh5mvv_/wheels/54/9c/d1/d2015aa0c34e791e64d65d19395e5a9a5528f0c63fd519b9ff Successfully built megatron_core DEPRECATION: apex 0.1-ascend-20240523 has a non-standard version number. pip 24.1 will enforce this behaviour change. A possible replacement is to upgrade to a newer version of apex or contact the author to suggest that they release a version with a conforming version number. Discussion can be found at https://github.com/pypa/pip/issues/12063 Installing collected packages: megatron_core Successfully installed megatron_core-0.6.0 WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv + pip3 install --no-index --find-links=/data/software/ -e /local_disk/cognitron_vl//third_party/MindSpeed_core_r0.6.0/ Looking in links: /data/software/ Obtaining file://local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0 Preparing metadata (setup.py): started Preparing metadata (setup.py): finished with status 'done' WARNING: Error parsing requirements for tokenizers: [Errno 2] No such file or directory: '/root/miniconda3/envs/py38/lib/python3.8/site-packages/tokenizers-0.19.1.dist-info/METADATA' WARNING: Error parsing requirements for transformers: [Errno 2] No such file or directory: '/root/miniconda3/envs/py38/lib/python3.8/site-packages/transformers-4.40.1.dist-info/METADATA' DEPRECATION: apex 0.1-ascend-20240523 has a non-standard version number. pip 24.1 will enforce this behaviour change. A possible replacement is to upgrade to a newer version of apex or contact the author to suggest that they release a version with a conforming version number. Discussion can be found at https://github.com/pypa/pip/issues/12063 Installing collected packages: mindspeed Running setup.py develop for mindspeed Successfully installed mindspeed-0.6.0 WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv + pip3 install --no-index --find-links=/data/software/ -e /local_disk/cognitron_vl//third_party/ModelLink/ Looking in links: /data/software/ Obtaining file://local_disk/cognitron_vl/third_party/ModelLink Preparing metadata (setup.py): started Preparing metadata (setup.py): finished with status 'done' Requirement already satisfied: numpy in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (1.24.4) Processing data/software/transformers-4.43.2-py3-none-any.whl (from modellink==0.0.1) Processing data/software/transformers-stream-generator-0.0.5.tar.gz (from modellink==0.0.1) Preparing metadata (setup.py): started Preparing metadata (setup.py): finished with status 'done' Requirement already satisfied: sympy in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (1.4) Requirement already satisfied: decorator in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (5.1.1) Requirement already satisfied: scipy in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (1.10.1) Requirement already satisfied: sentencepiece in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (0.2.0) Requirement already satisfied: einops in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (0.7.0) Requirement already satisfied: datasets in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (3.0.0) Requirement already satisfied: pybind11 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (2.13.6) Requirement already satisfied: accelerate in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (0.34.2) Requirement already satisfied: six in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (1.16.0) Requirement already satisfied: protobuf in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (4.25.3) Processing data/software/peft-0.7.1-py3-none-any.whl (from modellink==0.0.1) Requirement already satisfied: tiktoken in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (0.7.0) Requirement already satisfied: packaging>=20.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft==0.7.1->modellink==0.0.1) (23.2) Requirement already satisfied: psutil in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft==0.7.1->modellink==0.0.1) (5.9.8) Requirement already satisfied: pyyaml in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft==0.7.1->modellink==0.0.1) (5.4.1) Requirement already satisfied: torch>=1.13.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft==0.7.1->modellink==0.0.1) (2.1.0+cpu) Requirement already satisfied: tqdm in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft==0.7.1->modellink==0.0.1) (4.67.1) Requirement already satisfied: safetensors in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft==0.7.1->modellink==0.0.1) (0.4.5) Requirement already satisfied: huggingface-hub>=0.17.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft==0.7.1->modellink==0.0.1) (0.26.2) Requirement already satisfied: filelock in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from transformers==4.43.2->modellink==0.0.1) (3.13.1) Requirement already satisfied: regex!=2019.12.17 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from transformers==4.43.2->modellink==0.0.1) (2023.12.25) Requirement already satisfied: requests in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from transformers==4.43.2->modellink==0.0.1) (2.32.3) Processing data/software/tokenizers-0.19.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (from transformers==4.43.2->modellink==0.0.1) Requirement already satisfied: pyarrow>=15.0.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets->modellink==0.0.1) (17.0.0) Requirement already satisfied: dill<0.3.9,>=0.3.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets->modellink==0.0.1) (0.3.7) Requirement already satisfied: pandas in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets->modellink==0.0.1) (2.0.3) Requirement already satisfied: xxhash in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets->modellink==0.0.1) (3.4.1) Requirement already satisfied: multiprocess in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets->modellink==0.0.1) (0.70.15) Requirement already satisfied: fsspec<=2024.6.1,>=2023.1.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from fsspec[http]<=2024.6.1,>=2023.1.0->datasets->modellink==0.0.1) (2023.10.0) Requirement already satisfied: aiohttp in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets->modellink==0.0.1) (3.9.3) Requirement already satisfied: mpmath>=0.19 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from sympy->modellink==0.0.1) (1.3.0) Requirement already satisfied: aiosignal>=1.1.2 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets->modellink==0.0.1) (1.3.1) Requirement already satisfied: attrs>=17.3.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets->modellink==0.0.1) (23.2.0) Requirement already satisfied: frozenlist>=1.1.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets->modellink==0.0.1) (1.4.1) Requirement already satisfied: multidict<7.0,>=4.5 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets->modellink==0.0.1) (6.0.5) Requirement already satisfied: yarl<2.0,>=1.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets->modellink==0.0.1) (1.9.4) Requirement already satisfied: async-timeout<5.0,>=4.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets->modellink==0.0.1) (4.0.3) Requirement already satisfied: typing-extensions>=3.7.4.3 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from huggingface-hub>=0.17.0->peft==0.7.1->modellink==0.0.1) (4.10.0) Requirement already satisfied: charset-normalizer<4,>=2 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from requests->transformers==4.43.2->modellink==0.0.1) (3.3.2) Requirement already satisfied: idna<4,>=2.5 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from requests->transformers==4.43.2->modellink==0.0.1) (3.6) Requirement already satisfied: urllib3<3,>=1.21.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from requests->transformers==4.43.2->modellink==0.0.1) (1.26.18) Requirement already satisfied: certifi>=2017.4.17 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from requests->transformers==4.43.2->modellink==0.0.1) (2024.2.2) Requirement already satisfied: networkx in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from torch>=1.13.0->peft==0.7.1->modellink==0.0.1) (3.1) Requirement already satisfied: jinja2 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from torch>=1.13.0->peft==0.7.1->modellink==0.0.1) (3.1.3) Requirement already satisfied: python-dateutil>=2.8.2 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from pandas->datasets->modellink==0.0.1) (2.8.2) Requirement already satisfied: pytz>=2020.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from pandas->datasets->modellink==0.0.1) (2024.1) Requirement already satisfied: tzdata>=2022.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from pandas->datasets->modellink==0.0.1) (2024.1) Requirement already satisfied: MarkupSafe>=2.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from jinja2->torch>=1.13.0->peft==0.7.1->modellink==0.0.1) (2.1.5) Building wheels for collected packages: transformers_stream_generator Building wheel for transformers_stream_generator (setup.py): started Building wheel for transformers_stream_generator (setup.py): finished with status 'done' Created wheel for transformers_stream_generator: filename=transformers_stream_generator-0.0.5-py3-none-any.whl size=12425 sha256=7b79cc70e3e25aee266f004baeb02f49a26365cbf13fc2e075375f124a308963 Stored in directory: /root/.cache/pip/wheels/56/8c/42/5381d9c36bc85f28982f4cf8f98dc44d37a6d6c04897a5cb7c Successfully built transformers_stream_generator DEPRECATION: apex 0.1-ascend-20240523 has a non-standard version number. pip 24.1 will enforce this behaviour change. A possible replacement is to upgrade to a newer version of apex or contact the author to suggest that they release a version with a conforming version number. Discussion can be found at https://github.com/pypa/pip/issues/12063 Installing collected packages: tokenizers, transformers, transformers_stream_generator, peft, modellink Attempting uninstall: tokenizers Found existing installation: tokenizers 0.20.3 Uninstalling tokenizers-0.20.3: Successfully uninstalled tokenizers-0.20.3 Attempting uninstall: transformers Found existing installation: transformers 4.46.3 Uninstalling transformers-4.46.3: Successfully uninstalled transformers-4.46.3 Attempting uninstall: peft Found existing installation: peft 0.7.0 Uninstalling peft-0.7.0: Successfully uninstalled peft-0.7.0 Running setup.py develop for modellink Successfully installed modellink-0.0.1 peft-0.7.1 tokenizers-0.19.1 transformers-4.43.2 transformers_stream_generator-0.0.5 WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv + export PYTHONPATH=/local_disk/cognitron_vl//third_party/Megatron-LM_core_r0.6.0//:/usr/local/Ascend/ascend-toolkit/latest/python/site-packages:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe:/usr/local/Ascend/ascend-toolkit/latest/python/site-packages:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe: + PYTHONPATH=/local_disk/cognitron_vl//third_party/Megatron-LM_core_r0.6.0//:/usr/local/Ascend/ascend-toolkit/latest/python/site-packages:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe:/usr/local/Ascend/ascend-toolkit/latest/python/site-packages:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe: + GPUS_PER_NODE=16 + NNODES=32 + NODE_RANK=1 + MASTER_PORT=34567 + export CUDA_DEVICE_MAX_CONNECTIONS=1 + CUDA_DEVICE_MAX_CONNECTIONS=1 + export PYTORCH_NPU_ALLOC_CONF=expandable_segments:True + PYTORCH_NPU_ALLOC_CONF=expandable_segments:True + VISION_SEQ_LENGTH=1025 + IMAGE_TOKEN_LENGTH=256 + IMAGE_SIZE=448 + VISION_MODEL_TYPE=intern_300m + TP=8 + PP=1 + CP=8 + CP_ALGO=megatron_cp_algo + CP_MASK=causal + DISTRIBUTED_ARGS=' --nproc_per_node 16 --nnodes 32 --node_rank 1 --master_addr train-1198772881325351168-93vlj4s2getc-master-0.train-100034032793.svc.cluster.local --master_port 34567 ' + GPT_ARGS=' --use-mcore-models --tensor-model-parallel-size 8 --pipeline-model-parallel-size 1 --context-parallel-size 8 --context-parallel-algo megatron_cp_algo --cp-attention-mask-type causal --use-cp-send-recv-overlap --no-create-attention-mask-in-dataloader --sparse-mode 4 --sequence-parallel --recompute-method block --recompute-granularity full --recompute-num-layers 48 --num-layers 48 --hidden-size 5120 --ffn-hidden-size 13824 --num-attention-heads 40 --group-query-attention --num-query-groups 8 --tokenizer-type PretrainedFromHF --tokenizer-name-or-path /data_4/models/Qwen/Qwen2.5-14B-Instruct/ --seq-length 1048576 --max-position-embeddings 1048576 --micro-batch-size 1 --global-batch-size 8 --make-vocab-size-divisible-by 1 --padded-vocab-size 152064 --rotary-base 1000000.0 --lr 5.00e-6 --train-iters 500 --lr-decay-style cosine --untie-embeddings-and-output-weights --disable-bias-linear --attention-dropout 0.0 --init-method-std 0.01 --hidden-dropout 0.0 --position-embedding-type rope --normalization RMSNorm --use-fused-rmsnorm --norm-epsilon 1e-6 --swiglu --use-flash-attn --use-fused-rotary-pos-emb --use-rotary-position-embeddings --use-fused-swiglu --use-mc2 --no-masked-softmax-fusion --attention-softmax-in-fp32 --min-lr 1.00e-7 --weight-decay 0.0 --lr-warmup-fraction 0.03 --clip-grad 1.0 --adam-beta1 0.9 --adam-beta2 0.999 --add-qkv-bias --initial-loss-scale 4096 --no-gradient-accumulation-fusion --use-distributed-optimizer --bf16 --overlap-grad-reduce --finetune --vision-model-freeze --vision-model-type intern_300m --vision-downsample-ratio 0.5 --vision-projector-type mlp --vision-projector-pre-norm --vision-process-type dynamic --vision-normalize-type imagenet --vision-seq-length 1025 --image-token-length 256 --image-size 448 --prompt-format qwen2 --is-instruction-dataset --max-num-image 4096 --max-fps 1 --add-class-token --min-patch-grid 1 --max-patch-grid 12 --logit-mask --cross-dataset-joint ' + DATA_ARGS=' --data-path /local_disk/cognitron_vl//configs/lcvlm_finetune_stage4.yaml --split 100,0,0 --data-seq-length 1048576 --num-workers 8 ' + CKPT_ARGS=' --load /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/ --vit-load / --no-load-optim --no-load-rng --seed 42424242 --save /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743// ' + OUTPUT_ARGS=' --log-interval 1 --save-interval 20 --eval-interval 20 --eval-iters 0 --log-throughput --distributed-timeout-minutes 120 ' + torchrun --nproc_per_node 16 --nnodes 32 --node_rank 1 --master_addr train-1198772881325351168-93vlj4s2getc-master-0.train-100034032793.svc.cluster.local --master_port 34567 /local_disk/cognitron_vl//lcvlm_modellink/pretrain_lcvlm.py --use-mcore-models --tensor-model-parallel-size 8 --pipeline-model-parallel-size 1 --context-parallel-size 8 --context-parallel-algo megatron_cp_algo --cp-attention-mask-type causal --use-cp-send-recv-overlap --no-create-attention-mask-in-dataloader --sparse-mode 4 --sequence-parallel --recompute-method block --recompute-granularity full --recompute-num-layers 48 --num-layers 48 --hidden-size 5120 --ffn-hidden-size 13824 --num-attention-heads 40 --group-query-attention --num-query-groups 8 --tokenizer-type PretrainedFromHF --tokenizer-name-or-path /data_4/models/Qwen/Qwen2.5-14B-Instruct/ --seq-length 1048576 --max-position-embeddings 1048576 --micro-batch-size 1 --global-batch-size 8 --make-vocab-size-divisible-by 1 --padded-vocab-size 152064 --rotary-base 1000000.0 --lr 5.00e-6 --train-iters 500 --lr-decay-style cosine --untie-embeddings-and-output-weights --disable-bias-linear --attention-dropout 0.0 --init-method-std 0.01 --hidden-dropout 0.0 --position-embedding-type rope --normalization RMSNorm --use-fused-rmsnorm --norm-epsilon 1e-6 --swiglu --use-flash-attn --use-fused-rotary-pos-emb --use-rotary-position-embeddings --use-fused-swiglu --use-mc2 --no-masked-softmax-fusion --attention-softmax-in-fp32 --min-lr 1.00e-7 --weight-decay 0.0 --lr-warmup-fraction 0.03 --clip-grad 1.0 --adam-beta1 0.9 --adam-beta2 0.999 --add-qkv-bias --initial-loss-scale 4096 --no-gradient-accumulation-fusion --use-distributed-optimizer --bf16 --overlap-grad-reduce --finetune --vision-model-freeze --vision-model-type intern_300m --vision-downsample-ratio 0.5 --vision-projector-type mlp --vision-projector-pre-norm --vision-process-type dynamic --vision-normalize-type imagenet --vision-seq-length 1025 --image-token-length 256 --image-size 448 --prompt-format qwen2 --is-instruction-dataset --max-num-image 4096 --max-fps 1 --add-class-token --min-patch-grid 1 --max-patch-grid 12 --logit-mask --cross-dataset-joint --data-path /local_disk/cognitron_vl//configs/lcvlm_finetune_stage4.yaml --split 100,0,0 --data-seq-length 1048576 --num-workers 8 --load /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/ --vit-load / --no-load-optim --no-load-rng --seed 42424242 --save /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743// --log-interval 1 --save-interval 20 --eval-interval 20 --eval-iters 0 --log-throughput --distributed-timeout-minutes 120 --distributed-backend nccl [2024-11-28 15:50:27,179] torch.distributed.run: [WARNING] [2024-11-28 15:50:27,179] torch.distributed.run: [WARNING] ***************************************** [2024-11-28 15:50:27,179] torch.distributed.run: [WARNING] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. [2024-11-28 15:50:27,179] torch.distributed.run: [WARNING] ***************************************** Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root...Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Creating extension directory /root/.cache/torch_extensions/py38_cpu/adaptive_cp...Creating extension directory /root/.cache/torch_extensions/py38_cpu/adaptive_cp... Emitting ninja build file /root/.cache/torch_extensions/py38_cpu/adaptive_cp/build.ninja... Building extension module adaptive_cp... Allowing ninja to set a default number of workers... (overridable by setting the environment variable MAX_JOBS=N) Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... [1/2] c++ -MMD -MF adaptive_cp.o.d -DTORCH_EXTENSION_NAME=adaptive_cp -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE=\"_gcc\" -DPYBIND11_STDLIB=\"_libstdcpp\" -DPYBIND11_BUILD_ABI=\"_cxxabi1011\" -I/usr/local/Ascend/ascend-toolkit/latest/include -I/root/miniconda3/envs/py38/lib/python3.8/site-packages/torch_npu/include -I/root/miniconda3/envs/py38/lib/python3.8/site-packages/torch_npu/third_party -I/root/miniconda3/envs/py38/lib/python3.8/site-packages/torch_npu/acl -I/root/miniconda3/envs/py38/lib/python3.8/site-packages/torch_npu/inc -isystem /root/miniconda3/envs/py38/lib/python3.8/site-packages/torch/include -isystem /root/miniconda3/envs/py38/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -isystem /root/miniconda3/envs/py38/lib/python3.8/site-packages/torch/include/TH -isystem /root/miniconda3/envs/py38/lib/python3.8/site-packages/torch/include/THC -isystem /root/miniconda3/envs/py38/include/python3.8 -D_GLIBCXX_USE_CXX11_ABI=0 -fPIC -std=c++17 -fstack-protector-all -Wl,-z,relro,-z,now,-z,noexecstack -fPIC -pie -Wl,--disable-new-dtags,--rpath -s -O2 -c local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/ops/csrc/algorithm/adaptive_cp/adaptive_cp.cpp -o adaptive_cp.o [2/2] c++ adaptive_cp.o -shared -L/usr/local/Ascend/ascend-toolkit/latest/lib64 -lascendcl -L/root/miniconda3/envs/py38/lib/python3.8/site-packages/torch_npu/lib -ltorch_npu -L/root/miniconda3/envs/py38/lib/python3.8/site-packages/torch/lib -lc10 -ltorch_cpu -ltorch -ltorch_python -o adaptive_cp.so Loading extension module adaptive_cp... Loading extension module adaptive_cp... local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") Loading extension module adaptive_cp... local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") Loading extension module adaptive_cp... local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") Loading extension module adaptive_cp... Loading extension module adaptive_cp... Loading extension module adaptive_cp... Loading extension module adaptive_cp... local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") Loading extension module adaptive_cp... Loading extension module adaptive_cp... local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") Loading extension module adaptive_cp... Loading extension module adaptive_cp... Loading extension module adaptive_cp... local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") Loading extension module adaptive_cp... local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") Loading extension module adaptive_cp... Loading extension module adaptive_cp... local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( > compiling dataset index builder ... make: Entering directory 'local_disk/cognitron_vl/third_party/Megatron-LM_core_r0.6.0/megatron/core/datasets' make: Nothing to be done for 'default'. make: Leaving directory 'local_disk/cognitron_vl/third_party/Megatron-LM_core_r0.6.0/megatron/core/datasets' >>> done with dataset index builder. Compilation time: 0.273 seconds vision_projector_recompute Falsevision_projector_recompute False vision_projector_recompute Falsevision_projector_recompute Falsevision_projector_recompute False vision_projector_recompute False vision_projector_recompute Falsevision_projector_recompute False vision_projector_recompute False vision_projector_recompute False vision_projector_recompute False vision_projector_recompute False vision_projector_recompute False vision_projector_recompute False vision_projector_recompute False vision_projector_recompute False vision_model_freeze => set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False. => set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False. => set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. vision_model_freeze => set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False. => set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False. => set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.vision_model_freeze => set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False. => set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. vision_model_freeze => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. vision_model_freeze=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.vision_model_freeze=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False.vision_model_freeze => set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.vision_model_freeze=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False. model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() ) => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False. vision_model_freeze=> set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False.vision_model_freeze => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. vision_model_freeze => set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. vision_model_freeze => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. > number of parameters on (tensor, pipeline) model parallel rank (1, 0): 1887497216=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False. vision_model_freeze=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() )=> set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False.model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() )=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.vision_model_freeze=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() )=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() ) => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() ) => set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() ) => set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() )=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. > number of parameters on (tensor, pipeline) model parallel rank (1, 0): 1887497216 model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() ) model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() ) > number of parameters on (tensor, pipeline) model parallel rank (2, 0): 1887497216 model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() ) > number of parameters on (tensor, pipeline) model parallel rank (0, 0): 1887497216 vision_model_freeze => set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False. => set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False.model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() ) > number of parameters on (tensor, pipeline) model parallel rank (3, 0): 1887497216 => set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False. > number of parameters on (tensor, pipeline) model parallel rank (4, 0): 1887497216 => set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. > number of parameters on (tensor, pipeline) model parallel rank (5, 0): 1887497216=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False. model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() ) => set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False. model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() )=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False. > number of parameters on (tensor, pipeline) model parallel rank (0, 0): 1887497216 => set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. vision_model_freeze=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False. => set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. > number of parameters on (tensor, pipeline) model parallel rank (4, 0): 1887497216=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. > number of parameters on (tensor, pipeline) model parallel rank (6, 0): 1887497216=> set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. > number of parameters on (tensor, pipeline) model parallel rank (7, 0): 1887497216 => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. > number of parameters on (tensor, pipeline) model parallel rank (2, 0): 1887497216 > number of parameters on (tensor, pipeline) model parallel rank (7, 0): 1887497216 > number of parameters on (tensor, pipeline) model parallel rank (6, 0): 1887497216 model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() ) model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() ) > number of parameters on (tensor, pipeline) model parallel rank (5, 0): 1887497216 > number of parameters on (tensor, pipeline) model parallel rank (3, 0): 1887497216 _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False) _load_base_checkpoint iteration_load_base_checkpoint iteration_load_base_checkpoint iteration_load_base_checkpoint iteration_load_base_checkpoint iteration_load_base_checkpoint iteration_load_base_checkpoint iteration_load_base_checkpoint iteration_load_base_checkpoint iteration _load_base_checkpoint iteration_load_base_checkpoint iteration_load_base_checkpoint iteration _load_base_checkpoint iteration 1000 _load_base_checkpoint iteration_load_base_checkpoint iteration_load_base_checkpoint release1000 False1000 1000 1000 _load_base_checkpoint iteration 1000 10001000 1000 1000 10001000_load_base_checkpoint release 1000 1000_load_base_checkpoint release_load_base_checkpoint release1000_load_base_checkpoint release_load_base_checkpoint release False_load_base_checkpoint release _load_base_checkpoint releaseFalse_load_base_checkpoint release _load_base_checkpoint release _load_base_checkpoint release _load_base_checkpoint releaseFalse_load_base_checkpoint release1000 _load_base_checkpoint release False FalseFalseFalse False_load_base_checkpoint release FalseFalseFalse False False _load_base_checkpoint release False False _load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_02/model_optim_rng.pt _load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_01/model_optim_rng.pt _load_base_checkpoint_load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_07/model_optim_rng.pt/data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_01/model_optim_rng.pt _load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_04/model_optim_rng.pt _load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_03/model_optim_rng.pt _load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_04/model_optim_rng.pt _load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_00/model_optim_rng.pt _load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_00/model_optim_rng.pt _load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_07/model_optim_rng.pt _load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_05/model_optim_rng.pt _load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_02/model_optim_rng.pt _load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_03/model_optim_rng.pt _load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_06/model_optim_rng.pt _load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_06/model_optim_rng.pt _load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_05/model_optim_rng.pt load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True > rank 18 does not create GPT datasets ...> rank 31 does not create GPT datasets ... > rank 30 does not create GPT datasets ... > rank 21 does not create GPT datasets ... > rank 26 does not create GPT datasets ...> rank 23 does not create GPT datasets ... > rank 16 is creating GPT datasets ... > rank 19 does not create GPT datasets ... > rank 24 is creating GPT datasets ... > rank 22 does not create GPT datasets ... > rank 29 does not create GPT datasets ... > rank 28 does not create GPT datasets ... > rank 17 does not create GPT datasets ... > rank 27 does not create GPT datasets ... > rank 20 does not create GPT datasets ... > rank 25 does not create GPT datasets ... target_ratios [(1, 1), (1, 2), (2, 1), (3, 1), (1, 3), (2, 2), (4, 1), (1, 4), (5, 1), (1, 5), (1, 6), (6, 1), (3, 2), (2, 3), (7, 1), (1, 7), (4, 2), (2, 4), (1, 8), (8, 1), (1, 9), (3, 3), (9, 1), (2, 5), (5, 2), (10, 1), (1, 10), (11, 1), (1, 11), (12, 1), (3, 4), (4, 3), (1, 12), (6, 2), (2, 6)] possible_resolutions [[448, 448], [448, 896], [896, 448], [1344, 448], [448, 1344], [896, 896], [1792, 448], [448, 1792], [2240, 448], [448, 2240], [448, 2688], [2688, 448], [1344, 896], [896, 1344], [3136, 448], [448, 3136], [1792, 896], [896, 1792], [448, 3584], [3584, 448], [448, 4032], [1344, 1344], [4032, 448], [896, 2240], [2240, 896], [4480, 448], [448, 4480], [4928, 448], [448, 4928], [5376, 448], [1344, 1792], [1792, 1344], [448, 5376], [2688, 896], [896, 2688]] target_ratios [(1, 1), (1, 2), (2, 1), (3, 1), (1, 3), (2, 2), (4, 1), (1, 4), (5, 1), (1, 5), (1, 6), (6, 1), (3, 2), (2, 3), (7, 1), (1, 7), (4, 2), (2, 4), (1, 8), (8, 1), (1, 9), (3, 3), (9, 1), (2, 5), (5, 2), (10, 1), (1, 10), (11, 1), (1, 11), (12, 1), (3, 4), (4, 3), (1, 12), (6, 2), (2, 6)] possible_resolutions [[448, 448], [448, 896], [896, 448], [1344, 448], [448, 1344], [896, 896], [1792, 448], [448, 1792], [2240, 448], [448, 2240], [448, 2688], [2688, 448], [1344, 896], [896, 1344], [3136, 448], [448, 3136], [1792, 896], [896, 1792], [448, 3584], [3584, 448], [448, 4032], [1344, 1344], [4032, 448], [896, 2240], [2240, 896], [4480, 448], [448, 4480], [4928, 448], [448, 4928], [5376, 448], [1344, 1792], [1792, 1344], [448, 5376], [2688, 896], [896, 2688]] [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617ed3f3040] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure processed_samples 100 unjoint_samples 100 joint_samples 0 [136617, 135888] processed_samples 100 unjoint_samples 100 joint_samples 0 [136617, 135888] processed_samples 100 unjoint_samples 100 joint_samples 0 [151530, 153863] processed_samples 100 unjoint_samples 100 joint_samples 0 [151530, 153863] processed_samples 100 unjoint_samples 100 joint_samples 0 [169920, 169349] processed_samples 100 unjoint_samples 100 joint_samples 0 [169920, 169349] processed_samples 100 unjoint_samples 100 joint_samples 0 [182200, 178748] processed_samples 100 unjoint_samples 100 joint_samples 0 [182200, 178748] processed_samples 100 unjoint_samples 100 joint_samples 0 [118939, 117033] processed_samples 100 unjoint_samples 100 joint_samples 0 [118939, 117033] processed_samples 100 unjoint_samples 100 joint_samples 0 [147974, 136346] processed_samples 100 unjoint_samples 100 joint_samples 0 [147974, 136346] processed_samples 100 unjoint_samples 100 joint_samples 0 [133776, 193755] processed_samples 100 unjoint_samples 100 joint_samples 0 [133776, 193755] processed_samples 100 unjoint_samples 100 joint_samples 0 [170928, 172080] processed_samples 100 unjoint_samples 100 joint_samples 0 [170928, 172080] [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617e803bd40] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617ec009d00] mmco: unref short failure [h264 @ 0x5617ec009d00] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617e8964000] mmco: unref short failure processed_samples 200 unjoint_samples 200 joint_samples 0 [272600, 263494] processed_samples 200 unjoint_samples 200 joint_samples 0 [272600, 263494] processed_samples 200 unjoint_samples 200 joint_samples 0 [350560, 317454] processed_samples 200 unjoint_samples 200 joint_samples 0 [350560, 317454] processed_samples 200 unjoint_samples 200 joint_samples 0 [316234, 316044] processed_samples 200 unjoint_samples 200 joint_samples 0 [297706, 300263] processed_samples 200 unjoint_samples 200 joint_samples 0 [316234, 316044] processed_samples 200 unjoint_samples 200 joint_samples 0 [297706, 300263] processed_samples 200 unjoint_samples 200 joint_samples 0 [304906, 302332] processed_samples 200 unjoint_samples 200 joint_samples 0 [304906, 302332] processed_samples 200 unjoint_samples 200 joint_samples 0 [290433, 286771] processed_samples 200 unjoint_samples 200 joint_samples 0 [290433, 286771] processed_samples 200 unjoint_samples 200 joint_samples 0 [279295, 291940] processed_samples 200 unjoint_samples 200 joint_samples 0 [335236, 338160] processed_samples 200 unjoint_samples 200 joint_samples 0 [335236, 338160] processed_samples 200 unjoint_samples 200 joint_samples 0 [279295, 291940] [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure processed_samples 300 unjoint_samples 300 joint_samples 0 [458375, 466209] processed_samples 300 unjoint_samples 300 joint_samples 0 [458375, 466209] processed_samples 300 unjoint_samples 300 joint_samples 0 [443995, 449374] processed_samples 300 unjoint_samples 300 joint_samples 0 [443995, 449374] processed_samples 300 unjoint_samples 300 joint_samples 0 [448329, 445968] processed_samples 300 unjoint_samples 300 joint_samples 0 [448329, 445968] processed_samples 300 unjoint_samples 300 joint_samples 0 [467997, 468282] processed_samples 300 unjoint_samples 300 joint_samples 0 [467997, 468282] processed_samples 300 unjoint_samples 300 joint_samples 0 [545900, 484895] processed_samples 300 unjoint_samples 300 joint_samples 0 [545900, 484895] processed_samples 300 unjoint_samples 300 joint_samples 0 [403370, 402945] processed_samples 300 unjoint_samples 300 joint_samples 0 [403370, 402945] processed_samples 300 unjoint_samples 300 joint_samples 0 [475714, 474207] processed_samples 300 unjoint_samples 300 joint_samples 0 [475714, 474207] processed_samples 300 unjoint_samples 300 joint_samples 0 [440570, 438788] processed_samples 300 unjoint_samples 300 joint_samples 0 [440570, 438788] [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure processed_samples 400 unjoint_samples 400 joint_samples 0 [614095, 611512] processed_samples 400 unjoint_samples 400 joint_samples 0 [614095, 611512] processed_samples 400 unjoint_samples 400 joint_samples 0 [720038, 719894] processed_samples 400 unjoint_samples 400 joint_samples 0 [720038, 719894] processed_samples 400 unjoint_samples 400 joint_samples 0 [739305, 788503] processed_samples 400 unjoint_samples 400 joint_samples 0 [739305, 788503] processed_samples 400 unjoint_samples 400 joint_samples 0 [603932, 593783] processed_samples 400 unjoint_samples 400 joint_samples 0 [603932, 593783] processed_samples 400 unjoint_samples 400 joint_samples 0 [594854, 594937] processed_samples 400 unjoint_samples 400 joint_samples 0 [594854, 594937] processed_samples 400 unjoint_samples 400 joint_samples 0 [525518, 551856] processed_samples 400 unjoint_samples 400 joint_samples 0 [525518, 551856] processed_samples 400 unjoint_samples 400 joint_samples 0 [747235, 752195] processed_samples 400 unjoint_samples 400 joint_samples 0 [747235, 752195] processed_samples 400 unjoint_samples 400 joint_samples 0 [634439, 634523] processed_samples 400 unjoint_samples 400 joint_samples 0 [634439, 634523] [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure processed_samples 500 unjoint_samples 500 joint_samples 0 [953537, 952984] processed_samples 500 unjoint_samples 500 joint_samples 0 [953537, 952984] [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure processed_samples 500 unjoint_samples 500 joint_samples 0 [765674, 765089] processed_samples 500 unjoint_samples 500 joint_samples 0 [765674, 765089] [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure processed_samples 500 unjoint_samples 500 joint_samples 0 [697675, 698988] [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure processed_samples 500 unjoint_samples 500 joint_samples 0 [697675, 698988] [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure processed_samples 500 unjoint_samples 500 joint_samples 0 [873827, 875180] processed_samples 500 unjoint_samples 500 joint_samples 0 [873827, 875180] processed_samples 500 unjoint_samples 500 joint_samples 0 [899415, 901075] processed_samples 500 unjoint_samples 500 joint_samples 0 [899415, 901075] processed_samples 500 unjoint_samples 500 joint_samples 0 [746152, 746635] processed_samples 500 unjoint_samples 500 joint_samples 0 [746152, 746635] processed_samples 500 unjoint_samples 500 joint_samples 0 [762804, 760270] processed_samples 500 unjoint_samples 500 joint_samples 0 [762804, 760270] processed_samples 500 unjoint_samples 500 joint_samples 0 [781344, 781555] processed_samples 500 unjoint_samples 500 joint_samples 0 [781344, 781555] [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5596093d14c0] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617ebcdff40] mmco: unref short failure [h264 @ 0x5617ebcdff40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure ................................................................................................[Rank 27] (after 1 iterations) memory (MB) | allocated: 15488.0283203125 | max allocated: 31429.67822265625 | reserved: 33014.0 | max reserved: 33014.0 [Rank 29] (after 1 iterations) memory (MB) | allocated: 15488.0283203125 | max allocated: 31429.67822265625 | reserved: 33014.0 | max reserved: 33014.0 [Rank 28] (after 1 iterations) memory (MB) | allocated: 15488.0283203125 | max allocated: 31429.67822265625 | reserved: 33014.0 | max reserved: 33014.0 [Rank 30] (after 1 iterations) memory (MB) | allocated: 15488.0283203125 | max allocated: 31429.67822265625 | reserved: 33014.0 | max reserved: 33014.0 [Rank 25] (after 1 iterations) memory (MB) | allocated: 15488.0283203125 | max allocated: 31429.67822265625 | reserved: 33014.0 | max reserved: 33014.0 [Rank 31] (after 1 iterations) memory (MB) | allocated: 15488.0283203125 | max allocated: 31429.67822265625 | reserved: 33014.0 | max reserved: 33014.0 [Rank 18] (after 1 iterations) memory (MB) | allocated: 15617.53125 | max allocated: 31921.00537109375 | reserved: 33436.0 | max reserved: 33436.0 [Rank 19] (after 1 iterations) memory (MB) | allocated: 15617.53125 | max allocated: 31921.00537109375 | reserved: 33436.0 | max reserved: 33436.0 [Rank 22] (after 1 iterations) memory (MB) | allocated: 15617.53125 | max allocated: 31921.00537109375 | reserved: 33436.0 | max reserved: 33436.0 [Rank 20] (after 1 iterations) memory (MB) | allocated: 15617.53125 | max allocated: 31921.00537109375 | reserved: 33436.0 | max reserved: 33436.0 [Rank 24] (after 1 iterations) memory (MB) | allocated: 15488.0283203125 | max allocated: 31429.67822265625 | reserved: 33014.0 | max reserved: 33014.0[Rank 26] (after 1 iterations) memory (MB) | allocated: 15488.0283203125 | max allocated: 31429.67822265625 | reserved: 33014.0 | max reserved: 33014.0 [Rank 17] (after 1 iterations) memory (MB) | allocated: 15617.53125 | max allocated: 31921.00537109375 | reserved: 33436.0 | max reserved: 33436.0 [Rank 21] (after 1 iterations) memory (MB) | allocated: 15617.53125 | max allocated: 31921.00537109375 | reserved: 33436.0 | max reserved: 33436.0 [Rank 23] (after 1 iterations) memory (MB) | allocated: 15617.53125 | max allocated: 31921.00537109375 | reserved: 33436.0 | max reserved: 33436.0 [Rank 16] (after 1 iterations) memory (MB) | allocated: 15617.53125 | max allocated: 31921.00537109375 | reserved: 33436.0 | max reserved: 33436.0 [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure processed_samples 600 unjoint_samples 600 joint_samples 1 [67071, 1047472] processed_samples 600 unjoint_samples 600 joint_samples 0 [916974, 897727] processed_samples 600 unjoint_samples 600 joint_samples 1 [67071, 1047472] processed_samples 600 unjoint_samples 600 joint_samples 0 [916974, 897727] [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure processed_samples 600 unjoint_samples 600 joint_samples 0 [976521, 983222] processed_samples 600 unjoint_samples 600 joint_samples 0 [885968, 885155] processed_samples 600 unjoint_samples 600 joint_samples 0 [976521, 983222] processed_samples 600 unjoint_samples 600 joint_samples 0 [885968, 885155] processed_samples 600 unjoint_samples 600 joint_samples 0 [909708, 908926] processed_samples 600 unjoint_samples 600 joint_samples 0 [909708, 908926] processed_samples 600 unjoint_samples 600 joint_samples 0 [1031648, 1031813] processed_samples 600 unjoint_samples 600 joint_samples 0 [1031648, 1031813] processed_samples 600 unjoint_samples 600 joint_samples 0 [870105, 869287] processed_samples 600 unjoint_samples 600 joint_samples 0 [870105, 869287] processed_samples 600 unjoint_samples 600 joint_samples 0 [1041639, 1042256] processed_samples 600 unjoint_samples 600 joint_samples 0 [1041639, 1042256] [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617e803bd40] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure processed_samples 700 unjoint_samples 700 joint_samples 1 [1047221, 91603] processed_samples 700 unjoint_samples 700 joint_samples 1 [1047221, 91603] [h264 @ 0x5617e84806c0] mmco: unref short failure processed_samples 700 unjoint_samples 700 joint_samples 1 [116493, 1046583] [h264 @ 0x559613520f80] mmco: unref short failure processed_samples 700 unjoint_samples 700 joint_samples 1 [116493, 1046583] processed_samples 700 unjoint_samples 700 joint_samples 1 [155395, 1046374] processed_samples 700 unjoint_samples 700 joint_samples 1 [155395, 1046374] processed_samples 700 unjoint_samples 700 joint_samples 1 [296281, 1047472] processed_samples 700 unjoint_samples 700 joint_samples 1 [296281, 1047472] processed_samples 700 unjoint_samples 700 joint_samples 1 [373621, 1047617] processed_samples 700 unjoint_samples 700 joint_samples 1 [373621, 1047617] processed_samples 700 unjoint_samples 700 joint_samples 1 [302482, 1045187] processed_samples 700 unjoint_samples 700 joint_samples 1 [302482, 1045187] processed_samples 700 unjoint_samples 700 joint_samples 0 [1028777, 1037711] processed_samples 700 unjoint_samples 700 joint_samples 0 [1028777, 1037711] processed_samples 700 unjoint_samples 700 joint_samples 0 [1043242, 1041055] [h264 @ 0x5617e9274140] mmco: unref short failure processed_samples 700 unjoint_samples 700 joint_samples 0 [1043242, 1041055] [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure processed_samples 800 unjoint_samples 800 joint_samples 1 [408688, 1046374] processed_samples 800 unjoint_samples 800 joint_samples 1 [408688, 1046374] processed_samples 800 unjoint_samples 800 joint_samples 1 [285807, 1047649] processed_samples 800 unjoint_samples 800 joint_samples 1 [285807, 1047649] processed_samples 800 unjoint_samples 800 joint_samples 1 [1046313, 390935] processed_samples 800 unjoint_samples 800 joint_samples 1 [1046313, 390935] [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure processed_samples 800 unjoint_samples 800 joint_samples 1 [597137, 1047472] processed_samples 800 unjoint_samples 800 joint_samples 1 [597137, 1047472] processed_samples 800 unjoint_samples 800 joint_samples 1 [464041, 1046583] processed_samples 800 unjoint_samples 800 joint_samples 1 [464041, 1046583] processed_samples 800 unjoint_samples 800 joint_samples 1 [1047221, 623708] processed_samples 800 unjoint_samples 800 joint_samples 1 [1047221, 623708] processed_samples 800 unjoint_samples 800 joint_samples 1 [632307, 1047617] processed_samples 800 unjoint_samples 800 joint_samples 1 [632307, 1047617] processed_samples 800 unjoint_samples 800 joint_samples 1 [676248, 1045187] processed_samples 800 unjoint_samples 800 joint_samples 1 [676248, 1045187] [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure processed_samples 900 unjoint_samples 900 joint_samples 1 [1047221, 984881] processed_samples 900 unjoint_samples 900 joint_samples 1 [1047221, 984881] processed_samples 900 unjoint_samples 900 joint_samples 1 [624367, 1047649] processed_samples 900 unjoint_samples 900 joint_samples 1 [624367, 1047649] processed_samples 900 unjoint_samples 900 joint_samples 1 [766235, 1046583] processed_samples 900 unjoint_samples 900 joint_samples 1 [766235, 1046583] [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure processed_samples 900 unjoint_samples 900 joint_samples 1 [1046313, 713292] processed_samples 900 unjoint_samples 900 joint_samples 1 [1046313, 713292] processed_samples 900 unjoint_samples 900 joint_samples 1 [968545, 1047472] [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure processed_samples 900 unjoint_samples 900 joint_samples 1 [946444, 1045187] [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure processed_samples 900 unjoint_samples 900 joint_samples 1 [968545, 1047472] processed_samples 900 unjoint_samples 900 joint_samples 1 [776588, 1046374] processed_samples 900 unjoint_samples 900 joint_samples 1 [946444, 1045187] processed_samples 900 unjoint_samples 900 joint_samples 1 [776588, 1046374] processed_samples 900 unjoint_samples 900 joint_samples 1 [905577, 1047617] processed_samples 900 unjoint_samples 900 joint_samples 1 [905577, 1047617] [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x559616d35c80] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure processed_samples 1000 unjoint_samples 1000 joint_samples 2 [1037015, 189936] processed_samples 1000 unjoint_samples 1000 joint_samples 2 [1037015, 189936] processed_samples 1000 unjoint_samples 1000 joint_samples 2 [1042902, 63834] processed_samples 1000 unjoint_samples 1000 joint_samples 2 [1046476, 65937] processed_samples 1000 unjoint_samples 1000 joint_samples 2 [1042902, 63834] processed_samples 1000 unjoint_samples 1000 joint_samples 2 [1046476, 65937] processed_samples 1000 unjoint_samples 1000 joint_samples 2 [228076, 1034755] processed_samples 1000 unjoint_samples 1000 joint_samples 2 [228076, 1034755] processed_samples 1000 unjoint_samples 1000 joint_samples 2 [1034117, 231307] processed_samples 1000 unjoint_samples 1000 joint_samples 2 [1034117, 231307] processed_samples 1000 unjoint_samples 1000 joint_samples 2 [1047011, 178029] processed_samples 1000 unjoint_samples 1000 joint_samples 2 [1047011, 178029] processed_samples 1000 unjoint_samples 1000 joint_samples 1 [892167, 1047649] processed_samples 1000 unjoint_samples 1000 joint_samples 1 [892167, 1047649] processed_samples 1000 unjoint_samples 1000 joint_samples 1 [1046313, 1036581] processed_samples 1000 unjoint_samples 1000 joint_samples 1 [1046313, 1036581] [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e9531f40] [h264 @ 0x5596196a8b80] mmco: unref short failure mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure processed_samples 1100 unjoint_samples 1100 joint_samples 2 [1042902, 396719] processed_samples 1100 unjoint_samples 1100 joint_samples 2 [1042902, 396719] processed_samples 1100 unjoint_samples 1100 joint_samples 2 [1046476, 427953] processed_samples 1100 unjoint_samples 1100 joint_samples 2 [1046476, 427953] processed_samples 1100 unjoint_samples 1100 joint_samples 2 [996107, 341479] processed_samples 1100 unjoint_samples 1100 joint_samples 2 [996107, 341479] processed_samples 1100 unjoint_samples 1100 joint_samples 2 [1037015, 447603] processed_samples 1100 unjoint_samples 1100 joint_samples 2 [1037015, 447603] processed_samples 1100 unjoint_samples 1100 joint_samples 2 [592652, 1034755] processed_samples 1100 unjoint_samples 1100 joint_samples 2 [592652, 1034755] processed_samples 1100 unjoint_samples 1100 joint_samples 2 [1047011, 419894] processed_samples 1100 unjoint_samples 1100 joint_samples 2 [1034117, 509986] processed_samples 1100 unjoint_samples 1100 joint_samples 2 [311617, 1046462] processed_samples 1100 unjoint_samples 1100 joint_samples 2 [311617, 1046462] processed_samples 1100 unjoint_samples 1100 joint_samples 2 [1034117, 509986] [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure processed_samples 1100 unjoint_samples 1100 joint_samples 2 [1047011, 419894] [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559619420380] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x559619420380] mmco: unref short failure [h264 @ 0x559619420380] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x559618a89040] mmco: unref short failure [h264 @ 0x559618a89040] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure processed_samples 1200 unjoint_samples 1200 joint_samples 2 [1046476, 728508] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [1046476, 728508] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [996107, 727543] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [996107, 727543] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [1042902, 817009] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [1042902, 817009] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [815906, 1034755] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [815906, 1034755] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [1037015, 798549] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [1047011, 656722] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [1047011, 656722] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [1034117, 836947] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [1037015, 798549] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [657486, 1046462] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [1034117, 836947] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [657486, 1046462] [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e7e34fc0] mmco: unref short failure [h264 @ 0x5617e7e34fc0] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure processed_samples 1300 unjoint_samples 1300 joint_samples 3 [122448, 1047456] processed_samples 1300 unjoint_samples 1300 joint_samples 3 [122448, 1047456] processed_samples 1300 unjoint_samples 1300 joint_samples 2 [1001271, 1018164] processed_samples 1300 unjoint_samples 1300 joint_samples 2 [1001271, 1018164] processed_samples 1300 unjoint_samples 1300 joint_samples 3 [375099, 1030228] processed_samples 1300 unjoint_samples 1300 joint_samples 3 [375099, 1030228] processed_samples 1300 unjoint_samples 1300 joint_samples 3 [1040374, 144080] processed_samples 1300 unjoint_samples 1300 joint_samples 3 [1040374, 144080] processed_samples 1300 unjoint_samples 1300 joint_samples 2 [998275, 1046462] processed_samples 1300 unjoint_samples 1300 joint_samples 2 [998275, 1046462] processed_samples 1300 unjoint_samples 1300 joint_samples 2 [1046476, 1012906] processed_samples 1300 unjoint_samples 1300 joint_samples 2 [1046476, 1012906] processed_samples 1300 unjoint_samples 1300 joint_samples 3 [1041955, 8266] processed_samples 1300 unjoint_samples 1300 joint_samples 3 [1041955, 8266] processed_samples 1300 unjoint_samples 1300 joint_samples 2 [1047011, 1018252] processed_samples 1300 unjoint_samples 1300 joint_samples 2 [1047011, 1018252] [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ebc56240] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure processed_samples 1400 unjoint_samples 1400 joint_samples 3 [1044455, 234509] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [1044455, 234509] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [1046476, 200925] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [1046476, 200925] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [1041955, 265705] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [1040374, 556727] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [1041955, 265705] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [1040374, 556727] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [339616, 1027605] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [339616, 1027605] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [1002810, 416362] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [1002810, 416362] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [630033, 1030228] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [630033, 1030228] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [502478, 1047456] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [502478, 1047456] [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559616d35c80] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure processed_samples 1500 unjoint_samples 1500 joint_samples 3 [1044455, 662650] processed_samples 1500 unjoint_samples 1500 joint_samples 3 [1044455, 662650] processed_samples 1500 unjoint_samples 1500 joint_samples 3 [1046476, 572076] processed_samples 1500 unjoint_samples 1500 joint_samples 3 [1046476, 572076] processed_samples 1500 unjoint_samples 1500 joint_samples 3 [1040374, 852242] processed_samples 1500 unjoint_samples 1500 joint_samples 3 [1040374, 852242] processed_samples 1500 unjoint_samples 1500 joint_samples 3 [731588, 1027605] processed_samples 1500 unjoint_samples 1500 joint_samples 3 [801639, 1047456] processed_samples 1500 unjoint_samples 1500 joint_samples 3 [731588, 1027605] processed_samples 1500 unjoint_samples 1500 joint_samples 3 [801639, 1047456] processed_samples 1500 unjoint_samples 1500 joint_samples 3 [1002810, 708686] processed_samples 1500 unjoint_samples 1500 joint_samples 3 [1002810, 708686] processed_samples 1500 unjoint_samples 1500 joint_samples 3 [910760, 1030228] processed_samples 1500 unjoint_samples 1500 joint_samples 3 [910760, 1030228] processed_samples 1500 unjoint_samples 1500 joint_samples 3 [1041955, 553644] processed_samples 1500 unjoint_samples 1500 joint_samples 3 [1041955, 553644] [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559616fdd040] mmco: unref short failure [h264 @ 0x559616fdd040] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x559619420380] mmco: unref short failure [h264 @ 0x559619420380] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure processed_samples 1600 unjoint_samples 1600 joint_samples 4 [1046343, 118862] processed_samples 1600 unjoint_samples 1600 joint_samples 4 [1046343, 118862] processed_samples 1600 unjoint_samples 1600 joint_samples 4 [1032497, 204427] processed_samples 1600 unjoint_samples 1600 joint_samples 4 [18396, 1046778] processed_samples 1600 unjoint_samples 1600 joint_samples 4 [1032497, 204427] processed_samples 1600 unjoint_samples 1600 joint_samples 4 [18396, 1046778] processed_samples 1600 unjoint_samples 1600 joint_samples 3 [1002810, 987724] processed_samples 1600 unjoint_samples 1600 joint_samples 3 [1044455, 977139] processed_samples 1600 unjoint_samples 1600 joint_samples 3 [1002810, 987724] processed_samples 1600 unjoint_samples 1600 joint_samples 3 [1044455, 977139] processed_samples 1600 unjoint_samples 1600 joint_samples 3 [1041955, 843448] processed_samples 1600 unjoint_samples 1600 joint_samples 3 [1041955, 843448] processed_samples 1600 unjoint_samples 1600 joint_samples 3 [1046179, 1046260] processed_samples 1600 unjoint_samples 1600 joint_samples 3 [1046476, 810142] processed_samples 1600 unjoint_samples 1600 joint_samples 3 [1046476, 810142] processed_samples 1600 unjoint_samples 1600 joint_samples 3 [1046179, 1046260] [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x55961840f500] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure processed_samples 1700 unjoint_samples 1700 joint_samples 4 [43086, 1038963] processed_samples 1700 unjoint_samples 1700 joint_samples 4 [43086, 1038963] processed_samples 1700 unjoint_samples 1700 joint_samples 4 [229909, 1045767] processed_samples 1700 unjoint_samples 1700 joint_samples 4 [229909, 1045767] processed_samples 1700 unjoint_samples 1700 joint_samples 4 [351869, 1046260] processed_samples 1700 unjoint_samples 1700 joint_samples 4 [351869, 1046260] processed_samples 1700 unjoint_samples 1700 joint_samples 4 [1047258, 79130] processed_samples 1700 unjoint_samples 1700 joint_samples 4 [213625, 1046983] processed_samples 1700 unjoint_samples 1700 joint_samples 4 [1047258, 79130] processed_samples 1700 unjoint_samples 1700 joint_samples 4 [213625, 1046983] processed_samples 1700 unjoint_samples 1700 joint_samples 4 [1032497, 537910] processed_samples 1700 unjoint_samples 1700 joint_samples 4 [1032497, 537910] processed_samples 1700 unjoint_samples 1700 joint_samples 4 [366437, 1046778] processed_samples 1700 unjoint_samples 1700 joint_samples 4 [366437, 1046778] [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure processed_samples 1700 unjoint_samples 1700 joint_samples 4 [1046343, 445653] processed_samples 1700 unjoint_samples 1700 joint_samples 4 [1046343, 445653] [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x559616fdd040] mmco: unref short failure [h264 @ 0x559616fdd040] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e803bd40] mmco: unref short failure [h264 @ 0x5617e803bd40] mmco: unref short failure [h264 @ 0x5617e803bd40] mmco: unref short failure [h264 @ 0x559616fdd040] mmco: unref short failure [h264 @ 0x559616fdd040] mmco: unref short failure [h264 @ 0x559616fdd040] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617ea25da40] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure processed_samples 1800 unjoint_samples 1800 joint_samples 4 [1046343, 775225] processed_samples 1800 unjoint_samples 1800 joint_samples 4 [827571, 1046260] processed_samples 1800 unjoint_samples 1800 joint_samples 4 [827571, 1046260] processed_samples 1800 unjoint_samples 1800 joint_samples 4 [1046343, 775225] processed_samples 1800 unjoint_samples 1800 joint_samples 4 [363779, 1038963] processed_samples 1800 unjoint_samples 1800 joint_samples 4 [363779, 1038963] processed_samples 1800 unjoint_samples 1800 joint_samples 4 [632524, 1046778] processed_samples 1800 unjoint_samples 1800 joint_samples 4 [632524, 1046778] [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure processed_samples 1800 unjoint_samples 1800 joint_samples 4 [1047258, 403291] processed_samples 1800 unjoint_samples 1800 joint_samples 4 [1047258, 403291] processed_samples 1800 unjoint_samples 1800 joint_samples 4 [501727, 1045767] processed_samples 1800 unjoint_samples 1800 joint_samples 4 [501727, 1045767] processed_samples 1800 unjoint_samples 1800 joint_samples 4 [1032497, 830406] processed_samples 1800 unjoint_samples 1800 joint_samples 4 [1032497, 830406] processed_samples 1800 unjoint_samples 1800 joint_samples 4 [582639, 1046983] processed_samples 1800 unjoint_samples 1800 joint_samples 4 [582639, 1046983] [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e8e0d640] mmco: unref short failure [h264 @ 0x5617e8e0d640] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure processed_samples 1900 unjoint_samples 1900 joint_samples 5 [1039106, 124220] [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure processed_samples 1900 unjoint_samples 1900 joint_samples 5 [1046343, 149018] processed_samples 1900 unjoint_samples 1900 joint_samples 5 [1039106, 124220] processed_samples 1900 unjoint_samples 1900 joint_samples 5 [1046343, 149018] processed_samples 1900 unjoint_samples 1900 joint_samples 4 [609667, 1038963] processed_samples 1900 unjoint_samples 1900 joint_samples 4 [609667, 1038963] processed_samples 1900 unjoint_samples 1900 joint_samples 4 [1047258, 674481] processed_samples 1900 unjoint_samples 1900 joint_samples 4 [852354, 1046983] processed_samples 1900 unjoint_samples 1900 joint_samples 4 [1047258, 674481] processed_samples 1900 unjoint_samples 1900 joint_samples 4 [852354, 1046983] processed_samples 1900 unjoint_samples 1900 joint_samples 5 [66792, 1038508] processed_samples 1900 unjoint_samples 1900 joint_samples 4 [867657, 1045767] processed_samples 1900 unjoint_samples 1900 joint_samples 4 [867657, 1045767] processed_samples 1900 unjoint_samples 1900 joint_samples 5 [66792, 1038508] processed_samples 1900 unjoint_samples 1900 joint_samples 4 [907367, 1046778] processed_samples 1900 unjoint_samples 1900 joint_samples 4 [907367, 1046778] [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617e7e34fc0] mmco: unref short failure [h264 @ 0x5617e7e34fc0] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure processed_samples 2000 unjoint_samples 2000 joint_samples 5 [1046343, 489400] processed_samples 2000 unjoint_samples 2000 joint_samples 5 [1046343, 489400] processed_samples 2000 unjoint_samples 2000 joint_samples 5 [96503, 1046983] processed_samples 2000 unjoint_samples 2000 joint_samples 5 [39647, 1045425] processed_samples 2000 unjoint_samples 2000 joint_samples 5 [96503, 1046983] processed_samples 2000 unjoint_samples 2000 joint_samples 5 [39647, 1045425] processed_samples 2000 unjoint_samples 2000 joint_samples 5 [1048048, 191153] processed_samples 2000 unjoint_samples 2000 joint_samples 5 [1039106, 438263] processed_samples 2000 unjoint_samples 2000 joint_samples 5 [1039106, 438263] processed_samples 2000 unjoint_samples 2000 joint_samples 5 [1048048, 191153] processed_samples 2000 unjoint_samples 2000 joint_samples 5 [1045356, 128024] processed_samples 2000 unjoint_samples 2000 joint_samples 5 [1045356, 128024] processed_samples 2000 unjoint_samples 2000 joint_samples 5 [439135, 1038508] processed_samples 2000 unjoint_samples 2000 joint_samples 5 [439135, 1038508] processed_samples 2000 unjoint_samples 2000 joint_samples 4 [814629, 1038963] processed_samples 2000 unjoint_samples 2000 joint_samples 4 [814629, 1038963] [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure processed_samples 2100 unjoint_samples 2100 joint_samples 5 [1039106, 799275] processed_samples 2100 unjoint_samples 2100 joint_samples 5 [410596, 1045425] processed_samples 2100 unjoint_samples 2100 joint_samples 5 [1039106, 799275] processed_samples 2100 unjoint_samples 2100 joint_samples 5 [410596, 1045425] processed_samples 2100 unjoint_samples 2100 joint_samples 5 [1046686, 82475] processed_samples 2100 unjoint_samples 2100 joint_samples 5 [1046686, 82475] processed_samples 2100 unjoint_samples 2100 joint_samples 5 [449752, 1046983] processed_samples 2100 unjoint_samples 2100 joint_samples 5 [1045356, 475935] processed_samples 2100 unjoint_samples 2100 joint_samples 5 [449752, 1046983] processed_samples 2100 unjoint_samples 2100 joint_samples 5 [1045356, 475935] processed_samples 2100 unjoint_samples 2100 joint_samples 5 [1048048, 651728] processed_samples 2100 unjoint_samples 2100 joint_samples 5 [1048048, 651728] processed_samples 2100 unjoint_samples 2100 joint_samples 5 [776225, 1038508] processed_samples 2100 unjoint_samples 2100 joint_samples 5 [776225, 1038508] processed_samples 2100 unjoint_samples 2100 joint_samples 5 [1046343, 755308] processed_samples 2100 unjoint_samples 2100 joint_samples 5 [1046343, 755308] [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617e803bd40] mmco: unref short failure [h264 @ 0x5617e803bd40] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure processed_samples 2200 unjoint_samples 2200 joint_samples 5 [1048048, 884891] processed_samples 2200 unjoint_samples 2200 joint_samples 5 [1048048, 884891] processed_samples 2200 unjoint_samples 2200 joint_samples 6 [1042244, 222486] processed_samples 2200 unjoint_samples 2200 joint_samples 6 [1042244, 222486] processed_samples 2200 unjoint_samples 2200 joint_samples 6 [88290, 1006252] processed_samples 2200 unjoint_samples 2200 joint_samples 6 [88290, 1006252] processed_samples 2200 unjoint_samples 2200 joint_samples 5 [1046686, 356641] processed_samples 2200 unjoint_samples 2200 joint_samples 5 [1046686, 356641] processed_samples 2200 unjoint_samples 2200 joint_samples 5 [810650, 1045425] processed_samples 2200 unjoint_samples 2200 joint_samples 5 [810650, 1045425] processed_samples 2200 unjoint_samples 2200 joint_samples 5 [804654, 1046983] processed_samples 2200 unjoint_samples 2200 joint_samples 5 [804654, 1046983] processed_samples 2200 unjoint_samples 2200 joint_samples 5 [1044019, 1044924] processed_samples 2200 unjoint_samples 2200 joint_samples 5 [1044019, 1044924] processed_samples 2200 unjoint_samples 2200 joint_samples 5 [1045356, 716722] processed_samples 2200 unjoint_samples 2200 joint_samples 5 [1045356, 716722] [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559617e83980] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559617afc780] mmco: unref short failure [h264 @ 0x559617afc780] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure processed_samples 2300 unjoint_samples 2300 joint_samples 6 [1045356, 16654] processed_samples 2300 unjoint_samples 2300 joint_samples 6 [1045356, 16654] processed_samples 2300 unjoint_samples 2300 joint_samples 6 [29370, 1047600] processed_samples 2300 unjoint_samples 2300 joint_samples 6 [29370, 1047600] processed_samples 2300 unjoint_samples 2300 joint_samples 6 [358572, 1006252] processed_samples 2300 unjoint_samples 2300 joint_samples 6 [358572, 1006252] processed_samples 2300 unjoint_samples 2300 joint_samples 6 [362057, 982152] processed_samples 2300 unjoint_samples 2300 joint_samples 6 [362057, 982152] processed_samples 2300 unjoint_samples 2300 joint_samples 6 [1011991, 214606] processed_samples 2300 unjoint_samples 2300 joint_samples 6 [1011991, 214606] processed_samples 2300 unjoint_samples 2300 joint_samples 5 [1046686, 686059] processed_samples 2300 unjoint_samples 2300 joint_samples 5 [1046686, 686059] processed_samples 2300 unjoint_samples 2300 joint_samples 6 [206409, 1048071] processed_samples 2300 unjoint_samples 2300 joint_samples 6 [206409, 1048071] processed_samples 2300 unjoint_samples 2300 joint_samples 6 [1042244, 631889] processed_samples 2300 unjoint_samples 2300 joint_samples 6 [1042244, 631889] [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559617e83980] mmco: unref short failure [h264 @ 0x559617e83980] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596177086c0] mmco: unref short failure [h264 @ 0x5596177086c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596177086c0] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure processed_samples 2400 unjoint_samples 2400 joint_samples 6 [661212, 1006252] processed_samples 2400 unjoint_samples 2400 joint_samples 6 [661212, 1006252] processed_samples 2400 unjoint_samples 2400 joint_samples 6 [537827, 1048071] processed_samples 2400 unjoint_samples 2400 joint_samples 6 [537827, 1048071] processed_samples 2400 unjoint_samples 2400 joint_samples 6 [311696, 1047600] processed_samples 2400 unjoint_samples 2400 joint_samples 6 [311696, 1047600] [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5596196a8b80] [h264 @ 0x5617e8001bc0] mmco: unref short failure mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure processed_samples 2400 unjoint_samples 2400 joint_samples 6 [1045356, 431802] processed_samples 2400 unjoint_samples 2400 joint_samples 6 [1045356, 431802] processed_samples 2400 unjoint_samples 2400 joint_samples 6 [1042244, 928396] processed_samples 2400 unjoint_samples 2400 joint_samples 6 [1042244, 928396] processed_samples 2400 unjoint_samples 2400 joint_samples 6 [581936, 982152] processed_samples 2400 unjoint_samples 2400 joint_samples 6 [581936, 982152] [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure processed_samples 2400 unjoint_samples 2400 joint_samples 6 [1011991, 585883] processed_samples 2400 unjoint_samples 2400 joint_samples 6 [1011991, 585883] processed_samples 2400 unjoint_samples 2400 joint_samples 5 [1046686, 1007570] processed_samples 2400 unjoint_samples 2400 joint_samples 5 [1046686, 1007570] [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x559619420380] mmco: unref short failure [h264 @ 0x559619420380] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure processed_samples 2500 unjoint_samples 2500 joint_samples 7 [104676, 1041889] processed_samples 2500 unjoint_samples 2500 joint_samples 6 [348928, 1031263] processed_samples 2500 unjoint_samples 2500 joint_samples 7 [104676, 1041889] processed_samples 2500 unjoint_samples 2500 joint_samples 6 [348928, 1031263] processed_samples 2500 unjoint_samples 2500 joint_samples 6 [1011991, 882997] processed_samples 2500 unjoint_samples 2500 joint_samples 6 [1045356, 942067] processed_samples 2500 unjoint_samples 2500 joint_samples 6 [1011991, 882997] processed_samples 2500 unjoint_samples 2500 joint_samples 6 [1045356, 942067] processed_samples 2500 unjoint_samples 2500 joint_samples 6 [892280, 982152] processed_samples 2500 unjoint_samples 2500 joint_samples 6 [892280, 982152] processed_samples 2500 unjoint_samples 2500 joint_samples 6 [552828, 1047600] processed_samples 2500 unjoint_samples 2500 joint_samples 6 [843819, 1048071] processed_samples 2500 unjoint_samples 2500 joint_samples 6 [552828, 1047600] processed_samples 2500 unjoint_samples 2500 joint_samples 6 [843819, 1048071] [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure processed_samples 2500 unjoint_samples 2500 joint_samples 6 [894170, 1006252] [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure processed_samples 2500 unjoint_samples 2500 joint_samples 6 [894170, 1006252] [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596172068c0] [h264 @ 0x5617e7c5b300] mmco: unref short failure mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure processed_samples 2600 unjoint_samples 2600 joint_samples 7 [1044977, 46524] processed_samples 2600 unjoint_samples 2600 joint_samples 7 [1044977, 46524] processed_samples 2600 unjoint_samples 2600 joint_samples 7 [1011991, 233387] processed_samples 2600 unjoint_samples 2600 joint_samples 7 [1011991, 233387] processed_samples 2600 unjoint_samples 2600 joint_samples 7 [69064, 1045401] processed_samples 2600 unjoint_samples 2600 joint_samples 7 [69064, 1045401] processed_samples 2600 unjoint_samples 2600 joint_samples 7 [387378, 1041889] processed_samples 2600 unjoint_samples 2600 joint_samples 7 [387378, 1041889] processed_samples 2600 unjoint_samples 2600 joint_samples 6 [585507, 1031263] processed_samples 2600 unjoint_samples 2600 joint_samples 6 [585507, 1031263] processed_samples 2600 unjoint_samples 2600 joint_samples 7 [176320, 1035832] processed_samples 2600 unjoint_samples 2600 joint_samples 7 [176320, 1035832] processed_samples 2600 unjoint_samples 2600 joint_samples 7 [1045289, 74354] processed_samples 2600 unjoint_samples 2600 joint_samples 7 [1045289, 74354] processed_samples 2600 unjoint_samples 2600 joint_samples 6 [829989, 1047600] processed_samples 2600 unjoint_samples 2600 joint_samples 6 [829989, 1047600] [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure processed_samples 2700 unjoint_samples 2700 joint_samples 7 [372393, 1045401] processed_samples 2700 unjoint_samples 2700 joint_samples 7 [372393, 1045401] processed_samples 2700 unjoint_samples 2700 joint_samples 7 [1044977, 375974] processed_samples 2700 unjoint_samples 2700 joint_samples 7 [1044977, 375974] processed_samples 2700 unjoint_samples 2700 joint_samples 7 [1047983, 163782] processed_samples 2700 unjoint_samples 2700 joint_samples 7 [1047983, 163782] processed_samples 2700 unjoint_samples 2700 joint_samples 7 [1045289, 415502] processed_samples 2700 unjoint_samples 2700 joint_samples 7 [1045289, 415502] processed_samples 2700 unjoint_samples 2700 joint_samples 7 [1011991, 549434] processed_samples 2700 unjoint_samples 2700 joint_samples 7 [1011991, 549434] processed_samples 2700 unjoint_samples 2700 joint_samples 7 [427015, 1035832] processed_samples 2700 unjoint_samples 2700 joint_samples 7 [427015, 1035832] processed_samples 2700 unjoint_samples 2700 joint_samples 7 [688435, 1041889] processed_samples 2700 unjoint_samples 2700 joint_samples 7 [688435, 1041889] [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure processed_samples 2700 unjoint_samples 2700 joint_samples 6 [1026309, 1031263] processed_samples 2700 unjoint_samples 2700 joint_samples 6 [1026309, 1031263] [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617e8964000] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617e8964000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure processed_samples 2800 unjoint_samples 2800 joint_samples 8 [959794, 171251] processed_samples 2800 unjoint_samples 2800 joint_samples 8 [959794, 171251] processed_samples 2800 unjoint_samples 2800 joint_samples 7 [317690, 1044663] processed_samples 2800 unjoint_samples 2800 joint_samples 7 [317690, 1044663] processed_samples 2800 unjoint_samples 2800 joint_samples 7 [1044977, 580126] processed_samples 2800 unjoint_samples 2800 joint_samples 7 [1044977, 580126] processed_samples 2800 unjoint_samples 2800 joint_samples 7 [687635, 1045401] processed_samples 2800 unjoint_samples 2800 joint_samples 7 [687635, 1045401] [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e8964000] mmco: unref short failure [h264 @ 0x5617e8964000] mmco: unref short failure processed_samples 2800 unjoint_samples 2800 joint_samples 7 [1047983, 487808] processed_samples 2800 unjoint_samples 2800 joint_samples 7 [1045289, 634964] processed_samples 2800 unjoint_samples 2800 joint_samples 7 [1047983, 487808] processed_samples 2800 unjoint_samples 2800 joint_samples 7 [1045289, 634964] processed_samples 2800 unjoint_samples 2800 joint_samples 7 [1011991, 890406] processed_samples 2800 unjoint_samples 2800 joint_samples 7 [1011991, 890406] processed_samples 2800 unjoint_samples 2800 joint_samples 7 [861990, 1035832] processed_samples 2800 unjoint_samples 2800 joint_samples 7 [861990, 1035832] [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617e8964000] mmco: unref short failure [h264 @ 0x5617e8964000] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure processed_samples 2900 unjoint_samples 2900 joint_samples 8 [178333, 1046932] processed_samples 2900 unjoint_samples 2900 joint_samples 8 [178333, 1046932] processed_samples 2900 unjoint_samples 2900 joint_samples 8 [1047689, 67281] processed_samples 2900 unjoint_samples 2900 joint_samples 8 [1047689, 67281] processed_samples 2900 unjoint_samples 2900 joint_samples 7 [949445, 1045401] processed_samples 2900 unjoint_samples 2900 joint_samples 8 [184167, 1039689] processed_samples 2900 unjoint_samples 2900 joint_samples 8 [184167, 1039689] processed_samples 2900 unjoint_samples 2900 joint_samples 7 [642305, 1044663] processed_samples 2900 unjoint_samples 2900 joint_samples 7 [642305, 1044663] processed_samples 2900 unjoint_samples 2900 joint_samples 7 [949445, 1045401] processed_samples 2900 unjoint_samples 2900 joint_samples 7 [1047983, 785557] processed_samples 2900 unjoint_samples 2900 joint_samples 7 [1047983, 785557] processed_samples 2900 unjoint_samples 2900 joint_samples 8 [959794, 587501] processed_samples 2900 unjoint_samples 2900 joint_samples 8 [959794, 587501] processed_samples 2900 unjoint_samples 2900 joint_samples 7 [1045289, 908525] processed_samples 2900 unjoint_samples 2900 joint_samples 7 [1045289, 908525] [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ea25da40] mmco: unref short failure [h264 @ 0x5617ea25da40] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596196a8b80] [h264 @ 0x5617ec342580] mmco: unref short failure mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure processed_samples 3000 unjoint_samples 3000 joint_samples 8 [1047689, 427662] processed_samples 3000 unjoint_samples 3000 joint_samples 8 [1047689, 427662] processed_samples 3000 unjoint_samples 3000 joint_samples 8 [183240, 1046814] processed_samples 3000 unjoint_samples 3000 joint_samples 8 [455338, 1046932] processed_samples 3000 unjoint_samples 3000 joint_samples 8 [183240, 1046814] processed_samples 3000 unjoint_samples 3000 joint_samples 8 [455338, 1046932] processed_samples 3000 unjoint_samples 3000 joint_samples 8 [56738, 1047832] processed_samples 3000 unjoint_samples 3000 joint_samples 8 [959794, 892488] processed_samples 3000 unjoint_samples 3000 joint_samples 8 [56738, 1047832] processed_samples 3000 unjoint_samples 3000 joint_samples 8 [1047544, 178136] processed_samples 3000 unjoint_samples 3000 joint_samples 8 [1047544, 178136] processed_samples 3000 unjoint_samples 3000 joint_samples 8 [959794, 892488] processed_samples 3000 unjoint_samples 3000 joint_samples 7 [1000377, 1044663] processed_samples 3000 unjoint_samples 3000 joint_samples 7 [1000377, 1044663] processed_samples 3000 unjoint_samples 3000 joint_samples 8 [464923, 1039689] processed_samples 3000 unjoint_samples 3000 joint_samples 8 [464923, 1039689] [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure processed_samples 3100 unjoint_samples 3100 joint_samples 9 [993599, 216904] [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure processed_samples 3100 unjoint_samples 3100 joint_samples 9 [993599, 216904] [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure processed_samples 3100 unjoint_samples 3100 joint_samples 8 [214665, 1047012] processed_samples 3100 unjoint_samples 3100 joint_samples 8 [214665, 1047012] processed_samples 3100 unjoint_samples 3100 joint_samples 8 [1047689, 706620] processed_samples 3100 unjoint_samples 3100 joint_samples 8 [1047689, 706620] processed_samples 3100 unjoint_samples 3100 joint_samples 8 [490439, 1046814] processed_samples 3100 unjoint_samples 3100 joint_samples 8 [490439, 1046814] processed_samples 3100 unjoint_samples 3100 joint_samples 8 [1047544, 472092] processed_samples 3100 unjoint_samples 3100 joint_samples 8 [1047544, 472092] processed_samples 3100 unjoint_samples 3100 joint_samples 8 [762372, 1039689] processed_samples 3100 unjoint_samples 3100 joint_samples 8 [762372, 1039689] processed_samples 3100 unjoint_samples 3100 joint_samples 8 [331712, 1047832] processed_samples 3100 unjoint_samples 3100 joint_samples 8 [331712, 1047832] processed_samples 3100 unjoint_samples 3100 joint_samples 8 [738808, 1046932] processed_samples 3100 unjoint_samples 3100 joint_samples 8 [738808, 1046932] [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] [h264 @ 0x5617ebbed500] mmco: unref short failure mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x559618393880] mmco: unref short failure [h264 @ 0x559618393880] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure processed_samples 3200 unjoint_samples 3200 joint_samples 9 [13282, 1045187] processed_samples 3200 unjoint_samples 3200 joint_samples 9 [13282, 1045187] processed_samples 3200 unjoint_samples 3200 joint_samples 8 [1047544, 754965] processed_samples 3200 unjoint_samples 3200 joint_samples 8 [1047544, 754965] processed_samples 3200 unjoint_samples 3200 joint_samples 9 [1038321, 133087] processed_samples 3200 unjoint_samples 3200 joint_samples 9 [1038321, 133087] processed_samples 3200 unjoint_samples 3200 joint_samples 9 [993599, 508605] processed_samples 3200 unjoint_samples 3200 joint_samples 9 [993599, 508605] processed_samples 3200 unjoint_samples 3200 joint_samples 8 [446661, 1047012] processed_samples 3200 unjoint_samples 3200 joint_samples 8 [446661, 1047012] processed_samples 3200 unjoint_samples 3200 joint_samples 8 [889236, 1046814] processed_samples 3200 unjoint_samples 3200 joint_samples 8 [889236, 1046814] processed_samples 3200 unjoint_samples 3200 joint_samples 8 [1021765, 1039689] processed_samples 3200 unjoint_samples 3200 joint_samples 8 [1021765, 1039689] processed_samples 3200 unjoint_samples 3200 joint_samples 8 [676012, 1047832] processed_samples 3200 unjoint_samples 3200 joint_samples 8 [676012, 1047832] [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ec56a6c0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure processed_samples 3300 unjoint_samples 3300 joint_samples 8 [800528, 1047012] processed_samples 3300 unjoint_samples 3300 joint_samples 9 [1034687, 183997] processed_samples 3300 unjoint_samples 3300 joint_samples 9 [1038321, 457553] processed_samples 3300 unjoint_samples 3300 joint_samples 9 [1045983, 268230] processed_samples 3300 unjoint_samples 3300 joint_samples 8 [800528, 1047012] processed_samples 3300 unjoint_samples 3300 joint_samples 9 [281258, 1045187] processed_samples 3300 unjoint_samples 3300 joint_samples 9 [1038321, 457553] processed_samples 3300 unjoint_samples 3300 joint_samples 9 [1034687, 183997] processed_samples 3300 unjoint_samples 3300 joint_samples 8 [958958, 1047832] processed_samples 3300 unjoint_samples 3300 joint_samples 9 [1045983, 268230] processed_samples 3300 unjoint_samples 3300 joint_samples 9 [993599, 858491] processed_samples 3300 unjoint_samples 3300 joint_samples 9 [281258, 1045187] [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure processed_samples 3300 unjoint_samples 3300 joint_samples 8 [958958, 1047832] processed_samples 3300 unjoint_samples 3300 joint_samples 9 [993599, 858491] [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure processed_samples 3300 unjoint_samples 3300 joint_samples 8 [1047544, 1022951] processed_samples 3300 unjoint_samples 3300 joint_samples 8 [1047544, 1022951] [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure processed_samples 3400 unjoint_samples 3400 joint_samples 10 [1043359, 51491] processed_samples 3400 unjoint_samples 3400 joint_samples 10 [1043359, 51491] processed_samples 3400 unjoint_samples 3400 joint_samples 9 [1045983, 555276] processed_samples 3400 unjoint_samples 3400 joint_samples 9 [1045983, 555276] processed_samples 3400 unjoint_samples 3400 joint_samples 9 [284131, 1041969] processed_samples 3400 unjoint_samples 3400 joint_samples 9 [284131, 1041969] processed_samples 3400 unjoint_samples 3400 joint_samples 9 [25681, 1047012] processed_samples 3400 unjoint_samples 3400 joint_samples 9 [25681, 1047012] processed_samples 3400 unjoint_samples 3400 joint_samples 9 [355911, 1047832] processed_samples 3400 unjoint_samples 3400 joint_samples 9 [355911, 1047832] processed_samples 3400 unjoint_samples 3400 joint_samples 9 [566300, 1045187] processed_samples 3400 unjoint_samples 3400 joint_samples 9 [566300, 1045187] processed_samples 3400 unjoint_samples 3400 joint_samples 9 [1034687, 632703] processed_samples 3400 unjoint_samples 3400 joint_samples 9 [1034687, 632703] processed_samples 3400 unjoint_samples 3400 joint_samples 9 [1038321, 787157] processed_samples 3400 unjoint_samples 3400 joint_samples 9 [1038321, 787157] [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure processed_samples 3500 unjoint_samples 3500 joint_samples 10 [1043359, 465641] processed_samples 3500 unjoint_samples 3500 joint_samples 10 [1043359, 465641] [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure processed_samples 3500 unjoint_samples 3500 joint_samples 9 [334143, 1047012] processed_samples 3500 unjoint_samples 3500 joint_samples 9 [674296, 1041969] processed_samples 3500 unjoint_samples 3500 joint_samples 9 [674296, 1041969] processed_samples 3500 unjoint_samples 3500 joint_samples 9 [334143, 1047012] processed_samples 3500 unjoint_samples 3500 joint_samples 10 [36988, 1046950] processed_samples 3500 unjoint_samples 3500 joint_samples 10 [36988, 1046950] processed_samples 3500 unjoint_samples 3500 joint_samples 9 [1045983, 881820] processed_samples 3500 unjoint_samples 3500 joint_samples 9 [628603, 1047832] processed_samples 3500 unjoint_samples 3500 joint_samples 9 [628603, 1047832] processed_samples 3500 unjoint_samples 3500 joint_samples 9 [1045983, 881820] processed_samples 3500 unjoint_samples 3500 joint_samples 9 [903830, 1045187] processed_samples 3500 unjoint_samples 3500 joint_samples 9 [903830, 1045187] processed_samples 3500 unjoint_samples 3500 joint_samples 9 [1034687, 942767] processed_samples 3500 unjoint_samples 3500 joint_samples 9 [1034687, 942767] [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x559618471cc0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x559616fdd040] mmco: unref short failure [h264 @ 0x559616fdd040] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure processed_samples 3600 unjoint_samples 3600 joint_samples 10 [1013764, 240227] processed_samples 3600 unjoint_samples 3600 joint_samples 10 [1013764, 240227] processed_samples 3600 unjoint_samples 3600 joint_samples 10 [374024, 1046950] processed_samples 3600 unjoint_samples 3600 joint_samples 10 [374024, 1046950] processed_samples 3600 unjoint_samples 3600 joint_samples 10 [242992, 1045277] processed_samples 3600 unjoint_samples 3600 joint_samples 10 [242992, 1045277] [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure processed_samples 3600 unjoint_samples 3600 joint_samples 9 [595489, 1047012] processed_samples 3600 unjoint_samples 3600 joint_samples 9 [595489, 1047012] processed_samples 3600 unjoint_samples 3600 joint_samples 10 [1043359, 794141] processed_samples 3600 unjoint_samples 3600 joint_samples 10 [1043359, 794141] processed_samples 3600 unjoint_samples 3600 joint_samples 10 [156814, 1047940] processed_samples 3600 unjoint_samples 3600 joint_samples 10 [156814, 1047940] processed_samples 3600 unjoint_samples 3600 joint_samples 9 [959252, 1047832] processed_samples 3600 unjoint_samples 3600 joint_samples 9 [925933, 1041969] processed_samples 3600 unjoint_samples 3600 joint_samples 9 [959252, 1047832] processed_samples 3600 unjoint_samples 3600 joint_samples 9 [925933, 1041969] [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617ebf51f00] illegal short term buffer state detected [h264 @ 0x559614220440] illegal short term buffer state detected [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617f06c03c0] [h264 @ 0x55961d3b8040] mmco: unref short failure mmco: unref short failure [h264 @ 0x5617f06c03c0] [h264 @ 0x55961d3b8040] mmco: unref short failure mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure processed_samples 3700 unjoint_samples 3700 joint_samples 10 [1046909, 144299] processed_samples 3700 unjoint_samples 3700 joint_samples 10 [1046909, 144299] processed_samples 3700 unjoint_samples 3700 joint_samples 11 [61397, 1047349] processed_samples 3700 unjoint_samples 3700 joint_samples 11 [61397, 1047349] processed_samples 3700 unjoint_samples 3700 joint_samples 10 [498007, 1045277] processed_samples 3700 unjoint_samples 3700 joint_samples 10 [498007, 1045277] [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure processed_samples 3700 unjoint_samples 3700 joint_samples 10 [1028195, 200412] processed_samples 3700 unjoint_samples 3700 joint_samples 10 [1013764, 495071] [h264 @ 0x5617ebde1580] mmco: unref short failure processed_samples 3700 unjoint_samples 3700 joint_samples 10 [430344, 1047940] processed_samples 3700 unjoint_samples 3700 joint_samples 10 [1028195, 200412] [h264 @ 0x5596172068c0] mmco: unref short failure processed_samples 3700 unjoint_samples 3700 joint_samples 10 [1013764, 495071] processed_samples 3700 unjoint_samples 3700 joint_samples 10 [430344, 1047940] processed_samples 3700 unjoint_samples 3700 joint_samples 9 [880914, 1047012] processed_samples 3700 unjoint_samples 3700 joint_samples 9 [880914, 1047012] processed_samples 3700 unjoint_samples 3700 joint_samples 10 [726324, 1046950] processed_samples 3700 unjoint_samples 3700 joint_samples 10 [726324, 1046950] [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure processed_samples 3800 unjoint_samples 3800 joint_samples 10 [1046909, 417115] processed_samples 3800 unjoint_samples 3800 joint_samples 10 [1028195, 479393] processed_samples 3800 unjoint_samples 3800 joint_samples 10 [1021280, 225092] processed_samples 3800 unjoint_samples 3800 joint_samples 11 [389313, 1047349] processed_samples 3800 unjoint_samples 3800 joint_samples 10 [1046909, 417115] processed_samples 3800 unjoint_samples 3800 joint_samples 10 [1021280, 225092] processed_samples 3800 unjoint_samples 3800 joint_samples 10 [1028195, 479393] processed_samples 3800 unjoint_samples 3800 joint_samples 11 [389313, 1047349] processed_samples 3800 unjoint_samples 3800 joint_samples 10 [1013764, 790729] processed_samples 3800 unjoint_samples 3800 joint_samples 10 [1013764, 790729] processed_samples 3800 unjoint_samples 3800 joint_samples 10 [751464, 1045277] processed_samples 3800 unjoint_samples 3800 joint_samples 10 [751464, 1045277] processed_samples 3800 unjoint_samples 3800 joint_samples 10 [812341, 1047940] processed_samples 3800 unjoint_samples 3800 joint_samples 10 [812341, 1047940] processed_samples 3800 unjoint_samples 3800 joint_samples 11 [1032368, 43842] processed_samples 3800 unjoint_samples 3800 joint_samples 11 [1032368, 43842] [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x559614658100] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x559614658100] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559619420380] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure processed_samples 3900 unjoint_samples 3900 joint_samples 11 [643737, 1047349] processed_samples 3900 unjoint_samples 3900 joint_samples 11 [643737, 1047349] [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure processed_samples 3900 unjoint_samples 3900 joint_samples 10 [1021280, 494444] processed_samples 3900 unjoint_samples 3900 joint_samples 10 [1021280, 494444] processed_samples 3900 unjoint_samples 3900 joint_samples 10 [1046909, 730504] processed_samples 3900 unjoint_samples 3900 joint_samples 10 [1046909, 730504] processed_samples 3900 unjoint_samples 3900 joint_samples 11 [1044904, 77058] processed_samples 3900 unjoint_samples 3900 joint_samples 11 [1044904, 77058] processed_samples 3900 unjoint_samples 3900 joint_samples 10 [1025303, 1023192] processed_samples 3900 unjoint_samples 3900 joint_samples 10 [1028195, 765379] processed_samples 3900 unjoint_samples 3900 joint_samples 10 [1025303, 1023192] processed_samples 3900 unjoint_samples 3900 joint_samples 10 [1028195, 765379] processed_samples 3900 unjoint_samples 3900 joint_samples 10 [1026577, 1045277] processed_samples 3900 unjoint_samples 3900 joint_samples 10 [1026577, 1045277] processed_samples 3900 unjoint_samples 3900 joint_samples 11 [1032368, 510863] processed_samples 3900 unjoint_samples 3900 joint_samples 11 [1032368, 510863] [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure processed_samples 4000 unjoint_samples 4000 joint_samples 11 [1032368, 765792] processed_samples 4000 unjoint_samples 4000 joint_samples 11 [1032368, 765792] processed_samples 4000 unjoint_samples 4000 joint_samples 11 [1033418, 129400] processed_samples 4000 unjoint_samples 4000 joint_samples 11 [1033418, 129400] processed_samples 4000 unjoint_samples 4000 joint_samples 11 [1044904, 328296] processed_samples 4000 unjoint_samples 4000 joint_samples 11 [1044904, 328296] processed_samples 4000 unjoint_samples 4000 joint_samples 11 [1048497, 299731] processed_samples 4000 unjoint_samples 4000 joint_samples 11 [1048497, 299731] processed_samples 4000 unjoint_samples 4000 joint_samples 11 [1047384, 247018] processed_samples 4000 unjoint_samples 4000 joint_samples 11 [1047384, 247018] processed_samples 4000 unjoint_samples 4000 joint_samples 10 [1021280, 779934] processed_samples 4000 unjoint_samples 4000 joint_samples 10 [1021280, 779934] processed_samples 4000 unjoint_samples 4000 joint_samples 11 [945727, 1047349] processed_samples 4000 unjoint_samples 4000 joint_samples 11 [945727, 1047349] processed_samples 4000 unjoint_samples 4000 joint_samples 10 [1046909, 1027629] processed_samples 4000 unjoint_samples 4000 joint_samples 10 [1046909, 1027629] [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure processed_samples 4100 unjoint_samples 4100 joint_samples 11 [1048497, 558035] processed_samples 4100 unjoint_samples 4100 joint_samples 11 [1048497, 558035] processed_samples 4100 unjoint_samples 4100 joint_samples 12 [1046606, 222958] processed_samples 4100 unjoint_samples 4100 joint_samples 12 [1046606, 222958] processed_samples 4100 unjoint_samples 4100 joint_samples 12 [10839, 1046239] processed_samples 4100 unjoint_samples 4100 joint_samples 12 [10839, 1046239] processed_samples 4100 unjoint_samples 4100 joint_samples 11 [1046909, 297000] processed_samples 4100 unjoint_samples 4100 joint_samples 11 [1046909, 297000] processed_samples 4100 unjoint_samples 4100 joint_samples 11 [1033418, 447504] processed_samples 4100 unjoint_samples 4100 joint_samples 11 [1033418, 447504] processed_samples 4100 unjoint_samples 4100 joint_samples 11 [1044904, 757427] processed_samples 4100 unjoint_samples 4100 joint_samples 11 [1047384, 586744] processed_samples 4100 unjoint_samples 4100 joint_samples 11 [1047384, 586744] processed_samples 4100 unjoint_samples 4100 joint_samples 11 [1044904, 757427] processed_samples 4100 unjoint_samples 4100 joint_samples 10 [1021280, 1017719] processed_samples 4100 unjoint_samples 4100 joint_samples 10 [1021280, 1017719] [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e9262740] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure processed_samples 4200 unjoint_samples 4200 joint_samples 11 [1033418, 817176] processed_samples 4200 unjoint_samples 4200 joint_samples 11 [1033418, 817176] processed_samples 4200 unjoint_samples 4200 joint_samples 12 [1047353, 8576] processed_samples 4200 unjoint_samples 4200 joint_samples 12 [1047353, 8576] processed_samples 4200 unjoint_samples 4200 joint_samples 11 [248022, 1043492] processed_samples 4200 unjoint_samples 4200 joint_samples 11 [248022, 1043492] processed_samples 4200 unjoint_samples 4200 joint_samples 12 [1046606, 464255] processed_samples 4200 unjoint_samples 4200 joint_samples 12 [1046606, 464255] processed_samples 4200 unjoint_samples 4200 joint_samples 11 [1046909, 813051] processed_samples 4200 unjoint_samples 4200 joint_samples 11 [1046909, 813051] processed_samples 4200 unjoint_samples 4200 joint_samples 12 [290554, 1046239] processed_samples 4200 unjoint_samples 4200 joint_samples 12 [290554, 1046239] processed_samples 4200 unjoint_samples 4200 joint_samples 11 [1047384, 855521] processed_samples 4200 unjoint_samples 4200 joint_samples 11 [1047384, 855521] processed_samples 4200 unjoint_samples 4200 joint_samples 11 [1048497, 882663] processed_samples 4200 unjoint_samples 4200 joint_samples 11 [1048497, 882663] [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617e9262740] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure processed_samples 4300 unjoint_samples 4300 joint_samples 12 [119481, 1046566] processed_samples 4300 unjoint_samples 4300 joint_samples 12 [119481, 1046566] processed_samples 4300 unjoint_samples 4300 joint_samples 12 [1046996, 63417] processed_samples 4300 unjoint_samples 4300 joint_samples 12 [1046996, 63417] processed_samples 4300 unjoint_samples 4300 joint_samples 12 [179760, 1046236] processed_samples 4300 unjoint_samples 4300 joint_samples 12 [179760, 1046236] processed_samples 4300 unjoint_samples 4300 joint_samples 11 [642076, 1043492] processed_samples 4300 unjoint_samples 4300 joint_samples 11 [642076, 1043492] processed_samples 4300 unjoint_samples 4300 joint_samples 12 [254742, 1044425] processed_samples 4300 unjoint_samples 4300 joint_samples 12 [254742, 1044425] processed_samples 4300 unjoint_samples 4300 joint_samples 12 [627784, 1046239] processed_samples 4300 unjoint_samples 4300 joint_samples 12 [627784, 1046239] [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure processed_samples 4300 unjoint_samples 4300 joint_samples 12 [1047353, 336395] processed_samples 4300 unjoint_samples 4300 joint_samples 12 [1047353, 336395] processed_samples 4300 unjoint_samples 4300 joint_samples 12 [1046606, 723914] processed_samples 4300 unjoint_samples 4300 joint_samples 12 [1046606, 723914] [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] [h264 @ 0x55961c1ba780] mmco: unref short failure mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ed3f3040] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e803bd40] mmco: unref short failure [h264 @ 0x5617e803bd40] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e803bd40] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure processed_samples 4400 unjoint_samples 4400 joint_samples 12 [328452, 1046566] processed_samples 4400 unjoint_samples 4400 joint_samples 12 [328452, 1046566] processed_samples 4400 unjoint_samples 4400 joint_samples 12 [585715, 1046236] processed_samples 4400 unjoint_samples 4400 joint_samples 12 [1046606, 1007750] processed_samples 4400 unjoint_samples 4400 joint_samples 12 [585715, 1046236] processed_samples 4400 unjoint_samples 4400 joint_samples 12 [1046996, 469224] processed_samples 4400 unjoint_samples 4400 joint_samples 12 [553541, 1044425] processed_samples 4400 unjoint_samples 4400 joint_samples 11 [1003937, 1043492] processed_samples 4400 unjoint_samples 4400 joint_samples 12 [1046606, 1007750] processed_samples 4400 unjoint_samples 4400 joint_samples 12 [1046996, 469224] processed_samples 4400 unjoint_samples 4400 joint_samples 12 [1047353, 818228] processed_samples 4400 unjoint_samples 4400 joint_samples 12 [553541, 1044425] processed_samples 4400 unjoint_samples 4400 joint_samples 11 [1003937, 1043492] processed_samples 4400 unjoint_samples 4400 joint_samples 12 [930937, 1046239] processed_samples 4400 unjoint_samples 4400 joint_samples 12 [1047353, 818228] processed_samples 4400 unjoint_samples 4400 joint_samples 12 [930937, 1046239] [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e8001bc0] Missing reference picture, default is 65524 [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5596170d5600] Missing reference picture, default is 65524 [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ee8bc680] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure processed_samples 4500 unjoint_samples 4500 joint_samples 13 [1047508, 11801] processed_samples 4500 unjoint_samples 4500 joint_samples 13 [1047508, 11801] processed_samples 4500 unjoint_samples 4500 joint_samples 12 [186197, 1046901] processed_samples 4500 unjoint_samples 4500 joint_samples 12 [186197, 1046901] processed_samples 4500 unjoint_samples 4500 joint_samples 13 [1045393, 266242] processed_samples 4500 unjoint_samples 4500 joint_samples 13 [1045393, 266242] processed_samples 4500 unjoint_samples 4500 joint_samples 12 [1046996, 745529] processed_samples 4500 unjoint_samples 4500 joint_samples 12 [1046996, 745529] processed_samples 4500 unjoint_samples 4500 joint_samples 13 [380462, 1034451] processed_samples 4500 unjoint_samples 4500 joint_samples 13 [380462, 1034451] processed_samples 4500 unjoint_samples 4500 joint_samples 12 [872481, 1044425] processed_samples 4500 unjoint_samples 4500 joint_samples 12 [872481, 1044425] processed_samples 4500 unjoint_samples 4500 joint_samples 12 [864053, 1046236] processed_samples 4500 unjoint_samples 4500 joint_samples 12 [864053, 1046236] processed_samples 4500 unjoint_samples 4500 joint_samples 12 [631633, 1046566] processed_samples 4500 unjoint_samples 4500 joint_samples 12 [631633, 1046566] [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ed3f3040] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure processed_samples 4600 unjoint_samples 4600 joint_samples 13 [1046649, 126849] processed_samples 4600 unjoint_samples 4600 joint_samples 13 [1046649, 126849] processed_samples 4600 unjoint_samples 4600 joint_samples 13 [1039121, 191764] processed_samples 4600 unjoint_samples 4600 joint_samples 13 [1039121, 191764] processed_samples 4600 unjoint_samples 4600 joint_samples 12 [1000805, 1046566] processed_samples 4600 unjoint_samples 4600 joint_samples 12 [1000805, 1046566] processed_samples 4600 unjoint_samples 4600 joint_samples 13 [1047508, 266154] processed_samples 4600 unjoint_samples 4600 joint_samples 13 [1047508, 266154] processed_samples 4600 unjoint_samples 4600 joint_samples 13 [680443, 1034451] processed_samples 4600 unjoint_samples 4600 joint_samples 13 [680443, 1034451] processed_samples 4600 unjoint_samples 4600 joint_samples 13 [1045393, 600872] processed_samples 4600 unjoint_samples 4600 joint_samples 13 [1045393, 600872] processed_samples 4600 unjoint_samples 4600 joint_samples 12 [617981, 1046901] processed_samples 4600 unjoint_samples 4600 joint_samples 12 [617981, 1046901] [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure processed_samples 4600 unjoint_samples 4600 joint_samples 12 [1046996, 1007840] processed_samples 4600 unjoint_samples 4600 joint_samples 12 [1046996, 1007840] [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure processed_samples 4700 unjoint_samples 4700 joint_samples 13 [203382, 1046566] processed_samples 4700 unjoint_samples 4700 joint_samples 13 [203382, 1046566] processed_samples 4700 unjoint_samples 4700 joint_samples 13 [1039121, 537765] processed_samples 4700 unjoint_samples 4700 joint_samples 13 [309569, 1040750] processed_samples 4700 unjoint_samples 4700 joint_samples 13 [1039121, 537765] processed_samples 4700 unjoint_samples 4700 joint_samples 13 [309569, 1040750] processed_samples 4700 unjoint_samples 4700 joint_samples 13 [1046649, 419952] processed_samples 4700 unjoint_samples 4700 joint_samples 13 [1046649, 419952] processed_samples 4700 unjoint_samples 4700 joint_samples 13 [1047508, 631913] processed_samples 4700 unjoint_samples 4700 joint_samples 13 [1047508, 631913] processed_samples 4700 unjoint_samples 4700 joint_samples 12 [878001, 1046901] processed_samples 4700 unjoint_samples 4700 joint_samples 12 [878001, 1046901] processed_samples 4700 unjoint_samples 4700 joint_samples 13 [1045393, 959240] processed_samples 4700 unjoint_samples 4700 joint_samples 13 [1045393, 959240] processed_samples 4700 unjoint_samples 4700 joint_samples 13 [901220, 1034451] processed_samples 4700 unjoint_samples 4700 joint_samples 13 [901220, 1034451] [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559617afc780] mmco: unref short failure [h264 @ 0x5617e8298040] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e7fbb200] [h264 @ 0x559616359900] mmco: unref short failure mmco: unref short failure processed_samples 4800 unjoint_samples 4800 joint_samples 13 [697964, 1046566] processed_samples 4800 unjoint_samples 4800 joint_samples 13 [697964, 1046566] processed_samples 4800 unjoint_samples 4800 joint_samples 14 [166852, 1047852] processed_samples 4800 unjoint_samples 4800 joint_samples 14 [166852, 1047852] processed_samples 4800 unjoint_samples 4800 joint_samples 13 [1012646, 283821] processed_samples 4800 unjoint_samples 4800 joint_samples 13 [1012646, 283821] processed_samples 4800 unjoint_samples 4800 joint_samples 14 [344305, 1028606] processed_samples 4800 unjoint_samples 4800 joint_samples 14 [344305, 1028606] processed_samples 4800 unjoint_samples 4800 joint_samples 13 [734840, 1040750] processed_samples 4800 unjoint_samples 4800 joint_samples 13 [734840, 1040750] processed_samples 4800 unjoint_samples 4800 joint_samples 13 [1046649, 773398] processed_samples 4800 unjoint_samples 4800 joint_samples 13 [1046649, 773398] processed_samples 4800 unjoint_samples 4800 joint_samples 13 [1047508, 867406] processed_samples 4800 unjoint_samples 4800 joint_samples 13 [1047508, 867406] processed_samples 4800 unjoint_samples 4800 joint_samples 13 [1039121, 878627] processed_samples 4800 unjoint_samples 4800 joint_samples 13 [1039121, 878627] [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ec2f5f40] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure processed_samples 4900 unjoint_samples 4900 joint_samples 14 [507496, 1047852] processed_samples 4900 unjoint_samples 4900 joint_samples 14 [20429, 1046392] processed_samples 4900 unjoint_samples 4900 joint_samples 14 [507496, 1047852] processed_samples 4900 unjoint_samples 4900 joint_samples 14 [20429, 1046392] processed_samples 4900 unjoint_samples 4900 joint_samples 14 [1030200, 83163] processed_samples 4900 unjoint_samples 4900 joint_samples 14 [1030200, 83163] processed_samples 4900 unjoint_samples 4900 joint_samples 13 [1012646, 543229] processed_samples 4900 unjoint_samples 4900 joint_samples 13 [1012646, 543229] processed_samples 4900 unjoint_samples 4900 joint_samples 14 [124342, 1048011] processed_samples 4900 unjoint_samples 4900 joint_samples 14 [124342, 1048011] processed_samples 4900 unjoint_samples 4900 joint_samples 14 [1047008, 357503] processed_samples 4900 unjoint_samples 4900 joint_samples 14 [673323, 1028606] processed_samples 4900 unjoint_samples 4900 joint_samples 14 [1047008, 357503] processed_samples 4900 unjoint_samples 4900 joint_samples 14 [673323, 1028606] processed_samples 4900 unjoint_samples 4900 joint_samples 13 [1018464, 1040750] processed_samples 4900 unjoint_samples 4900 joint_samples 13 [1018464, 1040750] [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [mov,mp4,m4a,3gp,3g2,mj2 @ 0x5617e839f900] [mov,mp4,m4a,3gp,3g2,mj2 @ 0x55961c611dc0] stream 1, offset 0x1400a4d: partial file stream 1, offset 0x1400a4d: partial file [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x55961afcdd40] mmco: unref short failure [h264 @ 0x55961afcdd40] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure processed_samples 5000 unjoint_samples 5000 joint_samples 13 [1012646, 846074] processed_samples 5000 unjoint_samples 5000 joint_samples 13 [1012646, 846074] processed_samples 5000 unjoint_samples 5000 joint_samples 14 [1030200, 364791] processed_samples 5000 unjoint_samples 5000 joint_samples 14 [503859, 1048011] processed_samples 5000 unjoint_samples 5000 joint_samples 14 [1047008, 609019] processed_samples 5000 unjoint_samples 5000 joint_samples 14 [1030200, 364791] processed_samples 5000 unjoint_samples 5000 joint_samples 14 [259342, 1046392] processed_samples 5000 unjoint_samples 5000 joint_samples 14 [892190, 1047852] processed_samples 5000 unjoint_samples 5000 joint_samples 14 [503859, 1048011] processed_samples 5000 unjoint_samples 5000 joint_samples 14 [259342, 1046392] processed_samples 5000 unjoint_samples 5000 joint_samples 14 [1047008, 609019] processed_samples 5000 unjoint_samples 5000 joint_samples 14 [892190, 1047852] processed_samples 5001 unjoint_samples 5000 joint_samples 14 [284084, 1045179] processed_samples 5001 unjoint_samples 5000 joint_samples 14 [284084, 1045179] processed_samples 5000 unjoint_samples 5000 joint_samples 14 [993158, 1028606] processed_samples 5000 unjoint_samples 5000 joint_samples 14 [993158, 1028606] [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e84806c0] [h264 @ 0x5596183cd8c0] mmco: unref short failure mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure processed_samples 5100 unjoint_samples 5100 joint_samples 14 [176303, 1015128] processed_samples 5100 unjoint_samples 5100 joint_samples 14 [176303, 1015128] processed_samples 5100 unjoint_samples 5100 joint_samples 15 [1042528, 365017] processed_samples 5100 unjoint_samples 5100 joint_samples 15 [1042528, 365017] processed_samples 5100 unjoint_samples 5100 joint_samples 14 [1030200, 662822] processed_samples 5100 unjoint_samples 5100 joint_samples 14 [632412, 1046392] processed_samples 5100 unjoint_samples 5100 joint_samples 14 [632412, 1046392] processed_samples 5100 unjoint_samples 5100 joint_samples 14 [1030200, 662822] processed_samples 5100 unjoint_samples 5100 joint_samples 14 [1047008, 982031] processed_samples 5100 unjoint_samples 5100 joint_samples 15 [377599, 1028606] processed_samples 5100 unjoint_samples 5100 joint_samples 15 [377599, 1028606] processed_samples 5100 unjoint_samples 5100 joint_samples 14 [1047008, 982031] processed_samples 5101 unjoint_samples 5100 joint_samples 14 [668918, 1045179] processed_samples 5100 unjoint_samples 5100 joint_samples 14 [851664, 1048011] processed_samples 5101 unjoint_samples 5100 joint_samples 14 [668918, 1045179] processed_samples 5100 unjoint_samples 5100 joint_samples 14 [851664, 1048011] [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e9262740] mmco: unref short failure [h264 @ 0x5617e9262740] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559618471cc0] mmco: unref short failure [h264 @ 0x559618471cc0] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617ed3f3040] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure processed_samples 5200 unjoint_samples 5200 joint_samples 15 [1047935, 87041] [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure processed_samples 5200 unjoint_samples 5200 joint_samples 15 [1047935, 87041] processed_samples 5200 unjoint_samples 5200 joint_samples 14 [984745, 1046392] [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure processed_samples 5200 unjoint_samples 5200 joint_samples 15 [1042528, 649091] processed_samples 5200 unjoint_samples 5200 joint_samples 14 [984745, 1046392] processed_samples 5200 unjoint_samples 5200 joint_samples 15 [598013, 1028606] processed_samples 5200 unjoint_samples 5200 joint_samples 15 [1042528, 649091] processed_samples 5200 unjoint_samples 5200 joint_samples 14 [583371, 1015128] processed_samples 5200 unjoint_samples 5200 joint_samples 14 [583371, 1015128] processed_samples 5200 unjoint_samples 5200 joint_samples 15 [598013, 1028606] processed_samples 5200 unjoint_samples 5200 joint_samples 14 [1030200, 987161] processed_samples 5200 unjoint_samples 5200 joint_samples 14 [1030200, 987161] processed_samples 5200 unjoint_samples 5200 joint_samples 15 [1047008, 202665] processed_samples 5200 unjoint_samples 5200 joint_samples 15 [1047008, 202665] processed_samples 5201 unjoint_samples 5200 joint_samples 14 [987454, 1045179] processed_samples 5201 unjoint_samples 5200 joint_samples 14 [987454, 1045179] [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure processed_samples 5300 unjoint_samples 5300 joint_samples 16 [78425, 1033790] processed_samples 5300 unjoint_samples 5300 joint_samples 15 [1047008, 547758] processed_samples 5300 unjoint_samples 5300 joint_samples 15 [1047935, 336770] processed_samples 5300 unjoint_samples 5300 joint_samples 16 [78425, 1033790] processed_samples 5300 unjoint_samples 5300 joint_samples 15 [1047008, 547758] processed_samples 5300 unjoint_samples 5300 joint_samples 15 [1047935, 336770] processed_samples 5300 unjoint_samples 5300 joint_samples 15 [1043000, 505142] processed_samples 5300 unjoint_samples 5300 joint_samples 15 [1043000, 505142] processed_samples 5300 unjoint_samples 5300 joint_samples 15 [1018764, 320149] processed_samples 5300 unjoint_samples 5300 joint_samples 15 [1018764, 320149] processed_samples 5300 unjoint_samples 5300 joint_samples 14 [940482, 1015128] processed_samples 5300 unjoint_samples 5300 joint_samples 14 [940482, 1015128] processed_samples 5301 unjoint_samples 5300 joint_samples 15 [282445, 1047481] processed_samples 5301 unjoint_samples 5300 joint_samples 15 [282445, 1047481] processed_samples 5300 unjoint_samples 5300 joint_samples 15 [923052, 1028606] processed_samples 5300 unjoint_samples 5300 joint_samples 15 [923052, 1028606] [h264 @ 0x5617eb9e6580] mmco: unref short failure [h264 @ 0x5617eb9e6580] mmco: unref short failure [h264 @ 0x5617eb9e6580] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e7e34fc0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x5617e955dbc0] Missing reference picture, default is 65530 [h264 @ 0x5617e955dbc0] Missing reference picture, default is 65530 [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x559617100c80] Missing reference picture, default is 65530 [h264 @ 0x559617100c80] Missing reference picture, default is 65530 [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e955dbc0] Missing reference picture, default is 65530 [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] Missing reference picture, default is 65530 [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x559617100c80] Missing reference picture, default is 65530 [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] Missing reference picture, default is 65530 [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure processed_samples 5400 unjoint_samples 5400 joint_samples 16 [340095, 1033790] processed_samples 5400 unjoint_samples 5400 joint_samples 16 [340095, 1033790] processed_samples 5400 unjoint_samples 5400 joint_samples 15 [266282, 1043245] processed_samples 5400 unjoint_samples 5400 joint_samples 15 [266282, 1043245] processed_samples 5400 unjoint_samples 5400 joint_samples 16 [95087, 1046995] processed_samples 5400 unjoint_samples 5400 joint_samples 16 [95087, 1046995] processed_samples 5401 unjoint_samples 5400 joint_samples 15 [524457, 1047481] processed_samples 5400 unjoint_samples 5400 joint_samples 15 [1018764, 567012] [h264 @ 0x5617e897a680] mmco: unref short failure processed_samples 5401 unjoint_samples 5400 joint_samples 15 [524457, 1047481] processed_samples 5400 unjoint_samples 5400 joint_samples 15 [1018764, 567012] [h264 @ 0x559618395d00] mmco: unref short failure processed_samples 5400 unjoint_samples 5400 joint_samples 15 [1047008, 844655] processed_samples 5400 unjoint_samples 5400 joint_samples 15 [1047008, 844655] processed_samples 5400 unjoint_samples 5400 joint_samples 15 [1047935, 562325] processed_samples 5400 unjoint_samples 5400 joint_samples 15 [1047935, 562325] processed_samples 5400 unjoint_samples 5400 joint_samples 15 [1043000, 757202] processed_samples 5400 unjoint_samples 5400 joint_samples 15 [1043000, 757202] [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [mov,mp4,m4a,3gp,3g2,mj2 @ 0x55961acb34c0] stream 0, offset 0x90050f: partial file [mov,mp4,m4a,3gp,3g2,mj2 @ 0x5617ebf72380] stream 0, offset 0x90050f: partial file [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure processed_samples 5500 unjoint_samples 5500 joint_samples 15 [574737, 1043245] processed_samples 5500 unjoint_samples 5500 joint_samples 16 [1047008, 82331] processed_samples 5500 unjoint_samples 5500 joint_samples 15 [574737, 1043245] processed_samples 5500 unjoint_samples 5500 joint_samples 16 [1047008, 82331] processed_samples 5500 unjoint_samples 5500 joint_samples 15 [1043000, 1017350] processed_samples 5500 unjoint_samples 5500 joint_samples 15 [1043000, 1017350] processed_samples 5500 unjoint_samples 5500 joint_samples 16 [440345, 1046995] processed_samples 5500 unjoint_samples 5500 joint_samples 16 [440345, 1046995] processed_samples 5501 unjoint_samples 5500 joint_samples 16 [679506, 1033790] processed_samples 5500 unjoint_samples 5500 joint_samples 15 [1047935, 823331] processed_samples 5501 unjoint_samples 5500 joint_samples 15 [815713, 1047481] processed_samples 5500 unjoint_samples 5500 joint_samples 15 [1018764, 936633] processed_samples 5501 unjoint_samples 5500 joint_samples 16 [679506, 1033790] processed_samples 5500 unjoint_samples 5500 joint_samples 15 [1047935, 823331] processed_samples 5501 unjoint_samples 5500 joint_samples 15 [815713, 1047481] processed_samples 5500 unjoint_samples 5500 joint_samples 15 [1018764, 936633] [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ea25da40] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure processed_samples 5600 unjoint_samples 5600 joint_samples 16 [132095, 1046575] processed_samples 5600 unjoint_samples 5600 joint_samples 16 [132095, 1046575] processed_samples 5601 unjoint_samples 5600 joint_samples 16 [1046992, 68271] processed_samples 5601 unjoint_samples 5600 joint_samples 16 [1046992, 68271] processed_samples 5600 unjoint_samples 5600 joint_samples 16 [1047008, 484447] processed_samples 5600 unjoint_samples 5600 joint_samples 16 [1034011, 301015] processed_samples 5600 unjoint_samples 5600 joint_samples 16 [1034011, 301015] [h264 @ 0x5617e9426340] mmco: unref short failure processed_samples 5600 unjoint_samples 5600 joint_samples 16 [1045171, 284899] processed_samples 5600 unjoint_samples 5600 joint_samples 16 [1045171, 284899] processed_samples 5600 unjoint_samples 5600 joint_samples 16 [1047008, 484447] [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure processed_samples 5600 unjoint_samples 5600 joint_samples 15 [811627, 1043245] processed_samples 5600 unjoint_samples 5600 joint_samples 15 [811627, 1043245] processed_samples 5601 unjoint_samples 5600 joint_samples 17 [883114, 372746] processed_samples 5601 unjoint_samples 5600 joint_samples 17 [883114, 372746] processed_samples 5600 unjoint_samples 5600 joint_samples 16 [714517, 1046995] processed_samples 5600 unjoint_samples 5600 joint_samples 16 [714517, 1046995] [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e9262740] mmco: unref short failure [h264 @ 0x5617e9262740] mmco: unref short failure [h264 @ 0x5617e9262740] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e9262740] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure processed_samples 5700 unjoint_samples 5700 joint_samples 16 [111286, 1047557] processed_samples 5700 unjoint_samples 5700 joint_samples 16 [111286, 1047557] processed_samples 5700 unjoint_samples 5700 joint_samples 16 [455671, 1046575] processed_samples 5700 unjoint_samples 5700 joint_samples 16 [455671, 1046575] processed_samples 5700 unjoint_samples 5700 joint_samples 16 [1034011, 682162] processed_samples 5700 unjoint_samples 5700 joint_samples 16 [1047008, 740741] processed_samples 5700 unjoint_samples 5700 joint_samples 16 [1047008, 740741] processed_samples 5700 unjoint_samples 5700 joint_samples 16 [1034011, 682162] processed_samples 5700 unjoint_samples 5700 joint_samples 16 [1045171, 619598] processed_samples 5700 unjoint_samples 5700 joint_samples 16 [1045171, 619598] processed_samples 5701 unjoint_samples 5700 joint_samples 16 [1046992, 603906] processed_samples 5701 unjoint_samples 5700 joint_samples 16 [1046992, 603906] processed_samples 5701 unjoint_samples 5700 joint_samples 17 [883114, 674503] processed_samples 5701 unjoint_samples 5700 joint_samples 17 [883114, 674503] processed_samples 5700 unjoint_samples 5700 joint_samples 16 [1030890, 1046995] processed_samples 5700 unjoint_samples 5700 joint_samples 16 [1030890, 1046995] [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure processed_samples 5800 unjoint_samples 5800 joint_samples 16 [357365, 1047557] processed_samples 5800 unjoint_samples 5800 joint_samples 17 [141019, 1043346] processed_samples 5800 unjoint_samples 5800 joint_samples 17 [268926, 1046995] processed_samples 5800 unjoint_samples 5800 joint_samples 16 [357365, 1047557] processed_samples 5800 unjoint_samples 5800 joint_samples 16 [1045171, 942224] processed_samples 5800 unjoint_samples 5800 joint_samples 17 [141019, 1043346] processed_samples 5800 unjoint_samples 5800 joint_samples 16 [1045171, 942224] processed_samples 5800 unjoint_samples 5800 joint_samples 17 [268926, 1046995] processed_samples 5801 unjoint_samples 5800 joint_samples 16 [1046992, 874722] processed_samples 5800 unjoint_samples 5800 joint_samples 16 [736962, 1046575] processed_samples 5800 unjoint_samples 5800 joint_samples 16 [1034011, 940816] processed_samples 5800 unjoint_samples 5800 joint_samples 16 [1034011, 940816] processed_samples 5800 unjoint_samples 5800 joint_samples 16 [736962, 1046575] processed_samples 5801 unjoint_samples 5800 joint_samples 17 [939632, 940315] processed_samples 5801 unjoint_samples 5800 joint_samples 16 [1046992, 874722] processed_samples 5801 unjoint_samples 5800 joint_samples 17 [939632, 940315] [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x559616d35c80] mmco: unref short failure [h264 @ 0x559616d35c80] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure processed_samples 5900 unjoint_samples 5900 joint_samples 16 [608620, 1047557] processed_samples 5900 unjoint_samples 5900 joint_samples 16 [608620, 1047557] processed_samples 5900 unjoint_samples 5900 joint_samples 17 [223055, 1046860] [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure processed_samples 5900 unjoint_samples 5900 joint_samples 17 [223055, 1046860] processed_samples 5900 unjoint_samples 5900 joint_samples 17 [1045296, 228097] processed_samples 5900 unjoint_samples 5900 joint_samples 17 [1045296, 228097] processed_samples 5900 unjoint_samples 5900 joint_samples 17 [665648, 1046995] processed_samples 5900 unjoint_samples 5900 joint_samples 17 [665648, 1046995] processed_samples 5901 unjoint_samples 5900 joint_samples 18 [45413, 1043348] processed_samples 5900 unjoint_samples 5900 joint_samples 17 [494344, 1043346] processed_samples 5900 unjoint_samples 5900 joint_samples 17 [494344, 1043346] processed_samples 5901 unjoint_samples 5900 joint_samples 17 [1047479, 163515] processed_samples 5901 unjoint_samples 5900 joint_samples 18 [45413, 1043348] processed_samples 5901 unjoint_samples 5900 joint_samples 17 [1047479, 163515] processed_samples 5900 unjoint_samples 5900 joint_samples 16 [1019753, 1046575] processed_samples 5900 unjoint_samples 5900 joint_samples 16 [1019753, 1046575] [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure processed_samples 6000 unjoint_samples 6000 joint_samples 17 [1045296, 568500] processed_samples 6000 unjoint_samples 6000 joint_samples 17 [312676, 1046575] processed_samples 6001 unjoint_samples 6000 joint_samples 18 [368422, 1043348] processed_samples 6000 unjoint_samples 6000 joint_samples 17 [1045296, 568500] processed_samples 6000 unjoint_samples 6000 joint_samples 17 [800420, 1043346] [h264 @ 0x5596183cd8c0] mmco: unref short failure processed_samples 6000 unjoint_samples 6000 joint_samples 17 [967578, 1046995] processed_samples 6000 unjoint_samples 6000 joint_samples 17 [443508, 1046860] processed_samples 6000 unjoint_samples 6000 joint_samples 17 [312676, 1046575] processed_samples 6001 unjoint_samples 6000 joint_samples 18 [368422, 1043348] [h264 @ 0x5617ee580e80] mmco: unref short failure processed_samples 6001 unjoint_samples 6000 joint_samples 17 [1047479, 653145] processed_samples 6000 unjoint_samples 6000 joint_samples 17 [800420, 1043346] processed_samples 6000 unjoint_samples 6000 joint_samples 17 [967578, 1046995] processed_samples 6000 unjoint_samples 6000 joint_samples 17 [443508, 1046860] processed_samples 6000 unjoint_samples 6000 joint_samples 16 [864460, 1047557] processed_samples 6001 unjoint_samples 6000 joint_samples 17 [1047479, 653145] processed_samples 6000 unjoint_samples 6000 joint_samples 16 [864460, 1047557] [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure processed_samples 6100 unjoint_samples 6100 joint_samples 17 [1046841, 106609] processed_samples 6100 unjoint_samples 6100 joint_samples 17 [1046841, 106609] processed_samples 6100 unjoint_samples 6100 joint_samples 17 [1045296, 891062] processed_samples 6100 unjoint_samples 6100 joint_samples 17 [1045296, 891062] processed_samples 6100 unjoint_samples 6100 joint_samples 18 [182940, 1046995] processed_samples 6100 unjoint_samples 6100 joint_samples 18 [182940, 1046995] processed_samples 6101 unjoint_samples 6100 joint_samples 18 [607826, 1043348] processed_samples 6101 unjoint_samples 6100 joint_samples 18 [607826, 1043348] processed_samples 6100 unjoint_samples 6100 joint_samples 17 [700322, 1046860] processed_samples 6100 unjoint_samples 6100 joint_samples 17 [700322, 1046860] processed_samples 6100 unjoint_samples 6100 joint_samples 17 [592448, 1046575] processed_samples 6100 unjoint_samples 6100 joint_samples 17 [592448, 1046575] [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure processed_samples 6100 unjoint_samples 6100 joint_samples 17 [1012974, 1043346] processed_samples 6100 unjoint_samples 6100 joint_samples 17 [1012974, 1043346] processed_samples 6101 unjoint_samples 6100 joint_samples 18 [1047479, 2299] processed_samples 6101 unjoint_samples 6100 joint_samples 18 [1047479, 2299] [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5596177086c0] mmco: unref short failure [h264 @ 0x5596177086c0] mmco: unref short failure [h264 @ 0x5617e8298040] mmco: unref short failure [h264 @ 0x5617e8298040] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5596177086c0] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure processed_samples 6200 unjoint_samples 6200 joint_samples 18 [122759, 1036177] processed_samples 6200 unjoint_samples 6200 joint_samples 18 [122759, 1036177] processed_samples 6200 unjoint_samples 6200 joint_samples 18 [1046383, 239436] processed_samples 6200 unjoint_samples 6200 joint_samples 18 [1046383, 239436] processed_samples 6200 unjoint_samples 6200 joint_samples 17 [1046841, 549302] processed_samples 6200 unjoint_samples 6200 joint_samples 17 [1046841, 549302] processed_samples 6200 unjoint_samples 6200 joint_samples 18 [463434, 1046995] processed_samples 6200 unjoint_samples 6200 joint_samples 18 [463434, 1046995] processed_samples 6200 unjoint_samples 6200 joint_samples 17 [919897, 1046575] processed_samples 6200 unjoint_samples 6200 joint_samples 17 [919897, 1046575] processed_samples 6201 unjoint_samples 6200 joint_samples 18 [1047479, 283070] processed_samples 6201 unjoint_samples 6200 joint_samples 18 [1047479, 283070] processed_samples 6201 unjoint_samples 6200 joint_samples 18 [961064, 1043348] processed_samples 6200 unjoint_samples 6200 joint_samples 17 [1035765, 1046860] processed_samples 6201 unjoint_samples 6200 joint_samples 18 [961064, 1043348] processed_samples 6200 unjoint_samples 6200 joint_samples 17 [1035765, 1046860] [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e8dbeec0] mmco: unref short failure [h264 @ 0x5617e8dbeec0] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure processed_samples 6300 unjoint_samples 6300 joint_samples 18 [398829, 1036177] [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure processed_samples 6300 unjoint_samples 6300 joint_samples 18 [398829, 1036177] [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure processed_samples 6301 unjoint_samples 6300 joint_samples 19 [314495, 1047945] processed_samples 6300 unjoint_samples 6300 joint_samples 18 [352687, 1046860] processed_samples 6301 unjoint_samples 6300 joint_samples 19 [314495, 1047945] processed_samples 6300 unjoint_samples 6300 joint_samples 18 [1046383, 539200] processed_samples 6300 unjoint_samples 6300 joint_samples 18 [1045806, 230712] processed_samples 6300 unjoint_samples 6300 joint_samples 18 [352687, 1046860] processed_samples 6300 unjoint_samples 6300 joint_samples 18 [1046383, 539200] processed_samples 6300 unjoint_samples 6300 joint_samples 18 [1045806, 230712] processed_samples 6300 unjoint_samples 6300 joint_samples 18 [708238, 1046995] processed_samples 6300 unjoint_samples 6300 joint_samples 18 [708238, 1046995] processed_samples 6301 unjoint_samples 6300 joint_samples 18 [1047479, 566310] processed_samples 6300 unjoint_samples 6300 joint_samples 17 [1046841, 872638] [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure processed_samples 6301 unjoint_samples 6300 joint_samples 18 [1047479, 566310] processed_samples 6300 unjoint_samples 6300 joint_samples 17 [1046841, 872638] [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ed3f3040] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure processed_samples 6400 unjoint_samples 6400 joint_samples 18 [133567, 1044085] processed_samples 6400 unjoint_samples 6400 joint_samples 18 [133567, 1044085] processed_samples 6400 unjoint_samples 6400 joint_samples 18 [691352, 1036177] processed_samples 6400 unjoint_samples 6400 joint_samples 18 [1046383, 814089] processed_samples 6400 unjoint_samples 6400 joint_samples 18 [691352, 1036177] processed_samples 6400 unjoint_samples 6400 joint_samples 18 [582480, 1046860] processed_samples 6401 unjoint_samples 6400 joint_samples 19 [629452, 1047945] processed_samples 6400 unjoint_samples 6400 joint_samples 18 [1045806, 507050] processed_samples 6400 unjoint_samples 6400 joint_samples 18 [1046383, 814089] processed_samples 6400 unjoint_samples 6400 joint_samples 18 [582480, 1046860] processed_samples 6400 unjoint_samples 6400 joint_samples 18 [1045806, 507050] processed_samples 6401 unjoint_samples 6400 joint_samples 19 [629452, 1047945] [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure processed_samples 6400 unjoint_samples 6400 joint_samples 18 [1026235, 1046995] processed_samples 6400 unjoint_samples 6400 joint_samples 18 [1026235, 1046995] [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure processed_samples 6401 unjoint_samples 6400 joint_samples 18 [1047479, 839932] [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure processed_samples 6401 unjoint_samples 6400 joint_samples 18 [1047479, 839932] [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ed3f3040] mmco: unref short failure [h264 @ 0x5617ed3f3040] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617ebf4c300] mmco: unref short failure [h264 @ 0x5617ebf4c300] mmco: unref short failure [h264 @ 0x5617ebf4c300] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure processed_samples 6500 unjoint_samples 6500 joint_samples 19 [1045010, 359699] processed_samples 6500 unjoint_samples 6500 joint_samples 19 [1045010, 359699] processed_samples 6500 unjoint_samples 6500 joint_samples 18 [484586, 1044085] processed_samples 6500 unjoint_samples 6500 joint_samples 18 [484586, 1044085] processed_samples 6500 unjoint_samples 6500 joint_samples 18 [1045806, 809895] processed_samples 6500 unjoint_samples 6500 joint_samples 18 [1045806, 809895] processed_samples 6501 unjoint_samples 6500 joint_samples 19 [81644, 1047012] processed_samples 6501 unjoint_samples 6500 joint_samples 19 [81644, 1047012] processed_samples 6500 unjoint_samples 6500 joint_samples 18 [931091, 1036177] processed_samples 6500 unjoint_samples 6500 joint_samples 19 [76987, 1048417] processed_samples 6500 unjoint_samples 6500 joint_samples 19 [76987, 1048417] processed_samples 6500 unjoint_samples 6500 joint_samples 18 [931091, 1036177] processed_samples 6500 unjoint_samples 6500 joint_samples 18 [876695, 1046860] processed_samples 6500 unjoint_samples 6500 joint_samples 18 [876695, 1046860] processed_samples 6501 unjoint_samples 6500 joint_samples 19 [1007326, 1047945] processed_samples 6501 unjoint_samples 6500 joint_samples 19 [1007326, 1047945] [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure processed_samples 6600 unjoint_samples 6600 joint_samples 19 [377272, 1048417] [h264 @ 0x559619462b80] mmco: unref short failure processed_samples 6600 unjoint_samples 6600 joint_samples 19 [377272, 1048417] processed_samples 6600 unjoint_samples 6600 joint_samples 19 [173491, 1046860] [h264 @ 0x559619462b80] mmco: unref short failure processed_samples 6600 unjoint_samples 6600 joint_samples 19 [173491, 1046860] [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure processed_samples 6600 unjoint_samples 6600 joint_samples 19 [198164, 1046514] processed_samples 6600 unjoint_samples 6600 joint_samples 19 [198164, 1046514] [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure processed_samples 6600 unjoint_samples 6600 joint_samples 18 [796781, 1044085] processed_samples 6600 unjoint_samples 6600 joint_samples 18 [796781, 1044085] processed_samples 6600 unjoint_samples 6600 joint_samples 19 [1045010, 622903] processed_samples 6600 unjoint_samples 6600 joint_samples 19 [1045010, 622903] processed_samples 6601 unjoint_samples 6600 joint_samples 20 [1043934, 266104] processed_samples 6601 unjoint_samples 6600 joint_samples 19 [356152, 1047012] processed_samples 6601 unjoint_samples 6600 joint_samples 20 [1043934, 266104] processed_samples 6601 unjoint_samples 6600 joint_samples 19 [356152, 1047012] processed_samples 6600 unjoint_samples 6600 joint_samples 19 [3397, 1047684] processed_samples 6600 unjoint_samples 6600 joint_samples 19 [3397, 1047684] [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559618471cc0] mmco: unref short failure [h264 @ 0x559618471cc0] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure processed_samples 6700 unjoint_samples 6700 joint_samples 19 [1047679, 66109] processed_samples 6700 unjoint_samples 6700 joint_samples 19 [1047679, 66109] processed_samples 6700 unjoint_samples 6700 joint_samples 19 [475146, 1046514] processed_samples 6700 unjoint_samples 6700 joint_samples 19 [475146, 1046514] processed_samples 6700 unjoint_samples 6700 joint_samples 19 [295482, 1047684] processed_samples 6700 unjoint_samples 6700 joint_samples 19 [295482, 1047684] processed_samples 6701 unjoint_samples 6700 joint_samples 20 [1043934, 579726] processed_samples 6701 unjoint_samples 6700 joint_samples 20 [1043934, 579726] processed_samples 6700 unjoint_samples 6700 joint_samples 19 [1045010, 936122] processed_samples 6700 unjoint_samples 6700 joint_samples 19 [1045010, 936122] processed_samples 6700 unjoint_samples 6700 joint_samples 19 [474937, 1046860] processed_samples 6700 unjoint_samples 6700 joint_samples 19 [474937, 1046860] processed_samples 6700 unjoint_samples 6700 joint_samples 19 [613089, 1048417] processed_samples 6700 unjoint_samples 6700 joint_samples 19 [613089, 1048417] processed_samples 6701 unjoint_samples 6700 joint_samples 19 [730223, 1047012] processed_samples 6701 unjoint_samples 6700 joint_samples 19 [730223, 1047012] [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5596134bb5c0] [h264 @ 0x5617e8995380] mmco: unref short failure mmco: unref short failure [h264 @ 0x5617e8995380] [h264 @ 0x5596134bb5c0] mmco: unref short failure mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x559618471cc0] mmco: unref short failure [h264 @ 0x559618471cc0] mmco: unref short failure [h264 @ 0x5617e8298040] mmco: unref short failure [h264 @ 0x5617e8298040] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure processed_samples 6800 unjoint_samples 6800 joint_samples 19 [1047679, 427954] processed_samples 6800 unjoint_samples 6800 joint_samples 19 [1047679, 427954] processed_samples 6800 unjoint_samples 6800 joint_samples 20 [139029, 1047254] processed_samples 6800 unjoint_samples 6800 joint_samples 20 [139029, 1047254] processed_samples 6800 unjoint_samples 6800 joint_samples 19 [762694, 1047684] processed_samples 6800 unjoint_samples 6800 joint_samples 19 [762694, 1047684] processed_samples 6800 unjoint_samples 6800 joint_samples 19 [794001, 1046860] processed_samples 6800 unjoint_samples 6800 joint_samples 19 [794001, 1046860] processed_samples 6800 unjoint_samples 6800 joint_samples 19 [869170, 1046514] processed_samples 6800 unjoint_samples 6800 joint_samples 19 [869170, 1046514] [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure processed_samples 6801 unjoint_samples 6800 joint_samples 20 [988723, 204618] processed_samples 6801 unjoint_samples 6800 joint_samples 20 [988723, 204618] processed_samples 6800 unjoint_samples 6800 joint_samples 19 [897381, 1048417] processed_samples 6800 unjoint_samples 6800 joint_samples 19 [897381, 1048417] processed_samples 6801 unjoint_samples 6800 joint_samples 20 [1043934, 857046] processed_samples 6801 unjoint_samples 6800 joint_samples 20 [1043934, 857046] [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ee68bb40] mmco: unref short failure [h264 @ 0x5617ee68bb40] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x5617ed3f3040] mmco: unref short failure processed_samples 6900 unjoint_samples 6900 joint_samples 20 [1038289, 192432] processed_samples 6900 unjoint_samples 6900 joint_samples 20 [1038289, 192432] processed_samples 6900 unjoint_samples 6900 joint_samples 20 [56132, 1046860] processed_samples 6900 unjoint_samples 6900 joint_samples 20 [1047601, 21549] processed_samples 6900 unjoint_samples 6900 joint_samples 20 [56132, 1046860] processed_samples 6900 unjoint_samples 6900 joint_samples 20 [1047601, 21549] processed_samples 6900 unjoint_samples 6900 joint_samples 20 [1047186, 169242] processed_samples 6900 unjoint_samples 6900 joint_samples 20 [1047186, 169242] processed_samples 6900 unjoint_samples 6900 joint_samples 20 [441525, 1047254] processed_samples 6900 unjoint_samples 6900 joint_samples 20 [441525, 1047254] processed_samples 6900 unjoint_samples 6900 joint_samples 19 [1047679, 808318] processed_samples 6900 unjoint_samples 6900 joint_samples 19 [1047679, 808318] processed_samples 6901 unjoint_samples 6900 joint_samples 21 [1046536, 70545] processed_samples 6901 unjoint_samples 6900 joint_samples 21 [1046536, 70545] processed_samples 6901 unjoint_samples 6900 joint_samples 20 [988723, 533309] processed_samples 6901 unjoint_samples 6900 joint_samples 20 [988723, 533309] [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x5617e803bd40] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure processed_samples 7000 unjoint_samples 7000 joint_samples 20 [1047679, 99058] processed_samples 7000 unjoint_samples 7000 joint_samples 20 [1047679, 99058] processed_samples 7000 unjoint_samples 7000 joint_samples 20 [298973, 1046860] processed_samples 7000 unjoint_samples 7000 joint_samples 20 [298973, 1046860] processed_samples 7000 unjoint_samples 7000 joint_samples 20 [1038289, 653866] processed_samples 7000 unjoint_samples 7000 joint_samples 20 [1047186, 508288] processed_samples 7000 unjoint_samples 7000 joint_samples 20 [1038289, 653866] processed_samples 7000 unjoint_samples 7000 joint_samples 20 [1047186, 508288] processed_samples 7001 unjoint_samples 7000 joint_samples 20 [988723, 865432] processed_samples 7000 unjoint_samples 7000 joint_samples 20 [1047601, 378321] processed_samples 7000 unjoint_samples 7000 joint_samples 20 [1047601, 378321] processed_samples 7001 unjoint_samples 7000 joint_samples 20 [988723, 865432] processed_samples 7001 unjoint_samples 7000 joint_samples 21 [1046536, 365927] processed_samples 7001 unjoint_samples 7000 joint_samples 21 [1046536, 365927] processed_samples 7000 unjoint_samples 7000 joint_samples 20 [800208, 1047254] processed_samples 7000 unjoint_samples 7000 joint_samples 20 [800208, 1047254] [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [mov,mp4,m4a,3gp,3g2,mj2 @ 0x5596168a57c0] stream 1, offset 0x14000d8: partial file [mov,mp4,m4a,3gp,3g2,mj2 @ 0x5617eee852c0] stream 1, offset 0x14000d8: partial file [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure processed_samples 7100 unjoint_samples 7100 joint_samples 21 [52260, 1047429] processed_samples 7100 unjoint_samples 7100 joint_samples 21 [52260, 1047429] processed_samples 7101 unjoint_samples 7100 joint_samples 21 [1046536, 665022] processed_samples 7101 unjoint_samples 7100 joint_samples 21 [1046536, 665022] [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure processed_samples 7100 unjoint_samples 7100 joint_samples 20 [1047186, 909202] processed_samples 7100 unjoint_samples 7100 joint_samples 20 [1047186, 909202] processed_samples 7100 unjoint_samples 7100 joint_samples 20 [1047679, 422289] processed_samples 7100 unjoint_samples 7100 joint_samples 20 [1047679, 422289] processed_samples 7100 unjoint_samples 7100 joint_samples 20 [1047601, 675875] processed_samples 7100 unjoint_samples 7100 joint_samples 20 [1047601, 675875] processed_samples 7100 unjoint_samples 7100 joint_samples 20 [589508, 1046860] processed_samples 7100 unjoint_samples 7100 joint_samples 20 [589508, 1046860] processed_samples 7100 unjoint_samples 7100 joint_samples 21 [988611, 280527] processed_samples 7100 unjoint_samples 7100 joint_samples 21 [988611, 280527] processed_samples 7102 unjoint_samples 7100 joint_samples 20 [1032983, 1033572] processed_samples 7102 unjoint_samples 7100 joint_samples 20 [1032983, 1033572] [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure processed_samples 7200 unjoint_samples 7200 joint_samples 21 [338334, 1047429] processed_samples 7200 unjoint_samples 7200 joint_samples 21 [338334, 1047429] processed_samples 7200 unjoint_samples 7200 joint_samples 21 [128606, 1041708] processed_samples 7200 unjoint_samples 7200 joint_samples 21 [128606, 1041708] processed_samples 7200 unjoint_samples 7200 joint_samples 21 [1047186, 161716] processed_samples 7200 unjoint_samples 7200 joint_samples 21 [1047186, 161716] [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure processed_samples 7200 unjoint_samples 7200 joint_samples 20 [1047679, 691396] processed_samples 7200 unjoint_samples 7200 joint_samples 20 [1047679, 691396] processed_samples 7200 unjoint_samples 7200 joint_samples 20 [894377, 1046860] processed_samples 7200 unjoint_samples 7200 joint_samples 20 [894377, 1046860] processed_samples 7200 unjoint_samples 7200 joint_samples 21 [988611, 672873] processed_samples 7200 unjoint_samples 7200 joint_samples 21 [988611, 672873] processed_samples 7202 unjoint_samples 7200 joint_samples 21 [267286, 1047190] processed_samples 7202 unjoint_samples 7200 joint_samples 21 [267286, 1047190] [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure processed_samples 7201 unjoint_samples 7200 joint_samples 21 [1046536, 890317] processed_samples 7201 unjoint_samples 7200 joint_samples 21 [1046536, 890317] [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e8e0d640] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure processed_samples 7300 unjoint_samples 7300 joint_samples 21 [180830, 1034971] processed_samples 7300 unjoint_samples 7300 joint_samples 21 [180830, 1034971] processed_samples 7300 unjoint_samples 7300 joint_samples 21 [470274, 1041708] processed_samples 7300 unjoint_samples 7300 joint_samples 21 [470274, 1041708] processed_samples 7300 unjoint_samples 7300 joint_samples 21 [198994, 1046860] processed_samples 7300 unjoint_samples 7300 joint_samples 21 [198994, 1046860] processed_samples 7300 unjoint_samples 7300 joint_samples 21 [666188, 1047429] processed_samples 7300 unjoint_samples 7300 joint_samples 21 [666188, 1047429] processed_samples 7300 unjoint_samples 7300 joint_samples 21 [1047186, 475799] processed_samples 7300 unjoint_samples 7300 joint_samples 21 [1047186, 475799] processed_samples 7301 unjoint_samples 7300 joint_samples 22 [181670, 1035814] processed_samples 7301 unjoint_samples 7300 joint_samples 22 [181670, 1035814] processed_samples 7302 unjoint_samples 7300 joint_samples 21 [589092, 1047190] processed_samples 7302 unjoint_samples 7300 joint_samples 21 [589092, 1047190] processed_samples 7300 unjoint_samples 7300 joint_samples 21 [988611, 930059] processed_samples 7300 unjoint_samples 7300 joint_samples 21 [988611, 930059] [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure processed_samples 7400 unjoint_samples 7400 joint_samples 21 [634311, 1034971] processed_samples 7400 unjoint_samples 7400 joint_samples 21 [634311, 1034971] processed_samples 7400 unjoint_samples 7400 joint_samples 21 [615006, 1046860] processed_samples 7400 unjoint_samples 7400 joint_samples 21 [1047186, 816412] processed_samples 7400 unjoint_samples 7400 joint_samples 21 [1047186, 816412] processed_samples 7400 unjoint_samples 7400 joint_samples 21 [615006, 1046860] processed_samples 7400 unjoint_samples 7400 joint_samples 21 [742158, 1041708] processed_samples 7400 unjoint_samples 7400 joint_samples 21 [742158, 1041708] [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure processed_samples 7400 unjoint_samples 7400 joint_samples 22 [1044567, 289167] processed_samples 7400 unjoint_samples 7400 joint_samples 22 [1044567, 289167] processed_samples 7400 unjoint_samples 7400 joint_samples 21 [941675, 1047429] processed_samples 7401 unjoint_samples 7400 joint_samples 22 [545984, 1035814] processed_samples 7401 unjoint_samples 7400 joint_samples 22 [545984, 1035814] processed_samples 7400 unjoint_samples 7400 joint_samples 21 [941675, 1047429] processed_samples 7402 unjoint_samples 7400 joint_samples 21 [889077, 1047190] processed_samples 7402 unjoint_samples 7400 joint_samples 21 [889077, 1047190] [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure processed_samples 7500 unjoint_samples 7500 joint_samples 21 [899835, 1046860] processed_samples 7500 unjoint_samples 7500 joint_samples 21 [899835, 1046860] processed_samples 7500 unjoint_samples 7500 joint_samples 22 [115745, 1020989] processed_samples 7500 unjoint_samples 7500 joint_samples 22 [115745, 1020989] processed_samples 7500 unjoint_samples 7500 joint_samples 22 [10939, 1046596] processed_samples 7500 unjoint_samples 7500 joint_samples 22 [10939, 1046596] processed_samples 7500 unjoint_samples 7500 joint_samples 22 [1046261, 270940] processed_samples 7500 unjoint_samples 7500 joint_samples 22 [1046261, 270940] processed_samples 7500 unjoint_samples 7500 joint_samples 22 [1044567, 628248] processed_samples 7500 unjoint_samples 7500 joint_samples 22 [1044567, 628248] processed_samples 7500 unjoint_samples 7500 joint_samples 21 [941172, 1034971] processed_samples 7502 unjoint_samples 7500 joint_samples 22 [121268, 1047451] processed_samples 7500 unjoint_samples 7500 joint_samples 21 [941172, 1034971] processed_samples 7502 unjoint_samples 7500 joint_samples 22 [121268, 1047451] processed_samples 7501 unjoint_samples 7500 joint_samples 22 [838882, 1035814] processed_samples 7501 unjoint_samples 7500 joint_samples 22 [838882, 1035814] [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617ea25da40] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure processed_samples 7600 unjoint_samples 7600 joint_samples 22 [430200, 1020989] processed_samples 7600 unjoint_samples 7600 joint_samples 22 [430200, 1020989] processed_samples 7600 unjoint_samples 7600 joint_samples 22 [318923, 1046596] processed_samples 7600 unjoint_samples 7600 joint_samples 22 [985453, 375565] processed_samples 7600 unjoint_samples 7600 joint_samples 22 [318923, 1046596] processed_samples 7600 unjoint_samples 7600 joint_samples 22 [985453, 375565] processed_samples 7601 unjoint_samples 7600 joint_samples 23 [97298, 1045765] processed_samples 7601 unjoint_samples 7600 joint_samples 23 [97298, 1045765] processed_samples 7600 unjoint_samples 7600 joint_samples 22 [1046444, 136979] processed_samples 7600 unjoint_samples 7600 joint_samples 22 [1046444, 136979] processed_samples 7600 unjoint_samples 7600 joint_samples 22 [1046261, 597576] processed_samples 7600 unjoint_samples 7600 joint_samples 22 [1046261, 597576] processed_samples 7602 unjoint_samples 7600 joint_samples 22 [404443, 1047451] processed_samples 7602 unjoint_samples 7600 joint_samples 22 [404443, 1047451] processed_samples 7600 unjoint_samples 7600 joint_samples 22 [1044567, 963021] processed_samples 7600 unjoint_samples 7600 joint_samples 22 [1044567, 963021] [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559616fdd040] mmco: unref short failure [h264 @ 0x559616fdd040] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x55961d500c40] [h264 @ 0x5617ebde1580] mmco: unref short failure mmco: unref short failure processed_samples 7700 unjoint_samples 7700 joint_samples 22 [1046261, 854604] processed_samples 7700 unjoint_samples 7700 joint_samples 22 [985453, 714324] processed_samples 7700 unjoint_samples 7700 joint_samples 22 [985453, 714324] processed_samples 7700 unjoint_samples 7700 joint_samples 23 [355180, 985837] processed_samples 7700 unjoint_samples 7700 joint_samples 23 [355180, 985837] processed_samples 7700 unjoint_samples 7700 joint_samples 22 [1046261, 854604] processed_samples 7701 unjoint_samples 7700 joint_samples 23 [325221, 1045765] processed_samples 7701 unjoint_samples 7700 joint_samples 23 [325221, 1045765] processed_samples 7700 unjoint_samples 7700 joint_samples 22 [605890, 1046596] processed_samples 7700 unjoint_samples 7700 joint_samples 22 [605890, 1046596] processed_samples 7700 unjoint_samples 7700 joint_samples 22 [1046444, 430256] processed_samples 7700 unjoint_samples 7700 joint_samples 22 [1046444, 430256] processed_samples 7700 unjoint_samples 7700 joint_samples 22 [714904, 1020989] processed_samples 7700 unjoint_samples 7700 joint_samples 22 [714904, 1020989] processed_samples 7702 unjoint_samples 7700 joint_samples 22 [796922, 1047451] processed_samples 7702 unjoint_samples 7700 joint_samples 22 [796922, 1047451] [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ea25da40] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961adc42c0] mmco: unref short failure [h264 @ 0x5617ea25da40] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure processed_samples 7800 unjoint_samples 7800 joint_samples 23 [100643, 1048429] processed_samples 7800 unjoint_samples 7800 joint_samples 23 [100643, 1048429] [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure processed_samples 7800 unjoint_samples 7800 joint_samples 22 [1027093, 1025209] processed_samples 7800 unjoint_samples 7800 joint_samples 22 [1012235, 1013352] processed_samples 7800 unjoint_samples 7800 joint_samples 22 [1046444, 748638] processed_samples 7800 unjoint_samples 7800 joint_samples 22 [1012235, 1013352] processed_samples 7800 unjoint_samples 7800 joint_samples 22 [1027093, 1025209] processed_samples 7800 unjoint_samples 7800 joint_samples 22 [1046444, 748638] processed_samples 7801 unjoint_samples 7800 joint_samples 23 [610952, 1045765] processed_samples 7801 unjoint_samples 7800 joint_samples 23 [610952, 1045765] processed_samples 7802 unjoint_samples 7800 joint_samples 23 [56650, 1047451] processed_samples 7802 unjoint_samples 7800 joint_samples 23 [56650, 1047451] processed_samples 7800 unjoint_samples 7800 joint_samples 23 [819206, 985837] processed_samples 7800 unjoint_samples 7800 joint_samples 23 [819206, 985837] processed_samples 7800 unjoint_samples 7800 joint_samples 22 [867825, 1046596] processed_samples 7800 unjoint_samples 7800 joint_samples 22 [867825, 1046596] [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure processed_samples 7900 unjoint_samples 7900 joint_samples 23 [327859, 1048429] processed_samples 7900 unjoint_samples 7900 joint_samples 23 [327859, 1048429] [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure processed_samples 7900 unjoint_samples 7900 joint_samples 23 [1046535, 152417] processed_samples 7900 unjoint_samples 7900 joint_samples 23 [1046535, 152417] processed_samples 7900 unjoint_samples 7900 joint_samples 23 [1043101, 187740] processed_samples 7900 unjoint_samples 7900 joint_samples 23 [1047018, 97414] processed_samples 7900 unjoint_samples 7900 joint_samples 23 [1047018, 97414] processed_samples 7900 unjoint_samples 7900 joint_samples 23 [1043101, 187740] processed_samples 7900 unjoint_samples 7900 joint_samples 23 [1045331, 379873] processed_samples 7900 unjoint_samples 7900 joint_samples 23 [1045331, 379873] processed_samples 7900 unjoint_samples 7900 joint_samples 24 [1000627, 238985] processed_samples 7900 unjoint_samples 7900 joint_samples 24 [1000627, 238985] processed_samples 7901 unjoint_samples 7900 joint_samples 23 [924553, 1045765] processed_samples 7901 unjoint_samples 7900 joint_samples 23 [924553, 1045765] processed_samples 7902 unjoint_samples 7900 joint_samples 23 [413988, 1047451] processed_samples 7902 unjoint_samples 7900 joint_samples 23 [413988, 1047451] [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure processed_samples 8000 unjoint_samples 8000 joint_samples 23 [1046535, 452463] processed_samples 8000 unjoint_samples 8000 joint_samples 23 [1046535, 452463] processed_samples 8000 unjoint_samples 8000 joint_samples 24 [1000627, 508155] processed_samples 8000 unjoint_samples 8000 joint_samples 24 [1000627, 508155] processed_samples 8000 unjoint_samples 8000 joint_samples 23 [1047018, 425116] processed_samples 8000 unjoint_samples 8000 joint_samples 23 [1043101, 535220] processed_samples 8000 unjoint_samples 8000 joint_samples 23 [1043101, 535220] processed_samples 8000 unjoint_samples 8000 joint_samples 23 [1045331, 781552] processed_samples 8000 unjoint_samples 8000 joint_samples 23 [1047018, 425116] processed_samples 8000 unjoint_samples 8000 joint_samples 23 [1045331, 781552] [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure processed_samples 8001 unjoint_samples 8000 joint_samples 24 [1047232, 188285] processed_samples 8001 unjoint_samples 8000 joint_samples 24 [1047232, 188285] processed_samples 8000 unjoint_samples 8000 joint_samples 23 [615055, 1048429] processed_samples 8000 unjoint_samples 8000 joint_samples 23 [615055, 1048429] processed_samples 8002 unjoint_samples 8000 joint_samples 23 [728292, 1047451]processed_samples 8002 unjoint_samples 8000 joint_samples 23 [728292, 1047451] [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ea25da40] mmco: unref short failure [h264 @ 0x5617ea25da40] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ea25da40] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure processed_samples 8100 unjoint_samples 8100 joint_samples 24 [171611, 1045256] processed_samples 8100 unjoint_samples 8100 joint_samples 24 [171611, 1045256] processed_samples 8100 unjoint_samples 8100 joint_samples 23 [1046535, 846955] processed_samples 8100 unjoint_samples 8100 joint_samples 23 [1046535, 846955] processed_samples 8100 unjoint_samples 8100 joint_samples 23 [893108, 1048429] processed_samples 8100 unjoint_samples 8100 joint_samples 23 [893108, 1048429] processed_samples 8100 unjoint_samples 8100 joint_samples 23 [1043101, 905563] processed_samples 8100 unjoint_samples 8100 joint_samples 23 [1043101, 905563] [h264 @ 0x5596172068c0] mmco: unref short failure processed_samples 8100 unjoint_samples 8100 joint_samples 23 [1047018, 989840] processed_samples 8101 unjoint_samples 8100 joint_samples 24 [1047232, 634433] [h264 @ 0x55961b920980] mmco: unref short failure processed_samples 8101 unjoint_samples 8100 joint_samples 24 [1047232, 634433] [h264 @ 0x5617e9213b80] mmco: unref short failure processed_samples 8100 unjoint_samples 8100 joint_samples 23 [1047018, 989840] processed_samples 8100 unjoint_samples 8100 joint_samples 24 [1000627, 815362] processed_samples 8100 unjoint_samples 8100 joint_samples 24 [1000627, 815362] [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure processed_samples 8102 unjoint_samples 8100 joint_samples 24 [1030914, 55331] [h264 @ 0x5617eeca6a40] mmco: unref short failure processed_samples 8102 unjoint_samples 8100 joint_samples 24 [1030914, 55331] [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x55961adc42c0] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure processed_samples 8200 unjoint_samples 8200 joint_samples 24 [1028669, 269942] processed_samples 8200 unjoint_samples 8200 joint_samples 24 [1028669, 269942] processed_samples 8200 unjoint_samples 8200 joint_samples 24 [437099, 1045256] processed_samples 8200 unjoint_samples 8200 joint_samples 24 [437099, 1045256] processed_samples 8200 unjoint_samples 8200 joint_samples 24 [1046535, 62427] [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure processed_samples 8200 unjoint_samples 8200 joint_samples 24 [1046535, 62427] processed_samples 8200 unjoint_samples 8200 joint_samples 25 [41837, 1047353] processed_samples 8200 unjoint_samples 8200 joint_samples 25 [41837, 1047353] [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure processed_samples 8200 unjoint_samples 8200 joint_samples 24 [260373, 1034218] processed_samples 8200 unjoint_samples 8200 joint_samples 24 [260373, 1034218] processed_samples 8200 unjoint_samples 8200 joint_samples 24 [301705, 1038963] processed_samples 8200 unjoint_samples 8200 joint_samples 24 [301705, 1038963] processed_samples 8201 unjoint_samples 8200 joint_samples 24 [1047232, 840437] processed_samples 8201 unjoint_samples 8200 joint_samples 24 [1047232, 840437] [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure processed_samples 8202 unjoint_samples 8200 joint_samples 24 [1030914, 436780] processed_samples 8202 unjoint_samples 8200 joint_samples 24 [1030914, 436780] [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure processed_samples 8300 unjoint_samples 8300 joint_samples 24 [1046535, 387861] processed_samples 8300 unjoint_samples 8300 joint_samples 24 [1046535, 387861] processed_samples 8300 unjoint_samples 8300 joint_samples 24 [1028669, 577617] processed_samples 8300 unjoint_samples 8300 joint_samples 24 [520303, 1034218] processed_samples 8300 unjoint_samples 8300 joint_samples 24 [1028669, 577617] processed_samples 8300 unjoint_samples 8300 joint_samples 24 [520303, 1034218] processed_samples 8300 unjoint_samples 8300 joint_samples 25 [368705, 1047353] processed_samples 8300 unjoint_samples 8300 joint_samples 25 [368705, 1047353] processed_samples 8300 unjoint_samples 8300 joint_samples 24 [726225, 1038963] [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure processed_samples 8301 unjoint_samples 8300 joint_samples 25 [1047232, 88012] processed_samples 8301 unjoint_samples 8300 joint_samples 25 [1047232, 88012] processed_samples 8300 unjoint_samples 8300 joint_samples 24 [726225, 1038963] processed_samples 8300 unjoint_samples 8300 joint_samples 24 [708538, 1045256] processed_samples 8300 unjoint_samples 8300 joint_samples 24 [708538, 1045256] [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure processed_samples 8302 unjoint_samples 8300 joint_samples 24 [1030914, 758017] processed_samples 8302 unjoint_samples 8300 joint_samples 24 [1030914, 758017] [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559618471cc0] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ea25da40] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure processed_samples 8400 unjoint_samples 8400 joint_samples 24 [1046535, 683687] processed_samples 8400 unjoint_samples 8400 joint_samples 24 [1028669, 884681] processed_samples 8401 unjoint_samples 8400 joint_samples 25 [1047232, 412944] processed_samples 8400 unjoint_samples 8400 joint_samples 24 [838891, 1034218] processed_samples 8400 unjoint_samples 8400 joint_samples 25 [697009, 1047353] processed_samples 8400 unjoint_samples 8400 joint_samples 24 [1046535, 683687] processed_samples 8400 unjoint_samples 8400 joint_samples 24 [1037118, 1038963] processed_samples 8400 unjoint_samples 8400 joint_samples 24 [1023984, 1045256] processed_samples 8401 unjoint_samples 8400 joint_samples 25 [1047232, 412944] processed_samples 8400 unjoint_samples 8400 joint_samples 24 [1028669, 884681] processed_samples 8400 unjoint_samples 8400 joint_samples 24 [838891, 1034218] processed_samples 8402 unjoint_samples 8400 joint_samples 24 [1030914, 1000267] processed_samples 8400 unjoint_samples 8400 joint_samples 25 [697009, 1047353] processed_samples 8400 unjoint_samples 8400 joint_samples 24 [1037118, 1038963] processed_samples 8400 unjoint_samples 8400 joint_samples 24 [1023984, 1045256] processed_samples 8402 unjoint_samples 8400 joint_samples 24 [1030914, 1000267] [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e9262740] mmco: unref short failure [h264 @ 0x5617e9262740] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617e9262740] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure processed_samples 8500 unjoint_samples 8500 joint_samples 25 [40645, 1047363] processed_samples 8500 unjoint_samples 8500 joint_samples 25 [40645, 1047363] processed_samples 8500 unjoint_samples 8500 joint_samples 25 [163730, 989419] processed_samples 8500 unjoint_samples 8500 joint_samples 25 [1029244, 369655] processed_samples 8500 unjoint_samples 8500 joint_samples 25 [1029244, 369655] processed_samples 8500 unjoint_samples 8500 joint_samples 25 [163730, 989419] processed_samples 8500 unjoint_samples 8500 joint_samples 25 [1035024, 409201] processed_samples 8500 unjoint_samples 8500 joint_samples 25 [1035024, 409201] processed_samples 8500 unjoint_samples 8500 joint_samples 25 [318196, 1046562] processed_samples 8500 unjoint_samples 8500 joint_samples 25 [318196, 1046562] [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure processed_samples 8501 unjoint_samples 8500 joint_samples 25 [1047232, 814182] processed_samples 8501 unjoint_samples 8500 joint_samples 25 [1047232, 814182] processed_samples 8500 unjoint_samples 8500 joint_samples 25 [954132, 1047353] processed_samples 8500 unjoint_samples 8500 joint_samples 25 [954132, 1047353] processed_samples 8502 unjoint_samples 8500 joint_samples 25 [205293, 1040286] processed_samples 8502 unjoint_samples 8500 joint_samples 25 [205293, 1040286] [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure processed_samples 8600 unjoint_samples 8600 joint_samples 25 [332617, 1047363] processed_samples 8600 unjoint_samples 8600 joint_samples 25 [332617, 1047363] processed_samples 8600 unjoint_samples 8600 joint_samples 25 [1029244, 738726] processed_samples 8600 unjoint_samples 8600 joint_samples 25 [1029244, 738726] processed_samples 8600 unjoint_samples 8600 joint_samples 26 [1047093, 152028] processed_samples 8600 unjoint_samples 8600 joint_samples 26 [1047093, 152028] processed_samples 8600 unjoint_samples 8600 joint_samples 25 [646891, 1046562] processed_samples 8600 unjoint_samples 8600 joint_samples 25 [646891, 1046562] processed_samples 8601 unjoint_samples 8600 joint_samples 26 [39430, 1046832] processed_samples 8601 unjoint_samples 8600 joint_samples 26 [39430, 1046832] processed_samples 8600 unjoint_samples 8600 joint_samples 25 [1035024, 734063] processed_samples 8600 unjoint_samples 8600 joint_samples 25 [1035024, 734063] processed_samples 8600 unjoint_samples 8600 joint_samples 25 [414512, 989419] processed_samples 8600 unjoint_samples 8600 joint_samples 25 [414512, 989419] processed_samples 8602 unjoint_samples 8600 joint_samples 25 [556858, 1040286] processed_samples 8602 unjoint_samples 8600 joint_samples 25 [556858, 1040286] [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure processed_samples 8700 unjoint_samples 8700 joint_samples 25 [725288, 989419] processed_samples 8700 unjoint_samples 8700 joint_samples 25 [725288, 989419] processed_samples 8700 unjoint_samples 8700 joint_samples 25 [620873, 1047363] processed_samples 8700 unjoint_samples 8700 joint_samples 25 [620873, 1047363] processed_samples 8700 unjoint_samples 8700 joint_samples 26 [94643, 1046172] processed_samples 8700 unjoint_samples 8700 joint_samples 26 [94643, 1046172] processed_samples 8701 unjoint_samples 8700 joint_samples 26 [288675, 1046832] processed_samples 8701 unjoint_samples 8700 joint_samples 26 [288675, 1046832] processed_samples 8700 unjoint_samples 8700 joint_samples 26 [1047093, 488874] processed_samples 8700 unjoint_samples 8700 joint_samples 26 [1047093, 488874] processed_samples 8700 unjoint_samples 8700 joint_samples 25 [926171, 1046562] processed_samples 8700 unjoint_samples 8700 joint_samples 25 [926171, 1046562] [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure processed_samples 8700 unjoint_samples 8700 joint_samples 25 [1029244, 1012301] processed_samples 8700 unjoint_samples 8700 joint_samples 25 [1029244, 1012301] processed_samples 8702 unjoint_samples 8700 joint_samples 25 [926697, 1040286] processed_samples 8702 unjoint_samples 8700 joint_samples 25 [926697, 1040286] [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure processed_samples 8800 unjoint_samples 8800 joint_samples 26 [1046312, 10469] processed_samples 8800 unjoint_samples 8800 joint_samples 26 [1046312, 10469] processed_samples 8800 unjoint_samples 8800 joint_samples 26 [1046728, 163560] processed_samples 8800 unjoint_samples 8800 joint_samples 26 [1046728, 163560] processed_samples 8800 unjoint_samples 8800 joint_samples 26 [467211, 1046172] processed_samples 8800 unjoint_samples 8800 joint_samples 26 [467211, 1046172] processed_samples 8800 unjoint_samples 8800 joint_samples 26 [328007, 1034287] processed_samples 8800 unjoint_samples 8800 joint_samples 26 [328007, 1034287] processed_samples 8800 unjoint_samples 8800 joint_samples 25 [862812, 1047363] processed_samples 8800 unjoint_samples 8800 joint_samples 25 [862812, 1047363] processed_samples 8800 unjoint_samples 8800 joint_samples 26 [1047093, 771328] processed_samples 8800 unjoint_samples 8800 joint_samples 26 [1047093, 771328] processed_samples 8802 unjoint_samples 8800 joint_samples 26 [185183, 1044328] processed_samples 8802 unjoint_samples 8800 joint_samples 26 [185183, 1044328] processed_samples 8801 unjoint_samples 8800 joint_samples 26 [545868, 1046832] processed_samples 8801 unjoint_samples 8800 joint_samples 26 [545868, 1046832] [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5596177086c0] mmco: unref short failure [h264 @ 0x5596177086c0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure processed_samples 8900 unjoint_samples 8900 joint_samples 26 [55412, 1047363] processed_samples 8900 unjoint_samples 8900 joint_samples 26 [55412, 1047363] processed_samples 8900 unjoint_samples 8900 joint_samples 26 [1046312, 364291] processed_samples 8900 unjoint_samples 8900 joint_samples 26 [1046312, 364291] processed_samples 8900 unjoint_samples 8900 joint_samples 26 [592882, 1034287] processed_samples 8900 unjoint_samples 8900 joint_samples 26 [592882, 1034287] processed_samples 8900 unjoint_samples 8900 joint_samples 27 [26892, 1046376] processed_samples 8900 unjoint_samples 8900 joint_samples 27 [26892, 1046376] processed_samples 8900 unjoint_samples 8900 joint_samples 26 [1046728, 461784] processed_samples 8900 unjoint_samples 8900 joint_samples 26 [1046728, 461784] [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure processed_samples 8901 unjoint_samples 8900 joint_samples 26 [893489, 1046832] processed_samples 8901 unjoint_samples 8900 joint_samples 26 [893489, 1046832] processed_samples 8900 unjoint_samples 8900 joint_samples 26 [873062, 1046172] processed_samples 8900 unjoint_samples 8900 joint_samples 26 [873062, 1046172] processed_samples 8902 unjoint_samples 8900 joint_samples 26 [409484, 1044328] processed_samples 8902 unjoint_samples 8900 joint_samples 26 [409484, 1044328] [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617ed3f3040] mmco: unref short failure [h264 @ 0x5617ed3f3040] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x559618a89040] mmco: unref short failure [h264 @ 0x559618a89040] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure processed_samples 9000 unjoint_samples 9000 joint_samples 27 [85414, 1046516] processed_samples 9000 unjoint_samples 9000 joint_samples 27 [85414, 1046516] processed_samples 9000 unjoint_samples 9000 joint_samples 26 [1046312, 643866] processed_samples 9000 unjoint_samples 9000 joint_samples 26 [1046312, 643866] processed_samples 9000 unjoint_samples 9000 joint_samples 26 [316896, 1047363] processed_samples 9000 unjoint_samples 9000 joint_samples 26 [316896, 1047363] processed_samples 9001 unjoint_samples 9000 joint_samples 27 [150215, 1046832] processed_samples 9001 unjoint_samples 9000 joint_samples 27 [150215, 1046832] processed_samples 9000 unjoint_samples 9000 joint_samples 27 [330831, 1046376] processed_samples 9000 unjoint_samples 9000 joint_samples 27 [330831, 1046376] processed_samples 9000 unjoint_samples 9000 joint_samples 26 [870320, 1034287] processed_samples 9000 unjoint_samples 9000 joint_samples 26 [870320, 1034287] processed_samples 9000 unjoint_samples 9000 joint_samples 26 [1046728, 712998] processed_samples 9000 unjoint_samples 9000 joint_samples 26 [1046728, 712998] processed_samples 9002 unjoint_samples 9000 joint_samples 26 [809251, 1044328] processed_samples 9002 unjoint_samples 9000 joint_samples 26 [809251, 1044328] [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559616d35c80] mmco: unref short failure [h264 @ 0x559616d35c80] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure processed_samples 9100 unjoint_samples 9100 joint_samples 26 [619599, 1047363] processed_samples 9100 unjoint_samples 9100 joint_samples 26 [619599, 1047363] processed_samples 9100 unjoint_samples 9100 joint_samples 27 [114069, 1046690] processed_samples 9100 unjoint_samples 9100 joint_samples 27 [114069, 1046690] processed_samples 9100 unjoint_samples 9100 joint_samples 27 [398883, 1046516] processed_samples 9101 unjoint_samples 9100 joint_samples 27 [421384, 1046832] processed_samples 9101 unjoint_samples 9100 joint_samples 27 [421384, 1046832] processed_samples 9100 unjoint_samples 9100 joint_samples 27 [398883, 1046516] processed_samples 9100 unjoint_samples 9100 joint_samples 27 [575820, 1046376] processed_samples 9100 unjoint_samples 9100 joint_samples 27 [575820, 1046376] processed_samples 9100 unjoint_samples 9100 joint_samples 26 [1046728, 1006998] processed_samples 9100 unjoint_samples 9100 joint_samples 26 [1046728, 1006998] processed_samples 9100 unjoint_samples 9100 joint_samples 26 [1046312, 961612] processed_samples 9100 unjoint_samples 9100 joint_samples 26 [1046312, 961612] processed_samples 9102 unjoint_samples 9100 joint_samples 27 [1032172, 74139] processed_samples 9102 unjoint_samples 9100 joint_samples 27 [1032172, 74139] [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ed3f3040] mmco: unref short failure [h264 @ 0x5617ed3f3040] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure processed_samples 9200 unjoint_samples 9200 joint_samples 26 [875218, 1047363] processed_samples 9200 unjoint_samples 9200 joint_samples 27 [1046365, 333849] [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure processed_samples 9200 unjoint_samples 9200 joint_samples 27 [1046365, 333849] processed_samples 9200 unjoint_samples 9200 joint_samples 27 [212056, 1045452] processed_samples 9200 unjoint_samples 9200 joint_samples 26 [875218, 1047363] processed_samples 9200 unjoint_samples 9200 joint_samples 27 [212056, 1045452] processed_samples 9200 unjoint_samples 9200 joint_samples 27 [449971, 1046690] processed_samples 9200 unjoint_samples 9200 joint_samples 27 [700272, 1046516] processed_samples 9200 unjoint_samples 9200 joint_samples 27 [700272, 1046516] processed_samples 9202 unjoint_samples 9200 joint_samples 27 [1032172, 301612] processed_samples 9200 unjoint_samples 9200 joint_samples 27 [449971, 1046690] processed_samples 9202 unjoint_samples 9200 joint_samples 27 [1032172, 301612] processed_samples 9201 unjoint_samples 9200 joint_samples 27 [713147, 1046832] processed_samples 9200 unjoint_samples 9200 joint_samples 27 [872579, 1046376] [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure processed_samples 9200 unjoint_samples 9200 joint_samples 27 [872579, 1046376] [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure processed_samples 9201 unjoint_samples 9200 joint_samples 27 [713147, 1046832] [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure processed_samples 9300 unjoint_samples 9300 joint_samples 27 [1046365, 770944] processed_samples 9300 unjoint_samples 9300 joint_samples 27 [1047509, 57338] processed_samples 9300 unjoint_samples 9300 joint_samples 28 [155476, 1046376] processed_samples 9300 unjoint_samples 9300 joint_samples 27 [430588, 1045452] processed_samples 9301 unjoint_samples 9300 joint_samples 27 [981505, 1046832] processed_samples 9300 unjoint_samples 9300 joint_samples 27 [1046365, 770944] processed_samples 9300 unjoint_samples 9300 joint_samples 27 [705731, 1046690] processed_samples 9300 unjoint_samples 9300 joint_samples 27 [1047509, 57338] processed_samples 9300 unjoint_samples 9300 joint_samples 27 [997518, 1046516] processed_samples 9300 unjoint_samples 9300 joint_samples 28 [155476, 1046376] processed_samples 9300 unjoint_samples 9300 joint_samples 27 [430588, 1045452] processed_samples 9300 unjoint_samples 9300 joint_samples 27 [705731, 1046690] processed_samples 9302 unjoint_samples 9300 joint_samples 27 [1032172, 671365] processed_samples 9301 unjoint_samples 9300 joint_samples 27 [981505, 1046832] processed_samples 9300 unjoint_samples 9300 joint_samples 27 [997518, 1046516] processed_samples 9302 unjoint_samples 9300 joint_samples 27 [1032172, 671365] [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617e7e34fc0] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617ea25da40] mmco: unref short failure [h264 @ 0x5617ea25da40] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ea25da40] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961342b000] Missing reference picture, default is 65530 [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure processed_samples 9400 unjoint_samples 9400 joint_samples 27 [1047509, 538930] processed_samples 9400 unjoint_samples 9400 joint_samples 27 [1046365, 1047120] processed_samples 9400 unjoint_samples 9400 joint_samples 28 [544258, 1046376] processed_samples 9401 unjoint_samples 9400 joint_samples 28 [1039790, 235095] [h264 @ 0x5617ebfdebc0] mmco: unref short failure processed_samples 9400 unjoint_samples 9400 joint_samples 28 [1042777, 280938] processed_samples 9400 unjoint_samples 9400 joint_samples 27 [681230, 1045452] processed_samples 9400 unjoint_samples 9400 joint_samples 27 [953764, 1046690] [h264 @ 0x5617e829eac0] Missing reference picture, default is 65530 [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure processed_samples 9400 unjoint_samples 9400 joint_samples 27 [1047509, 538930] processed_samples 9400 unjoint_samples 9400 joint_samples 27 [1046365, 1047120] processed_samples 9400 unjoint_samples 9400 joint_samples 28 [544258, 1046376] processed_samples 9401 unjoint_samples 9400 joint_samples 28 [1039790, 235095] processed_samples 9400 unjoint_samples 9400 joint_samples 28 [1042777, 280938] processed_samples 9400 unjoint_samples 9400 joint_samples 27 [681230, 1045452] processed_samples 9402 unjoint_samples 9400 joint_samples 27 [1032172, 1026499] processed_samples 9400 unjoint_samples 9400 joint_samples 27 [953764, 1046690] processed_samples 9402 unjoint_samples 9400 joint_samples 27 [1032172, 1026499] [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617e9262740] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617e9262740] mmco: unref short failure [h264 @ 0x5617e9262740] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure processed_samples 9500 unjoint_samples 9500 joint_samples 28 [404485, 1047120] processed_samples 9500 unjoint_samples 9500 joint_samples 28 [404485, 1047120] processed_samples 9500 unjoint_samples 9500 joint_samples 28 [201099, 1046690] processed_samples 9500 unjoint_samples 9500 joint_samples 28 [201099, 1046690] processed_samples 9500 unjoint_samples 9500 joint_samples 28 [1042777, 618561] processed_samples 9500 unjoint_samples 9500 joint_samples 28 [1042777, 618561] processed_samples 9501 unjoint_samples 9500 joint_samples 28 [1039790, 481479] processed_samples 9501 unjoint_samples 9500 joint_samples 28 [1039790, 481479] processed_samples 9500 unjoint_samples 9500 joint_samples 28 [925250, 1046376] [h264 @ 0x559618461980] mmco: unref short failure processed_samples 9500 unjoint_samples 9500 joint_samples 28 [925250, 1046376] [h264 @ 0x5617f0b8e140] mmco: unref short failure processed_samples 9500 unjoint_samples 9500 joint_samples 27 [1047509, 852241] processed_samples 9500 unjoint_samples 9500 joint_samples 27 [1047509, 852241] processed_samples 9502 unjoint_samples 9500 joint_samples 28 [1045090, 356434] processed_samples 9502 unjoint_samples 9500 joint_samples 28 [1045090, 356434] processed_samples 9500 unjoint_samples 9500 joint_samples 27 [945960, 1045452] processed_samples 9500 unjoint_samples 9500 joint_samples 27 [945960, 1045452] [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure processed_samples 9600 unjoint_samples 9600 joint_samples 28 [127384, 1046571] processed_samples 9600 unjoint_samples 9600 joint_samples 28 [127384, 1046571] processed_samples 9600 unjoint_samples 9600 joint_samples 28 [503812, 1046690] processed_samples 9600 unjoint_samples 9600 joint_samples 28 [1047189, 172518] processed_samples 9600 unjoint_samples 9600 joint_samples 28 [1047189, 172518] processed_samples 9600 unjoint_samples 9600 joint_samples 29 [1034198, 306237] processed_samples 9600 unjoint_samples 9600 joint_samples 28 [702739, 1047120] processed_samples 9600 unjoint_samples 9600 joint_samples 28 [503812, 1046690] processed_samples 9600 unjoint_samples 9600 joint_samples 29 [1034198, 306237] processed_samples 9600 unjoint_samples 9600 joint_samples 28 [702739, 1047120] processed_samples 9600 unjoint_samples 9600 joint_samples 28 [1042777, 863365] [h264 @ 0x55961d016580] mmco: unref short failure processed_samples 9600 unjoint_samples 9600 joint_samples 28 [1042777, 863365] [h264 @ 0x5617e9426340] mmco: unref short failure processed_samples 9601 unjoint_samples 9600 joint_samples 28 [1039790, 751048] processed_samples 9602 unjoint_samples 9600 joint_samples 28 [1045090, 629462] processed_samples 9601 unjoint_samples 9600 joint_samples 28 [1039790, 751048] processed_samples 9602 unjoint_samples 9600 joint_samples 28 [1045090, 629462] [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e840ae40] mmco: unref short failure [h264 @ 0x5617e840ae40] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x559619420380] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559616d35c80] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure processed_samples 9700 unjoint_samples 9700 joint_samples 28 [497028, 1046571] processed_samples 9700 unjoint_samples 9700 joint_samples 28 [497028, 1046571] processed_samples 9700 unjoint_samples 9700 joint_samples 28 [781727, 1046690] processed_samples 9700 unjoint_samples 9700 joint_samples 28 [781727, 1046690] processed_samples 9700 unjoint_samples 9700 joint_samples 29 [146097, 1023182] processed_samples 9700 unjoint_samples 9700 joint_samples 29 [146097, 1023182] processed_samples 9701 unjoint_samples 9700 joint_samples 29 [1047533, 7263] processed_samples 9701 unjoint_samples 9700 joint_samples 29 [1047533, 7263] processed_samples 9700 unjoint_samples 9700 joint_samples 28 [1047189, 507840] processed_samples 9700 unjoint_samples 9700 joint_samples 28 [1047189, 507840] processed_samples 9700 unjoint_samples 9700 joint_samples 29 [1034198, 679486] processed_samples 9700 unjoint_samples 9700 joint_samples 29 [1034198, 679486] processed_samples 9700 unjoint_samples 9700 joint_samples 28 [962532, 1047120] processed_samples 9700 unjoint_samples 9700 joint_samples 28 [962532, 1047120] processed_samples 9702 unjoint_samples 9700 joint_samples 28 [1045090, 977030] processed_samples 9702 unjoint_samples 9700 joint_samples 28 [1045090, 977030] [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure local_disk/cognitron_vl/cognitron_vl/data/processor/image_processor.py:265: ResourceWarning: unclosed file <_io.BufferedReader name='/data/data/LMM/Vision-Flan/vision-flan_191-task_1k/images_191task_1k/WIT+detailed_description_66_Kinmen_Bridge_27922114926_e954daa682_o.jpg'> image = Image.open(img_or_path).convert("RGB") [h264 @ 0x5617ee5b7480] mmco: unref short failure local_disk/cognitron_vl/cognitron_vl/data/processor/image_processor.py:265: ResourceWarning: unclosed file <_io.BufferedReader name='/data/data/LMM/Vision-Flan/vision-flan_191-task_1k/images_191task_1k/WIT+detailed_description_66_Kinmen_Bridge_27922114926_e954daa682_o.jpg'> image = Image.open(img_or_path).convert("RGB") [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x559616d35c80] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure processed_samples 9800 unjoint_samples 9800 joint_samples 29 [1046214, 83209] processed_samples 9800 unjoint_samples 9800 joint_samples 29 [1046214, 83209] processed_samples 9800 unjoint_samples 9800 joint_samples 29 [1028861, 278488] processed_samples 9800 unjoint_samples 9800 joint_samples 29 [1028861, 278488] processed_samples 9800 unjoint_samples 9800 joint_samples 29 [416390, 1023182] processed_samples 9800 unjoint_samples 9800 joint_samples 29 [416390, 1023182] processed_samples 9800 unjoint_samples 9800 joint_samples 28 [1047189, 953498] processed_samples 9800 unjoint_samples 9800 joint_samples 28 [735088, 1046571] processed_samples 9800 unjoint_samples 9800 joint_samples 28 [1047189, 953498] processed_samples 9800 unjoint_samples 9800 joint_samples 28 [735088, 1046571] [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure processed_samples 9801 unjoint_samples 9800 joint_samples 29 [1047533, 303929] processed_samples 9801 unjoint_samples 9800 joint_samples 29 [1047533, 303929] processed_samples 9800 unjoint_samples 9800 joint_samples 29 [1034198, 950516] processed_samples 9802 unjoint_samples 9800 joint_samples 29 [193448, 1047656] processed_samples 9800 unjoint_samples 9800 joint_samples 29 [1034198, 950516] processed_samples 9802 unjoint_samples 9800 joint_samples 29 [193448, 1047656] [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure processed_samples 9900 unjoint_samples 9900 joint_samples 30 [1039364, 152377] [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure processed_samples 9900 unjoint_samples 9900 joint_samples 29 [274053, 1034779] [h264 @ 0x5596186f0980] mmco: unref short failure processed_samples 9900 unjoint_samples 9900 joint_samples 29 [1046214, 361387] processed_samples 9900 unjoint_samples 9900 joint_samples 29 [1028861, 547372] processed_samples 9900 unjoint_samples 9900 joint_samples 29 [689465, 1023182] processed_samples 9901 unjoint_samples 9900 joint_samples 29 [1047533, 582119] processed_samples 9900 unjoint_samples 9900 joint_samples 30 [1039364, 152377] processed_samples 9900 unjoint_samples 9900 joint_samples 28 [1002085, 1046571] [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure processed_samples 9900 unjoint_samples 9900 joint_samples 29 [274053, 1034779] processed_samples 9900 unjoint_samples 9900 joint_samples 29 [1046214, 361387] processed_samples 9900 unjoint_samples 9900 joint_samples 29 [1028861, 547372] processed_samples 9902 unjoint_samples 9900 joint_samples 29 [448945, 1047656] processed_samples 9900 unjoint_samples 9900 joint_samples 29 [689465, 1023182] processed_samples 9901 unjoint_samples 9900 joint_samples 29 [1047533, 582119] processed_samples 9900 unjoint_samples 9900 joint_samples 28 [1002085, 1046571] processed_samples 9902 unjoint_samples 9900 joint_samples 29 [448945, 1047656] [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure processed_samples 10000 unjoint_samples 10000 joint_samples 29 [205199, 1046571] [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure processed_samples 10000 unjoint_samples 10000 joint_samples 29 [527858, 1034779] processed_samples 10000 unjoint_samples 10000 joint_samples 30 [1039364, 405191] processed_samples 10000 unjoint_samples 10000 joint_samples 29 [1046214, 621133] processed_samples 10000 unjoint_samples 10000 joint_samples 29 [1028861, 863469] [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure processed_samples 10000 unjoint_samples 10000 joint_samples 29 [938667, 1023182] processed_samples 10001 unjoint_samples 10000 joint_samples 29 [1047533, 883817] [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure processed_samples 10002 unjoint_samples 10000 joint_samples 29 [765514, 1047656] [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure processed_samples 10000 unjoint_samples 10000 joint_samples 29 [205199, 1046571] processed_samples 10000 unjoint_samples 10000 joint_samples 29 [527858, 1034779] processed_samples 10000 unjoint_samples 10000 joint_samples 30 [1039364, 405191] processed_samples 10000 unjoint_samples 10000 joint_samples 29 [1046214, 621133] processed_samples 10000 unjoint_samples 10000 joint_samples 29 [1028861, 863469] processed_samples 10000 unjoint_samples 10000 joint_samples 29 [938667, 1023182] processed_samples 10001 unjoint_samples 10000 joint_samples 29 [1047533, 883817] [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure processed_samples 10002 unjoint_samples 10000 joint_samples 29 [765514, 1047656] [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x55961adc42c0] mmco: unref short failure [h264 @ 0x55961adc42c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure processed_samples 10100 unjoint_samples 10100 joint_samples 30 [244994, 1026689] processed_samples 10100 unjoint_samples 10100 joint_samples 30 [244994, 1026689] processed_samples 10100 unjoint_samples 10100 joint_samples 30 [253222, 1036754] processed_samples 10100 unjoint_samples 10100 joint_samples 29 [582573, 1046571] processed_samples 10101 unjoint_samples 10100 joint_samples 30 [1047533, 127841] processed_samples 10100 unjoint_samples 10100 joint_samples 29 [582573, 1046571] processed_samples 10100 unjoint_samples 10100 joint_samples 30 [1039364, 682401] processed_samples 10101 unjoint_samples 10100 joint_samples 30 [1047533, 127841] processed_samples 10100 unjoint_samples 10100 joint_samples 29 [1046214, 971261] processed_samples 10100 unjoint_samples 10100 joint_samples 30 [253222, 1036754] processed_samples 10100 unjoint_samples 10100 joint_samples 30 [1039364, 682401] processed_samples 10100 unjoint_samples 10100 joint_samples 29 [1046214, 971261] processed_samples 10102 unjoint_samples 10100 joint_samples 30 [59867, 1048086] processed_samples 10100 unjoint_samples 10100 joint_samples 29 [912312, 1034779] processed_samples 10102 unjoint_samples 10100 joint_samples 30 [59867, 1048086] processed_samples 10100 unjoint_samples 10100 joint_samples 29 [912312, 1034779] [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ed3f3040] mmco: unref short failure [h264 @ 0x5617ed3f3040] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure processed_samples 10200 unjoint_samples 10200 joint_samples 30 [264794, 1036524] [h264 @ 0x5617e9213b80] mmco: unref short failure processed_samples 10200 unjoint_samples 10200 joint_samples 30 [547504, 1026689] processed_samples 10200 unjoint_samples 10200 joint_samples 29 [895382, 1046571] processed_samples 10200 unjoint_samples 10200 joint_samples 30 [1032084, 373143] processed_samples 10201 unjoint_samples 10200 joint_samples 30 [1047533, 592193] processed_samples 10200 unjoint_samples 10200 joint_samples 30 [1039364, 953711] processed_samples 10200 unjoint_samples 10200 joint_samples 30 [696957, 1036754] processed_samples 10202 unjoint_samples 10200 joint_samples 30 [500790, 1048086] [h264 @ 0x559615eea1c0] mmco: unref short failure processed_samples 10200 unjoint_samples 10200 joint_samples 30 [264794, 1036524] processed_samples 10200 unjoint_samples 10200 joint_samples 30 [1032084, 373143] processed_samples 10200 unjoint_samples 10200 joint_samples 30 [547504, 1026689] [h264 @ 0x559617100c80] mmco: unref short failure processed_samples 10200 unjoint_samples 10200 joint_samples 30 [696957, 1036754] processed_samples 10201 unjoint_samples 10200 joint_samples 30 [1047533, 592193] processed_samples 10200 unjoint_samples 10200 joint_samples 29 [895382, 1046571] processed_samples 10202 unjoint_samples 10200 joint_samples 30 [500790, 1048086] processed_samples 10200 unjoint_samples 10200 joint_samples 30 [1039364, 953711] [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559618497a80] mmco: unref short failure [h264 @ 0x559618497a80] mmco: unref short failure [h264 @ 0x559618497a80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure processed_samples 10300 unjoint_samples 10300 joint_samples 30 [1046358, 128744] processed_samples 10300 unjoint_samples 10300 joint_samples 30 [1046358, 128744] processed_samples 10300 unjoint_samples 10300 joint_samples 31 [69557, 1037210] processed_samples 10300 unjoint_samples 10300 joint_samples 31 [69557, 1037210] processed_samples 10300 unjoint_samples 10300 joint_samples 30 [948686, 1036754] processed_samples 10300 unjoint_samples 10300 joint_samples 30 [948686, 1036754] processed_samples 10300 unjoint_samples 10300 joint_samples 30 [597819, 1036524] processed_samples 10300 unjoint_samples 10300 joint_samples 31 [260833, 1047211] processed_samples 10300 unjoint_samples 10300 joint_samples 31 [260833, 1047211] processed_samples 10300 unjoint_samples 10300 joint_samples 30 [1032084, 764670] processed_samples 10300 unjoint_samples 10300 joint_samples 30 [597819, 1036524] processed_samples 10300 unjoint_samples 10300 joint_samples 30 [1032084, 764670] processed_samples 10301 unjoint_samples 10300 joint_samples 30 [1047533, 875980] processed_samples 10302 unjoint_samples 10300 joint_samples 30 [759178, 1048086] processed_samples 10302 unjoint_samples 10300 joint_samples 30 [759178, 1048086] processed_samples 10301 unjoint_samples 10300 joint_samples 30 [1047533, 875980] [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617ed3f3040] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure processed_samples 10400 unjoint_samples 10400 joint_samples 31 [282902, 1037210] processed_samples 10400 unjoint_samples 10400 joint_samples 31 [282902, 1037210] processed_samples 10400 unjoint_samples 10400 joint_samples 30 [1046358, 443890] processed_samples 10400 unjoint_samples 10400 joint_samples 30 [1046358, 443890] processed_samples 10401 unjoint_samples 10400 joint_samples 31 [169963, 1035963] processed_samples 10400 unjoint_samples 10400 joint_samples 31 [214309, 1036754] processed_samples 10401 unjoint_samples 10400 joint_samples 31 [169963, 1035963] processed_samples 10400 unjoint_samples 10400 joint_samples 31 [214309, 1036754] [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure processed_samples 10400 unjoint_samples 10400 joint_samples 31 [556751, 1047211] processed_samples 10400 unjoint_samples 10400 joint_samples 31 [556751, 1047211] processed_samples 10400 unjoint_samples 10400 joint_samples 30 [1032084, 967476] processed_samples 10400 unjoint_samples 10400 joint_samples 30 [1032084, 967476] processed_samples 10400 unjoint_samples 10400 joint_samples 30 [920112, 1036524] processed_samples 10400 unjoint_samples 10400 joint_samples 30 [920112, 1036524] processed_samples 10402 unjoint_samples 10400 joint_samples 31 [1036162, 97455] processed_samples 10402 unjoint_samples 10400 joint_samples 31 [1036162, 97455] [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617ee8bc680] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617e8e0d640] mmco: unref short failure [h264 @ 0x5617e8e0d640] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure processed_samples 10500 unjoint_samples 10500 joint_samples 30 [1046358, 799845] processed_samples 10500 unjoint_samples 10500 joint_samples 30 [1046358, 799845] [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure processed_samples 10500 unjoint_samples 10500 joint_samples 31 [1046949, 170732] processed_samples 10500 unjoint_samples 10500 joint_samples 31 [1046949, 170732] processed_samples 10501 unjoint_samples 10500 joint_samples 31 [516782, 1035963] processed_samples 10500 unjoint_samples 10500 joint_samples 31 [687041, 1036754] processed_samples 10501 unjoint_samples 10500 joint_samples 31 [516782, 1035963] processed_samples 10500 unjoint_samples 10500 joint_samples 31 [687041, 1036754] processed_samples 10500 unjoint_samples 10500 joint_samples 31 [1032084, 291707] processed_samples 10500 unjoint_samples 10500 joint_samples 31 [1032084, 291707] processed_samples 10500 unjoint_samples 10500 joint_samples 31 [980075, 1047211] processed_samples 10500 unjoint_samples 10500 joint_samples 31 [980075, 1047211] processed_samples 10500 unjoint_samples 10500 joint_samples 31 [647245, 1037210] processed_samples 10500 unjoint_samples 10500 joint_samples 31 [647245, 1037210] processed_samples 10502 unjoint_samples 10500 joint_samples 31 [1036162, 430944] processed_samples 10502 unjoint_samples 10500 joint_samples 31 [1036162, 430944] [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559619420380] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559619420380] mmco: unref short failure [h264 @ 0x559619420380] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559619420380] mmco: unref short failure [h264 @ 0x559619420380] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559617afc780] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure processed_samples 10600 unjoint_samples 10600 joint_samples 31 [1048226, 166444] processed_samples 10600 unjoint_samples 10600 joint_samples 31 [1048226, 166444] processed_samples 10600 unjoint_samples 10600 joint_samples 32 [1019418, 45385] processed_samples 10600 unjoint_samples 10600 joint_samples 32 [1019418, 45385] processed_samples 10600 unjoint_samples 10600 joint_samples 32 [1047870, 283733] processed_samples 10600 unjoint_samples 10600 joint_samples 32 [1047870, 283733] processed_samples 10600 unjoint_samples 10600 joint_samples 31 [1032084, 574066] processed_samples 10600 unjoint_samples 10600 joint_samples 31 [1032084, 574066] processed_samples 10600 unjoint_samples 10600 joint_samples 31 [1017419, 1037210] processed_samples 10600 unjoint_samples 10600 joint_samples 31 [1017419, 1037210] processed_samples 10600 unjoint_samples 10600 joint_samples 31 [1046949, 496459] processed_samples 10600 unjoint_samples 10600 joint_samples 31 [1046949, 496459] processed_samples 10601 unjoint_samples 10600 joint_samples 31 [889831, 1035963] processed_samples 10601 unjoint_samples 10600 joint_samples 31 [889831, 1035963] processed_samples 10602 unjoint_samples 10600 joint_samples 31 [1036162, 754364] processed_samples 10602 unjoint_samples 10600 joint_samples 31 [1036162, 754364] [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559619420380] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559619420380] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x559616f5be80] mmco: unref short failure [h264 @ 0x559616f5be80] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure processed_samples 10700 unjoint_samples 10700 joint_samples 31 [1032084, 900352] processed_samples 10700 unjoint_samples 10700 joint_samples 32 [1019418, 352653] processed_samples 10700 unjoint_samples 10700 joint_samples 32 [1047870, 580347] processed_samples 10700 unjoint_samples 10700 joint_samples 31 [1048226, 417576] processed_samples 10700 unjoint_samples 10700 joint_samples 31 [1032084, 900352] processed_samples 10700 unjoint_samples 10700 joint_samples 31 [1046949, 785393] processed_samples 10701 unjoint_samples 10700 joint_samples 32 [999081, 218096] processed_samples 10700 unjoint_samples 10700 joint_samples 32 [1019418, 352653] [h264 @ 0x55961d3b8040] mmco: unref short failure processed_samples 10700 unjoint_samples 10700 joint_samples 32 [334874, 1047057] processed_samples 10700 unjoint_samples 10700 joint_samples 31 [1048226, 417576] processed_samples 10700 unjoint_samples 10700 joint_samples 31 [1046949, 785393] [h264 @ 0x5617f06c03c0] mmco: unref short failure processed_samples 10701 unjoint_samples 10700 joint_samples 32 [999081, 218096] processed_samples 10700 unjoint_samples 10700 joint_samples 32 [1047870, 580347] [h264 @ 0x5596186f0980] mmco: unref short failure processed_samples 10702 unjoint_samples 10700 joint_samples 31 [1040542, 1041087] processed_samples 10700 unjoint_samples 10700 joint_samples 32 [334874, 1047057] [h264 @ 0x5617ec417900] mmco: unref short failure processed_samples 10702 unjoint_samples 10700 joint_samples 31 [1040542, 1041087] [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure processed_samples 10800 unjoint_samples 10800 joint_samples 32 [238107, 1033744] processed_samples 10800 unjoint_samples 10800 joint_samples 32 [238107, 1033744] processed_samples 10800 unjoint_samples 10800 joint_samples 32 [1046949, 78052] processed_samples 10800 unjoint_samples 10800 joint_samples 32 [1046949, 78052] processed_samples 10800 unjoint_samples 10800 joint_samples 32 [555840, 1047057] processed_samples 10800 unjoint_samples 10800 joint_samples 32 [555840, 1047057] processed_samples 10800 unjoint_samples 10800 joint_samples 32 [1019418, 625106] processed_samples 10800 unjoint_samples 10800 joint_samples 32 [1019418, 625106] processed_samples 10800 unjoint_samples 10800 joint_samples 32 [1047870, 955860] processed_samples 10800 unjoint_samples 10800 joint_samples 32 [1047870, 955860] processed_samples 10801 unjoint_samples 10800 joint_samples 32 [999081, 518272] processed_samples 10801 unjoint_samples 10800 joint_samples 32 [999081, 518272] processed_samples 10800 unjoint_samples 10800 joint_samples 31 [1048226, 785824] processed_samples 10800 unjoint_samples 10800 joint_samples 31 [1048226, 785824] [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure processed_samples 10802 unjoint_samples 10800 joint_samples 32 [1045695, 249025] processed_samples 10802 unjoint_samples 10800 joint_samples 32 [1045695, 249025] [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x559618a89040] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure processed_samples 10900 unjoint_samples 10900 joint_samples 32 [81517, 1024678] processed_samples 10900 unjoint_samples 10900 joint_samples 32 [81517, 1024678] processed_samples 10900 unjoint_samples 10900 joint_samples 32 [1046949, 490170] processed_samples 10900 unjoint_samples 10900 joint_samples 32 [1046949, 490170] processed_samples 10900 unjoint_samples 10900 joint_samples 32 [795641, 1047057] processed_samples 10900 unjoint_samples 10900 joint_samples 32 [795641, 1047057] processed_samples 10901 unjoint_samples 10900 joint_samples 32 [999081, 839922] processed_samples 10901 unjoint_samples 10900 joint_samples 32 [999081, 839922] processed_samples 10900 unjoint_samples 10900 joint_samples 33 [1047870, 466877] processed_samples 10900 unjoint_samples 10900 joint_samples 33 [1047870, 466877] processed_samples 10900 unjoint_samples 10900 joint_samples 32 [477386, 1033744] processed_samples 10900 unjoint_samples 10900 joint_samples 32 [477386, 1033744] processed_samples 10900 unjoint_samples 10900 joint_samples 32 [1019418, 919297] processed_samples 10900 unjoint_samples 10900 joint_samples 32 [1019418, 919297] processed_samples 10902 unjoint_samples 10900 joint_samples 32 [1045695, 527152] processed_samples 10902 unjoint_samples 10900 joint_samples 32 [1045695, 527152] [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559613eda640] mmco: unref short failure [h264 @ 0x559613eda640] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x559613eda640] mmco: unref short failure [h264 @ 0x559613eda640] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559619420380] mmco: unref short failure [h264 @ 0x559619420380] mmco: unref short failure [h264 @ 0x559619420380] mmco: unref short failure [h264 @ 0x559619420380] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure processed_samples 11000 unjoint_samples 11000 joint_samples 33 [1048001, 129969] processed_samples 11000 unjoint_samples 11000 joint_samples 33 [1048001, 129969] processed_samples 11000 unjoint_samples 11000 joint_samples 32 [539209, 1024678] processed_samples 11000 unjoint_samples 11000 joint_samples 32 [539209, 1024678] processed_samples 11000 unjoint_samples 11000 joint_samples 32 [1046949, 945891] processed_samples 11000 unjoint_samples 11000 joint_samples 33 [1046381, 89011] processed_samples 11000 unjoint_samples 11000 joint_samples 32 [809154, 1033744] processed_samples 11000 unjoint_samples 11000 joint_samples 33 [1046381, 89011] processed_samples 11000 unjoint_samples 11000 joint_samples 32 [1046949, 945891] processed_samples 11000 unjoint_samples 11000 joint_samples 32 [809154, 1033744] processed_samples 11000 unjoint_samples 11000 joint_samples 33 [1047870, 794971] processed_samples 11001 unjoint_samples 11000 joint_samples 33 [91366, 1026237] processed_samples 11000 unjoint_samples 11000 joint_samples 33 [1047870, 794971] processed_samples 11001 unjoint_samples 11000 joint_samples 33 [91366, 1026237] processed_samples 11002 unjoint_samples 11000 joint_samples 32 [1045695, 806913] processed_samples 11002 unjoint_samples 11000 joint_samples 32 [1045695, 806913] [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure processed_samples 11100 unjoint_samples 11100 joint_samples 33 [1046949, 205768] processed_samples 11100 unjoint_samples 11100 joint_samples 33 [1046949, 205768] processed_samples 11100 unjoint_samples 11100 joint_samples 33 [126689, 1033744] processed_samples 11100 unjoint_samples 11100 joint_samples 33 [126689, 1033744] processed_samples 11100 unjoint_samples 11100 joint_samples 33 [1046381, 397425] processed_samples 11100 unjoint_samples 11100 joint_samples 33 [1046381, 397425] processed_samples 11100 unjoint_samples 11100 joint_samples 32 [775955, 1024678] processed_samples 11100 unjoint_samples 11100 joint_samples 32 [775955, 1024678] processed_samples 11100 unjoint_samples 11100 joint_samples 33 [1048001, 463107] processed_samples 11100 unjoint_samples 11100 joint_samples 33 [1048001, 463107] processed_samples 11100 unjoint_samples 11100 joint_samples 33 [1047870, 1039811] processed_samples 11100 unjoint_samples 11100 joint_samples 33 [1047870, 1039811] processed_samples 11101 unjoint_samples 11100 joint_samples 33 [357046, 1026237] processed_samples 11101 unjoint_samples 11100 joint_samples 33 [357046, 1026237] processed_samples 11102 unjoint_samples 11100 joint_samples 33 [211691, 999200] processed_samples 11102 unjoint_samples 11100 joint_samples 33 [211691, 999200] [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure processed_samples 11200 unjoint_samples 11200 joint_samples 33 [1048190, 59036] processed_samples 11200 unjoint_samples 11200 joint_samples 33 [1048190, 59036] [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure processed_samples 11200 unjoint_samples 11200 joint_samples 33 [1046381, 770553] processed_samples 11200 unjoint_samples 11200 joint_samples 33 [1046949, 561695] processed_samples 11200 unjoint_samples 11200 joint_samples 33 [1046381, 770553] processed_samples 11200 unjoint_samples 11200 joint_samples 34 [252231, 1045545] processed_samples 11200 unjoint_samples 11200 joint_samples 33 [1046949, 561695] processed_samples 11200 unjoint_samples 11200 joint_samples 34 [252231, 1045545] processed_samples 11200 unjoint_samples 11200 joint_samples 33 [1048001, 759032] processed_samples 11200 unjoint_samples 11200 joint_samples 33 [1048001, 759032] processed_samples 11200 unjoint_samples 11200 joint_samples 33 [396498, 1033744] processed_samples 11200 unjoint_samples 11200 joint_samples 33 [396498, 1033744] processed_samples 11201 unjoint_samples 11200 joint_samples 33 [772968, 1026237] processed_samples 11202 unjoint_samples 11200 joint_samples 33 [539981, 999200] processed_samples 11202 unjoint_samples 11200 joint_samples 33 [539981, 999200] processed_samples 11201 unjoint_samples 11200 joint_samples 33 [772968, 1026237] [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x559616fdd040] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617e980ab40] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559616d35c80] mmco: unref short failure [h264 @ 0x559616d35c80] mmco: unref short failure [h264 @ 0x559616d35c80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559616d35c80] mmco: unref short failure [h264 @ 0x559616d35c80] mmco: unref short failure [h264 @ 0x559616d35c80] mmco: unref short failure [h264 @ 0x559616d35c80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure processed_samples 11300 unjoint_samples 11300 joint_samples 34 [108312, 1046711] [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure processed_samples 11300 unjoint_samples 11300 joint_samples 34 [108312, 1046711] processed_samples 11300 unjoint_samples 11300 joint_samples 33 [1048190, 422022] processed_samples 11300 unjoint_samples 11300 joint_samples 33 [1046949, 820776] processed_samples 11300 unjoint_samples 11300 joint_samples 33 [1048190, 422022] processed_samples 11300 unjoint_samples 11300 joint_samples 34 [650086, 1045545] processed_samples 11300 unjoint_samples 11300 joint_samples 34 [650086, 1045545] processed_samples 11300 unjoint_samples 11300 joint_samples 33 [1046381, 1003909] processed_samples 11301 unjoint_samples 11300 joint_samples 34 [990990, 135428] processed_samples 11301 unjoint_samples 11300 joint_samples 34 [990990, 135428] processed_samples 11300 unjoint_samples 11300 joint_samples 33 [1046381, 1003909] processed_samples 11300 unjoint_samples 11300 joint_samples 33 [1046949, 820776] processed_samples 11300 unjoint_samples 11300 joint_samples 33 [752560, 1033744] processed_samples 11300 unjoint_samples 11300 joint_samples 33 [752560, 1033744] processed_samples 11302 unjoint_samples 11300 joint_samples 33 [854465, 999200] processed_samples 11302 unjoint_samples 11300 joint_samples 33 [854465, 999200] [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x559617afc780] mmco: unref short failure [h264 @ 0x559617afc780] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure processed_samples 11400 unjoint_samples 11400 joint_samples 34 [253323, 1046250] processed_samples 11400 unjoint_samples 11400 joint_samples 34 [253323, 1046250] processed_samples 11400 unjoint_samples 11400 joint_samples 34 [407812, 1046711] processed_samples 11400 unjoint_samples 11400 joint_samples 34 [147824, 1022132] processed_samples 11400 unjoint_samples 11400 joint_samples 34 [147824, 1022132] processed_samples 11400 unjoint_samples 11400 joint_samples 34 [239513, 1037545] processed_samples 11400 unjoint_samples 11400 joint_samples 34 [239513, 1037545] processed_samples 11400 unjoint_samples 11400 joint_samples 34 [407812, 1046711] processed_samples 11400 unjoint_samples 11400 joint_samples 34 [1012189, 1045545] processed_samples 11400 unjoint_samples 11400 joint_samples 34 [1012189, 1045545] processed_samples 11401 unjoint_samples 11400 joint_samples 34 [990990, 411019] processed_samples 11401 unjoint_samples 11400 joint_samples 34 [990990, 411019] processed_samples 11400 unjoint_samples 11400 joint_samples 33 [1048190, 785086] processed_samples 11400 unjoint_samples 11400 joint_samples 33 [1048190, 785086] processed_samples 11402 unjoint_samples 11400 joint_samples 34 [47547, 1041874] processed_samples 11402 unjoint_samples 11400 joint_samples 34 [47547, 1041874] [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617e9262740] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure processed_samples 11500 unjoint_samples 11500 joint_samples 34 [117226, 1046708] processed_samples 11500 unjoint_samples 11500 joint_samples 34 [117226, 1046708] processed_samples 11500 unjoint_samples 11500 joint_samples 34 [481959, 1022132] processed_samples 11500 unjoint_samples 11500 joint_samples 34 [588287, 1037545] processed_samples 11500 unjoint_samples 11500 joint_samples 34 [481959, 1022132] processed_samples 11500 unjoint_samples 11500 joint_samples 34 [588287, 1037545] processed_samples 11500 unjoint_samples 11500 joint_samples 34 [770554, 1046711] processed_samples 11500 unjoint_samples 11500 joint_samples 35 [1046751, 272203] processed_samples 11500 unjoint_samples 11500 joint_samples 35 [1046751, 272203] processed_samples 11500 unjoint_samples 11500 joint_samples 34 [770554, 1046711] processed_samples 11501 unjoint_samples 11500 joint_samples 34 [990990, 692665] processed_samples 11500 unjoint_samples 11500 joint_samples 34 [666788, 1046250] processed_samples 11500 unjoint_samples 11500 joint_samples 34 [666788, 1046250] processed_samples 11501 unjoint_samples 11500 joint_samples 34 [990990, 692665] processed_samples 11502 unjoint_samples 11500 joint_samples 34 [350269, 1041874] processed_samples 11502 unjoint_samples 11500 joint_samples 34 [350269, 1041874] [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure processed_samples 11600 unjoint_samples 11600 joint_samples 35 [985569, 81933] processed_samples 11600 unjoint_samples 11600 joint_samples 35 [985569, 81933] processed_samples 11600 unjoint_samples 11600 joint_samples 35 [57715, 1046711] processed_samples 11600 unjoint_samples 11600 joint_samples 35 [57715, 1046711] processed_samples 11600 unjoint_samples 11600 joint_samples 34 [459233, 1046708] processed_samples 11600 unjoint_samples 11600 joint_samples 34 [459233, 1046708] processed_samples 11600 unjoint_samples 11600 joint_samples 34 [861056, 1037545] processed_samples 11600 unjoint_samples 11600 joint_samples 35 [1046751, 730243] processed_samples 11600 unjoint_samples 11600 joint_samples 35 [1046751, 730243] processed_samples 11600 unjoint_samples 11600 joint_samples 34 [861056, 1037545] processed_samples 11601 unjoint_samples 11600 joint_samples 34 [1028594, 1030455] processed_samples 11601 unjoint_samples 11600 joint_samples 34 [1028594, 1030455] processed_samples 11600 unjoint_samples 11600 joint_samples 34 [854490, 1022132] processed_samples 11600 unjoint_samples 11600 joint_samples 34 [854490, 1022132] [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure processed_samples 11602 unjoint_samples 11600 joint_samples 34 [662976, 1041874] processed_samples 11602 unjoint_samples 11600 joint_samples 34 [662976, 1041874] [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617e9100f80] [h264 @ 0x55961864b640] mmco: unref short failure mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure processed_samples 11700 unjoint_samples 11700 joint_samples 36 [22002, 1046501] processed_samples 11700 unjoint_samples 11700 joint_samples 36 [22002, 1046501] processed_samples 11700 unjoint_samples 11700 joint_samples 35 [97623, 1047855] processed_samples 11700 unjoint_samples 11700 joint_samples 35 [97623, 1047855] processed_samples 11700 unjoint_samples 11700 joint_samples 35 [484010, 1046711] processed_samples 11700 unjoint_samples 11700 joint_samples 35 [484010, 1046711] processed_samples 11700 unjoint_samples 11700 joint_samples 35 [962603, 320927] processed_samples 11700 unjoint_samples 11700 joint_samples 35 [962603, 320927] [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure processed_samples 11701 unjoint_samples 11700 joint_samples 35 [336199, 1045546] processed_samples 11700 unjoint_samples 11700 joint_samples 35 [985569, 398007] processed_samples 11700 unjoint_samples 11700 joint_samples 35 [985569, 398007] processed_samples 11701 unjoint_samples 11700 joint_samples 35 [336199, 1045546] processed_samples 11700 unjoint_samples 11700 joint_samples 34 [766708, 1046708] [h264 @ 0x5617ebf51f00] mmco: unref short failure processed_samples 11700 unjoint_samples 11700 joint_samples 34 [766708, 1046708] [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure processed_samples 11702 unjoint_samples 11700 joint_samples 34 [1029590, 1041874] processed_samples 11702 unjoint_samples 11700 joint_samples 34 [1029590, 1041874] [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961adc42c0] mmco: unref short failure [h264 @ 0x55961adc42c0] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure processed_samples 11800 unjoint_samples 11800 joint_samples 35 [471541, 1047855] processed_samples 11800 unjoint_samples 11800 joint_samples 36 [306389, 1046501] processed_samples 11800 unjoint_samples 11800 joint_samples 35 [985569, 706218] processed_samples 11800 unjoint_samples 11800 joint_samples 34 [1026423, 1046708] processed_samples 11801 unjoint_samples 11800 joint_samples 35 [650400, 1045546] processed_samples 11800 unjoint_samples 11800 joint_samples 35 [741525, 1046711] processed_samples 11802 unjoint_samples 11800 joint_samples 35 [1045557, 373766] processed_samples 11800 unjoint_samples 11800 joint_samples 35 [962603, 585262] [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x559616d35c80] mmco: unref short failure [h264 @ 0x559616d35c80] mmco: unref short failure processed_samples 11800 unjoint_samples 11800 joint_samples 36 [306389, 1046501] processed_samples 11800 unjoint_samples 11800 joint_samples 34 [1026423, 1046708] processed_samples 11800 unjoint_samples 11800 joint_samples 35 [471541, 1047855] processed_samples 11800 unjoint_samples 11800 joint_samples 35 [985569, 706218] processed_samples 11800 unjoint_samples 11800 joint_samples 35 [741525, 1046711] [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure processed_samples 11802 unjoint_samples 11800 joint_samples 35 [1045557, 373766] processed_samples 11801 unjoint_samples 11800 joint_samples 35 [650400, 1045546] processed_samples 11800 unjoint_samples 11800 joint_samples 35 [962603, 585262] [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559616d35c80] mmco: unref short failure [h264 @ 0x559616d35c80] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617ea25da40] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559618a89040] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure processed_samples 11900 unjoint_samples 11900 joint_samples 35 [262734, 1046708] processed_samples 11900 unjoint_samples 11900 joint_samples 35 [985569, 985205] processed_samples 11900 unjoint_samples 11900 joint_samples 35 [1037545, 1046711] processed_samples 11900 unjoint_samples 11900 joint_samples 35 [830860, 1047855] [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure processed_samples 11900 unjoint_samples 11900 joint_samples 35 [262734, 1046708] processed_samples 11900 unjoint_samples 11900 joint_samples 36 [593539, 1046501] processed_samples 11900 unjoint_samples 11900 joint_samples 35 [985569, 985205] processed_samples 11901 unjoint_samples 11900 joint_samples 35 [1004598, 1045546] processed_samples 11900 unjoint_samples 11900 joint_samples 35 [830860, 1047855] [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure processed_samples 11900 unjoint_samples 11900 joint_samples 35 [1037545, 1046711] processed_samples 11902 unjoint_samples 11900 joint_samples 35 [1045557, 881592] processed_samples 11900 unjoint_samples 11900 joint_samples 36 [593539, 1046501] processed_samples 11901 unjoint_samples 11900 joint_samples 35 [1004598, 1045546] processed_samples 11900 unjoint_samples 11900 joint_samples 35 [962603, 887800] processed_samples 11902 unjoint_samples 11900 joint_samples 35 [1045557, 881592] processed_samples 11900 unjoint_samples 11900 joint_samples 35 [962603, 887800] [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e803bd40] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure processed_samples 12000 unjoint_samples 12000 joint_samples 36 [89760, 1027168] processed_samples 12000 unjoint_samples 12000 joint_samples 36 [89760, 1027168] processed_samples 12000 unjoint_samples 12000 joint_samples 36 [1027577, 140421] processed_samples 12000 unjoint_samples 12000 joint_samples 36 [1027577, 140421] processed_samples 12000 unjoint_samples 12000 joint_samples 36 [221901, 1040958] processed_samples 12000 unjoint_samples 12000 joint_samples 36 [221901, 1040958] processed_samples 12000 unjoint_samples 12000 joint_samples 36 [302257, 1046711] processed_samples 12000 unjoint_samples 12000 joint_samples 36 [302257, 1046711] processed_samples 12000 unjoint_samples 12000 joint_samples 36 [868734, 1046501] processed_samples 12000 unjoint_samples 12000 joint_samples 36 [868734, 1046501] processed_samples 12000 unjoint_samples 12000 joint_samples 35 [514701, 1046708] [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure processed_samples 12000 unjoint_samples 12000 joint_samples 35 [514701, 1046708] [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure processed_samples 12001 unjoint_samples 12000 joint_samples 36 [270403, 1047713] processed_samples 12001 unjoint_samples 12000 joint_samples 36 [270403, 1047713] processed_samples 12002 unjoint_samples 12000 joint_samples 36 [1045557, 181757] processed_samples 12002 unjoint_samples 12000 joint_samples 36 [1045557, 181757] [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure processed_samples 12100 unjoint_samples 12100 joint_samples 37 [101945, 1046501] processed_samples 12100 unjoint_samples 12100 joint_samples 37 [101945, 1046501] processed_samples 12100 unjoint_samples 12100 joint_samples 36 [425407, 1027168] processed_samples 12100 unjoint_samples 12100 joint_samples 36 [425407, 1027168] processed_samples 12100 unjoint_samples 12100 joint_samples 36 [1027577, 509968] processed_samples 12100 unjoint_samples 12100 joint_samples 36 [1027577, 509968] processed_samples 12100 unjoint_samples 12100 joint_samples 36 [473805, 1040958] processed_samples 12100 unjoint_samples 12100 joint_samples 36 [473805, 1040958] processed_samples 12100 unjoint_samples 12100 joint_samples 36 [562059, 1046711] processed_samples 12100 unjoint_samples 12100 joint_samples 36 [562059, 1046711] processed_samples 12101 unjoint_samples 12100 joint_samples 36 [518050, 1047713] processed_samples 12101 unjoint_samples 12100 joint_samples 36 [518050, 1047713] processed_samples 12100 unjoint_samples 12100 joint_samples 35 [852133, 1046708] processed_samples 12100 unjoint_samples 12100 joint_samples 35 [852133, 1046708] [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure processed_samples 12102 unjoint_samples 12100 joint_samples 36 [1045557, 517754] processed_samples 12102 unjoint_samples 12100 joint_samples 36 [1045557, 517754] [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x559618471cc0] mmco: unref short failure [h264 @ 0x559618471cc0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure processed_samples 12200 unjoint_samples 12200 joint_samples 36 [90269, 1046708] processed_samples 12200 unjoint_samples 12200 joint_samples 36 [90269, 1046708] processed_samples 12200 unjoint_samples 12200 joint_samples 36 [762756, 1040958] processed_samples 12200 unjoint_samples 12200 joint_samples 36 [743799, 1027168] processed_samples 12200 unjoint_samples 12200 joint_samples 36 [837934, 1046711] processed_samples 12200 unjoint_samples 12200 joint_samples 37 [372575, 1046501] processed_samples 12200 unjoint_samples 12200 joint_samples 36 [837934, 1046711] processed_samples 12200 unjoint_samples 12200 joint_samples 36 [1027577, 787086] processed_samples 12200 unjoint_samples 12200 joint_samples 36 [1027577, 787086] processed_samples 12201 unjoint_samples 12200 joint_samples 36 [916529, 1047713] processed_samples 12200 unjoint_samples 12200 joint_samples 36 [762756, 1040958] processed_samples 12200 unjoint_samples 12200 joint_samples 37 [372575, 1046501] processed_samples 12200 unjoint_samples 12200 joint_samples 36 [743799, 1027168] [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure processed_samples 12201 unjoint_samples 12200 joint_samples 36 [916529, 1047713] [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure processed_samples 12202 unjoint_samples 12200 joint_samples 36 [1045557, 790227] processed_samples 12202 unjoint_samples 12200 joint_samples 36 [1045557, 790227] [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [mov,mp4,m4a,3gp,3g2,mj2 @ 0x5596172b7c40] [mov,mp4,m4a,3gp,3g2,mj2 @ 0x5617ec2f1680] stream 0, offset 0x101588b: partial file stream 0, offset 0x101588b: partial file [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure processed_samples 12300 unjoint_samples 12300 joint_samples 37 [34570, 1046608] processed_samples 12300 unjoint_samples 12300 joint_samples 37 [1046584, 123148] [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure processed_samples 12300 unjoint_samples 12300 joint_samples 37 [63687, 1040958] processed_samples 12300 unjoint_samples 12300 joint_samples 37 [34570, 1046608] processed_samples 12300 unjoint_samples 12300 joint_samples 37 [1046584, 123148] [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure processed_samples 12300 unjoint_samples 12300 joint_samples 37 [63687, 1040958] processed_samples 12300 unjoint_samples 12300 joint_samples 36 [401376, 1046708] processed_samples 12300 unjoint_samples 12300 joint_samples 36 [401376, 1046708] processed_samples 12301 unjoint_samples 12300 joint_samples 37 [1020396, 72773] processed_samples 12301 unjoint_samples 12300 joint_samples 37 [1020396, 72773] processed_samples 12301 unjoint_samples 12300 joint_samples 37 [1040711, 184337] processed_samples 12301 unjoint_samples 12300 joint_samples 37 [1040711, 184337] processed_samples 12300 unjoint_samples 12300 joint_samples 37 [644909, 1046501] processed_samples 12300 unjoint_samples 12300 joint_samples 37 [644909, 1046501] processed_samples 12302 unjoint_samples 12300 joint_samples 37 [5827, 1046973] processed_samples 12302 unjoint_samples 12300 joint_samples 37 [5827, 1046973] [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5596145f5e40] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617ed3f3040] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617e803bd40] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure processed_samples 12400 unjoint_samples 12400 joint_samples 37 [341909, 1046608] processed_samples 12400 unjoint_samples 12400 joint_samples 37 [341909, 1046608] processed_samples 12400 unjoint_samples 12400 joint_samples 38 [24177, 1046501] [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure processed_samples 12400 unjoint_samples 12400 joint_samples 38 [24177, 1046501] processed_samples 12400 unjoint_samples 12400 joint_samples 36 [689414, 1046708] processed_samples 12400 unjoint_samples 12400 joint_samples 37 [1046584, 390260] processed_samples 12400 unjoint_samples 12400 joint_samples 37 [1046584, 390260] processed_samples 12400 unjoint_samples 12400 joint_samples 36 [689414, 1046708] processed_samples 12400 unjoint_samples 12400 joint_samples 37 [378643, 1040958] processed_samples 12400 unjoint_samples 12400 joint_samples 37 [378643, 1040958] processed_samples 12401 unjoint_samples 12400 joint_samples 37 [1020396, 379352] processed_samples 12401 unjoint_samples 12400 joint_samples 37 [1040711, 421051] processed_samples 12401 unjoint_samples 12400 joint_samples 37 [1020396, 379352] processed_samples 12401 unjoint_samples 12400 joint_samples 37 [1040711, 421051] processed_samples 12402 unjoint_samples 12400 joint_samples 37 [355393, 1046973] processed_samples 12402 unjoint_samples 12400 joint_samples 37 [355393, 1046973] [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5596143e2540] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure processed_samples 12500 unjoint_samples 12500 joint_samples 37 [740512, 1040958] processed_samples 12500 unjoint_samples 12500 joint_samples 37 [79252, 1046708] processed_samples 12500 unjoint_samples 12500 joint_samples 37 [598983, 1046608] [h264 @ 0x5596156ec940] mmco: unref short failure processed_samples 12500 unjoint_samples 12500 joint_samples 38 [416733, 1046501] processed_samples 12501 unjoint_samples 12500 joint_samples 37 [1040711, 790970] processed_samples 12500 unjoint_samples 12500 joint_samples 37 [740512, 1040958] [h264 @ 0x5617ebf1ebc0] mmco: unref short failure processed_samples 12500 unjoint_samples 12500 joint_samples 37 [79252, 1046708] processed_samples 12500 unjoint_samples 12500 joint_samples 37 [598983, 1046608] processed_samples 12500 unjoint_samples 12500 joint_samples 38 [416733, 1046501] processed_samples 12501 unjoint_samples 12500 joint_samples 37 [1040711, 790970] processed_samples 12500 unjoint_samples 12500 joint_samples 37 [1046584, 837487] processed_samples 12500 unjoint_samples 12500 joint_samples 37 [1046584, 837487] processed_samples 12502 unjoint_samples 12500 joint_samples 37 [639689, 1046973] processed_samples 12501 unjoint_samples 12500 joint_samples 37 [1020396, 668953] processed_samples 12501 unjoint_samples 12500 joint_samples 37 [1020396, 668953] processed_samples 12502 unjoint_samples 12500 joint_samples 37 [639689, 1046973] [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e8636300] mmco: unref short failure [h264 @ 0x5617e8636300] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure processed_samples 12600 unjoint_samples 12600 joint_samples 38 [1046584, 52330] processed_samples 12600 unjoint_samples 12600 joint_samples 37 [492357, 1046708] [h264 @ 0x5617ec342580] mmco: unref short failure processed_samples 12601 unjoint_samples 12600 joint_samples 38 [1042871, 194801] processed_samples 12600 unjoint_samples 12600 joint_samples 37 [1015004, 1040958] processed_samples 12600 unjoint_samples 12600 joint_samples 37 [984746, 1046608] processed_samples 12600 unjoint_samples 12600 joint_samples 38 [745489, 1046501] [h264 @ 0x5617e84806c0] mmco: unref short failure processed_samples 12601 unjoint_samples 12600 joint_samples 37 [1027477, 1025493] [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure processed_samples 12602 unjoint_samples 12600 joint_samples 37 [969269, 1046973] processed_samples 12600 unjoint_samples 12600 joint_samples 37 [492357, 1046708] processed_samples 12600 unjoint_samples 12600 joint_samples 38 [1046584, 52330] processed_samples 12601 unjoint_samples 12600 joint_samples 38 [1042871, 194801] processed_samples 12600 unjoint_samples 12600 joint_samples 38 [745489, 1046501] processed_samples 12600 unjoint_samples 12600 joint_samples 37 [984746, 1046608] [h264 @ 0x55961672f480] mmco: unref short failure processed_samples 12601 unjoint_samples 12600 joint_samples 37 [1027477, 1025493] processed_samples 12600 unjoint_samples 12600 joint_samples 37 [1015004, 1040958] [h264 @ 0x559616d0c780] mmco: unref short failure processed_samples 12602 unjoint_samples 12600 joint_samples 37 [969269, 1046973] [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5596177086c0] mmco: unref short failure [h264 @ 0x5596177086c0] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x559618a89040] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e9262740] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e7e34fc0] mmco: unref short failure [h264 @ 0x5617e7e34fc0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure processed_samples 12700 unjoint_samples 12700 joint_samples 38 [1046584, 396286] processed_samples 12700 unjoint_samples 12700 joint_samples 38 [1046584, 396286] processed_samples 12700 unjoint_samples 12700 joint_samples 38 [1043744, 262065] processed_samples 12700 unjoint_samples 12700 joint_samples 38 [1043744, 262065] processed_samples 12700 unjoint_samples 12700 joint_samples 37 [727565, 1046708] processed_samples 12700 unjoint_samples 12700 joint_samples 37 [727565, 1046708] processed_samples 12700 unjoint_samples 12700 joint_samples 38 [1047204, 311223] processed_samples 12700 unjoint_samples 12700 joint_samples 38 [1047204, 311223] processed_samples 12701 unjoint_samples 12700 joint_samples 38 [1042431, 261056] processed_samples 12701 unjoint_samples 12700 joint_samples 38 [1042431, 261056] processed_samples 12701 unjoint_samples 12700 joint_samples 38 [1042871, 502805] processed_samples 12701 unjoint_samples 12700 joint_samples 38 [1042871, 502805] processed_samples 12700 unjoint_samples 12700 joint_samples 38 [1021557, 1046501] processed_samples 12700 unjoint_samples 12700 joint_samples 38 [1021557, 1046501] processed_samples 12702 unjoint_samples 12700 joint_samples 38 [351973, 1046973] processed_samples 12702 unjoint_samples 12700 joint_samples 38 [351973, 1046973] [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617eeb18bc0] mmco: unref short failure [h264 @ 0x5617eeb18bc0] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617eeb18bc0] mmco: unref short failure [h264 @ 0x5617eeb18bc0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure processed_samples 12800 unjoint_samples 12800 joint_samples 38 [1047204, 650566] [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure processed_samples 12800 unjoint_samples 12800 joint_samples 38 [1043744, 550477] processed_samples 12800 unjoint_samples 12800 joint_samples 39 [1025129, 427229] processed_samples 12801 unjoint_samples 12800 joint_samples 38 [1042871, 848979] processed_samples 12800 unjoint_samples 12800 joint_samples 37 [1017221, 1046708] [h264 @ 0x55961864b640] mmco: unref short failure processed_samples 12800 unjoint_samples 12800 joint_samples 38 [1046584, 697841] processed_samples 12802 unjoint_samples 12800 joint_samples 38 [608794, 1046973] processed_samples 12800 unjoint_samples 12800 joint_samples 38 [1047204, 650566] processed_samples 12800 unjoint_samples 12800 joint_samples 39 [1025129, 427229] [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure processed_samples 12800 unjoint_samples 12800 joint_samples 38 [1046584, 697841] processed_samples 12800 unjoint_samples 12800 joint_samples 37 [1017221, 1046708] processed_samples 12801 unjoint_samples 12800 joint_samples 38 [1042431, 641317] processed_samples 12801 unjoint_samples 12800 joint_samples 38 [1042871, 848979] processed_samples 12800 unjoint_samples 12800 joint_samples 38 [1043744, 550477] [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure processed_samples 12802 unjoint_samples 12800 joint_samples 38 [608794, 1046973] [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure processed_samples 12801 unjoint_samples 12800 joint_samples 38 [1042431, 641317] [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617e8298040] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure processed_samples 12900 unjoint_samples 12900 joint_samples 38 [1046584, 956901] processed_samples 12900 unjoint_samples 12900 joint_samples 38 [1047204, 961018] processed_samples 12900 unjoint_samples 12900 joint_samples 38 [1036684, 254325] processed_samples 12900 unjoint_samples 12900 joint_samples 38 [1036684, 254325] processed_samples 12901 unjoint_samples 12900 joint_samples 39 [56648, 1046657] processed_samples 12901 unjoint_samples 12900 joint_samples 39 [56648, 1046657] processed_samples 12900 unjoint_samples 12900 joint_samples 39 [1025129, 741010] processed_samples 12900 unjoint_samples 12900 joint_samples 39 [1025129, 741010] [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure processed_samples 12900 unjoint_samples 12900 joint_samples 38 [1043744, 938536] processed_samples 12900 unjoint_samples 12900 joint_samples 38 [1043744, 938536] processed_samples 12900 unjoint_samples 12900 joint_samples 38 [1046584, 956901] [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure processed_samples 12900 unjoint_samples 12900 joint_samples 38 [1047204, 961018] processed_samples 12902 unjoint_samples 12900 joint_samples 38 [1007388, 1046973] processed_samples 12901 unjoint_samples 12900 joint_samples 38 [1042431, 968215] processed_samples 12901 unjoint_samples 12900 joint_samples 38 [1042431, 968215] [h264 @ 0x5617e9213b80] mmco: unref short failure processed_samples 12902 unjoint_samples 12900 joint_samples 38 [1007388, 1046973] [h264 @ 0x5617e7e34fc0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure processed_samples 13000 unjoint_samples 13000 joint_samples 39 [152095, 1046630] processed_samples 13000 unjoint_samples 13000 joint_samples 39 [152095, 1046630] processed_samples 13000 unjoint_samples 13000 joint_samples 40 [58826, 1035704] processed_samples 13000 unjoint_samples 13000 joint_samples 40 [58826, 1035704] processed_samples 13000 unjoint_samples 13000 joint_samples 39 [211581, 1046361] processed_samples 13000 unjoint_samples 13000 joint_samples 39 [211581, 1046361] [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure processed_samples 13001 unjoint_samples 13000 joint_samples 39 [407598, 1046657] processed_samples 13001 unjoint_samples 13000 joint_samples 39 [407598, 1046657] processed_samples 13000 unjoint_samples 13000 joint_samples 38 [1036684, 546200] processed_samples 13000 unjoint_samples 13000 joint_samples 38 [1036684, 546200] processed_samples 13000 unjoint_samples 13000 joint_samples 39 [232245, 1038953] processed_samples 13000 unjoint_samples 13000 joint_samples 39 [232245, 1038953] processed_samples 13001 unjoint_samples 13000 joint_samples 39 [183306, 1046413] processed_samples 13001 unjoint_samples 13000 joint_samples 39 [183306, 1046413] processed_samples 13002 unjoint_samples 13000 joint_samples 39 [1041616, 426865] processed_samples 13002 unjoint_samples 13000 joint_samples 39 [1041616, 426865] [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x559619420380] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559618471cc0] mmco: unref short failure [h264 @ 0x559618471cc0] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure processed_samples 13100 unjoint_samples 13100 joint_samples 38 [1036684, 824601] [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure processed_samples 13100 unjoint_samples 13100 joint_samples 38 [1036684, 824601] processed_samples 13100 unjoint_samples 13100 joint_samples 39 [457469, 1046630] processed_samples 13100 unjoint_samples 13100 joint_samples 40 [326528, 1035704] processed_samples 13100 unjoint_samples 13100 joint_samples 39 [506725, 1046361] processed_samples 13100 unjoint_samples 13100 joint_samples 39 [457469, 1046630] processed_samples 13100 unjoint_samples 13100 joint_samples 40 [326528, 1035704] processed_samples 13100 unjoint_samples 13100 joint_samples 39 [506725, 1046361] processed_samples 13100 unjoint_samples 13100 joint_samples 39 [564096, 1038953] processed_samples 13100 unjoint_samples 13100 joint_samples 39 [564096, 1038953] processed_samples 13102 unjoint_samples 13100 joint_samples 39 [1041616, 984423] processed_samples 13101 unjoint_samples 13100 joint_samples 39 [534744, 1046413] processed_samples 13102 unjoint_samples 13100 joint_samples 39 [1041616, 984423] processed_samples 13101 unjoint_samples 13100 joint_samples 39 [534744, 1046413] processed_samples 13101 unjoint_samples 13100 joint_samples 39 [846167, 1046657] processed_samples 13101 unjoint_samples 13100 joint_samples 39 [846167, 1046657] [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5596177086c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure processed_samples 13200 unjoint_samples 13200 joint_samples 40 [1030746, 71518] processed_samples 13200 unjoint_samples 13200 joint_samples 40 [1030746, 71518] processed_samples 13200 unjoint_samples 13200 joint_samples 39 [1047807, 106691] processed_samples 13200 unjoint_samples 13200 joint_samples 39 [1047807, 106691] processed_samples 13201 unjoint_samples 13200 joint_samples 40 [109162, 1046657] processed_samples 13201 unjoint_samples 13200 joint_samples 40 [109162, 1046657] processed_samples 13200 unjoint_samples 13200 joint_samples 39 [795509, 1046630] processed_samples 13200 unjoint_samples 13200 joint_samples 39 [795509, 1046630] processed_samples 13200 unjoint_samples 13200 joint_samples 39 [719624, 1046361] processed_samples 13200 unjoint_samples 13200 joint_samples 39 [719624, 1046361] processed_samples 13201 unjoint_samples 13200 joint_samples 39 [795397, 1046413] processed_samples 13201 unjoint_samples 13200 joint_samples 39 [795397, 1046413] processed_samples 13202 unjoint_samples 13200 joint_samples 40 [1046994, 258833] processed_samples 13202 unjoint_samples 13200 joint_samples 40 [1046994, 258833] processed_samples 13200 unjoint_samples 13200 joint_samples 40 [786945, 1035704] processed_samples 13200 unjoint_samples 13200 joint_samples 40 [786945, 1035704] [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x559617cd2580] mmco: unref short failure [h264 @ 0x559617cd2580] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure processed_samples 13300 unjoint_samples 13300 joint_samples 40 [1030746, 441110] processed_samples 13300 unjoint_samples 13300 joint_samples 40 [1030746, 441110] processed_samples 13300 unjoint_samples 13300 joint_samples 40 [69629, 1046673] processed_samples 13300 unjoint_samples 13300 joint_samples 40 [69629, 1046673] [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure processed_samples 13300 unjoint_samples 13300 joint_samples 41 [47228, 1040927] processed_samples 13300 unjoint_samples 13300 joint_samples 41 [47228, 1040927] processed_samples 13300 unjoint_samples 13300 joint_samples 39 [1047807, 621423] processed_samples 13300 unjoint_samples 13300 joint_samples 39 [1047807, 621423] processed_samples 13301 unjoint_samples 13300 joint_samples 40 [1043222, 91066] processed_samples 13301 unjoint_samples 13300 joint_samples 40 [1043222, 91066] processed_samples 13301 unjoint_samples 13300 joint_samples 40 [455890, 1046657] processed_samples 13301 unjoint_samples 13300 joint_samples 40 [455890, 1046657] processed_samples 13300 unjoint_samples 13300 joint_samples 39 [1019529, 1046361] processed_samples 13300 unjoint_samples 13300 joint_samples 39 [1019529, 1046361] processed_samples 13302 unjoint_samples 13300 joint_samples 40 [1046994, 531504] processed_samples 13302 unjoint_samples 13300 joint_samples 40 [1046994, 531504] [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ed3f3040] mmco: unref short failure [h264 @ 0x5617ed3f3040] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure processed_samples 13400 unjoint_samples 13400 joint_samples 39 [1047807, 983436] processed_samples 13400 unjoint_samples 13400 joint_samples 40 [417737, 1046673] [h264 @ 0x5617e9274140] mmco: unref short failure processed_samples 13400 unjoint_samples 13400 joint_samples 41 [323899, 1040927] processed_samples 13400 unjoint_samples 13400 joint_samples 40 [1030817, 341183] processed_samples 13400 unjoint_samples 13400 joint_samples 39 [1047807, 983436] processed_samples 13400 unjoint_samples 13400 joint_samples 40 [1030746, 660247] processed_samples 13401 unjoint_samples 13400 joint_samples 40 [1043222, 367098] processed_samples 13400 unjoint_samples 13400 joint_samples 40 [417737, 1046673] processed_samples 13401 unjoint_samples 13400 joint_samples 40 [896015, 1046657] processed_samples 13400 unjoint_samples 13400 joint_samples 41 [323899, 1040927] processed_samples 13400 unjoint_samples 13400 joint_samples 40 [1030817, 341183] processed_samples 13402 unjoint_samples 13400 joint_samples 40 [1046994, 934793] processed_samples 13400 unjoint_samples 13400 joint_samples 40 [1030746, 660247] processed_samples 13401 unjoint_samples 13400 joint_samples 40 [1043222, 367098] processed_samples 13402 unjoint_samples 13400 joint_samples 40 [1046994, 934793] processed_samples 13401 unjoint_samples 13400 joint_samples 40 [896015, 1046657] [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x55961adc42c0] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure processed_samples 13500 unjoint_samples 13500 joint_samples 40 [155464, 1040018] processed_samples 13500 unjoint_samples 13500 joint_samples 40 [155464, 1040018] processed_samples 13500 unjoint_samples 13500 joint_samples 41 [575767, 1040927] [h264 @ 0x55961adc42c0] mmco: unref short failure [h264 @ 0x55961adc42c0] mmco: unref short failure processed_samples 13500 unjoint_samples 13500 joint_samples 41 [575767, 1040927] [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure processed_samples 13500 unjoint_samples 13500 joint_samples 40 [646936, 1046673] processed_samples 13501 unjoint_samples 13500 joint_samples 41 [193704, 1046657] processed_samples 13501 unjoint_samples 13500 joint_samples 41 [193704, 1046657] processed_samples 13500 unjoint_samples 13500 joint_samples 40 [646936, 1046673] processed_samples 13500 unjoint_samples 13500 joint_samples 40 [1030817, 604871] processed_samples 13500 unjoint_samples 13500 joint_samples 40 [1030746, 960251] processed_samples 13500 unjoint_samples 13500 joint_samples 40 [1030817, 604871] processed_samples 13500 unjoint_samples 13500 joint_samples 40 [1030746, 960251] processed_samples 13502 unjoint_samples 13500 joint_samples 41 [250520, 1026509] processed_samples 13502 unjoint_samples 13500 joint_samples 41 [250520, 1026509] processed_samples 13501 unjoint_samples 13500 joint_samples 40 [1043222, 621596] processed_samples 13501 unjoint_samples 13500 joint_samples 40 [1043222, 621596] [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596177086c0] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure processed_samples 13600 unjoint_samples 13600 joint_samples 41 [1046464, 197408] processed_samples 13600 unjoint_samples 13600 joint_samples 40 [1030817, 960021] [h264 @ 0x559617e83980] mmco: unref short failure [h264 @ 0x559617e83980] mmco: unref short failure processed_samples 13600 unjoint_samples 13600 joint_samples 40 [878486, 1046673] [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure processed_samples 13601 unjoint_samples 13600 joint_samples 41 [577826, 1046657] processed_samples 13600 unjoint_samples 13600 joint_samples 40 [597133, 1040018] processed_samples 13600 unjoint_samples 13600 joint_samples 41 [819451, 1040927] [h264 @ 0x55961d016580] mmco: unref short failure processed_samples 13602 unjoint_samples 13600 joint_samples 41 [529592, 1026509] processed_samples 13601 unjoint_samples 13600 joint_samples 40 [1043222, 966820] [h264 @ 0x559614220440] mmco: unref short failure processed_samples 13600 unjoint_samples 13600 joint_samples 40 [1030817, 960021] processed_samples 13600 unjoint_samples 13600 joint_samples 41 [1046464, 197408] processed_samples 13600 unjoint_samples 13600 joint_samples 40 [878486, 1046673] [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure processed_samples 13600 unjoint_samples 13600 joint_samples 41 [819451, 1040927] processed_samples 13600 unjoint_samples 13600 joint_samples 40 [597133, 1040018] [h264 @ 0x55961c1ba780] mmco: unref short failure processed_samples 13601 unjoint_samples 13600 joint_samples 41 [577826, 1046657] [h264 @ 0x5617e84806c0] mmco: unref short failure processed_samples 13602 unjoint_samples 13600 joint_samples 41 [529592, 1026509] [h264 @ 0x5617e87a0740] mmco: unref short failure processed_samples 13601 unjoint_samples 13600 joint_samples 40 [1043222, 966820] [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure processed_samples 13700 unjoint_samples 13700 joint_samples 41 [1045722, 275703] processed_samples 13700 unjoint_samples 13700 joint_samples 41 [1045722, 275703] processed_samples 13700 unjoint_samples 13700 joint_samples 41 [1046345, 308461] processed_samples 13700 unjoint_samples 13700 joint_samples 41 [1046345, 308461] processed_samples 13700 unjoint_samples 13700 joint_samples 42 [1027739, 106581] processed_samples 13700 unjoint_samples 13700 joint_samples 42 [1027739, 106581] processed_samples 13700 unjoint_samples 13700 joint_samples 41 [1046464, 505939] processed_samples 13700 unjoint_samples 13700 joint_samples 41 [1046464, 505939] processed_samples 13700 unjoint_samples 13700 joint_samples 40 [1011663, 1040018] processed_samples 13700 unjoint_samples 13700 joint_samples 40 [1011663, 1040018] [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure processed_samples 13701 unjoint_samples 13700 joint_samples 41 [1043222, 251644] processed_samples 13701 unjoint_samples 13700 joint_samples 41 [1043222, 251644] [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure processed_samples 13701 unjoint_samples 13700 joint_samples 41 [1046838, 1046657] processed_samples 13701 unjoint_samples 13700 joint_samples 41 [1046838, 1046657] processed_samples 13702 unjoint_samples 13700 joint_samples 41 [904533, 1026509] processed_samples 13702 unjoint_samples 13700 joint_samples 41 [904533, 1026509] [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ea25da40] mmco: unref short failure [h264 @ 0x5617ea25da40] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure processed_samples 13800 unjoint_samples 13800 joint_samples 41 [211038, 1047013] processed_samples 13800 unjoint_samples 13800 joint_samples 41 [211038, 1047013] processed_samples 13800 unjoint_samples 13800 joint_samples 41 [1045722, 694986] processed_samples 13800 unjoint_samples 13800 joint_samples 41 [1045722, 694986] processed_samples 13801 unjoint_samples 13800 joint_samples 42 [303717, 1046657] processed_samples 13801 unjoint_samples 13800 joint_samples 42 [303717, 1046657] processed_samples 13800 unjoint_samples 13800 joint_samples 42 [1027739, 385136] processed_samples 13800 unjoint_samples 13800 joint_samples 42 [1027739, 385136] processed_samples 13800 unjoint_samples 13800 joint_samples 41 [1046464, 910153] processed_samples 13800 unjoint_samples 13800 joint_samples 41 [1046464, 910153] processed_samples 13801 unjoint_samples 13800 joint_samples 41 [1043222, 566325] processed_samples 13801 unjoint_samples 13800 joint_samples 41 [1043222, 566325] processed_samples 13800 unjoint_samples 13800 joint_samples 41 [1046345, 674999] processed_samples 13800 unjoint_samples 13800 joint_samples 41 [1046345, 674999] processed_samples 13802 unjoint_samples 13800 joint_samples 42 [1040355, 251441] processed_samples 13802 unjoint_samples 13800 joint_samples 42 [1040355, 251441] [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure processed_samples 13900 unjoint_samples 13900 joint_samples 42 [133328, 1040381] processed_samples 13900 unjoint_samples 13900 joint_samples 42 [133328, 1040381] processed_samples 13900 unjoint_samples 13900 joint_samples 42 [1027739, 729495] processed_samples 13900 unjoint_samples 13900 joint_samples 42 [1027739, 729495] [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure processed_samples 13900 unjoint_samples 13900 joint_samples 41 [502307, 1047013] [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure processed_samples 13900 unjoint_samples 13900 joint_samples 41 [502307, 1047013] processed_samples 13901 unjoint_samples 13900 joint_samples 42 [603347, 1046657] processed_samples 13901 unjoint_samples 13900 joint_samples 42 [603347, 1046657] processed_samples 13900 unjoint_samples 13900 joint_samples 41 [1045722, 988491] processed_samples 13900 unjoint_samples 13900 joint_samples 41 [1045722, 988491] processed_samples 13900 unjoint_samples 13900 joint_samples 41 [1046345, 931975] processed_samples 13900 unjoint_samples 13900 joint_samples 41 [1046345, 931975] [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure processed_samples 13901 unjoint_samples 13900 joint_samples 41 [1043222, 808171] [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure processed_samples 13902 unjoint_samples 13900 joint_samples 42 [1040355, 482687] processed_samples 13902 unjoint_samples 13900 joint_samples 42 [1040355, 482687] processed_samples 13901 unjoint_samples 13900 joint_samples 41 [1043222, 808171] [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure processed_samples 14000 unjoint_samples 14000 joint_samples 42 [224939, 1046531] processed_samples 14000 unjoint_samples 14000 joint_samples 42 [300434, 1039116] processed_samples 14000 unjoint_samples 14000 joint_samples 42 [648178, 1040381] processed_samples 14000 unjoint_samples 14000 joint_samples 42 [648178, 1040381] processed_samples 14000 unjoint_samples 14000 joint_samples 41 [847942, 1047013] processed_samples 14000 unjoint_samples 14000 joint_samples 42 [224939, 1046531] processed_samples 14000 unjoint_samples 14000 joint_samples 42 [300434, 1039116] processed_samples 14000 unjoint_samples 14000 joint_samples 41 [847942, 1047013] processed_samples 14001 unjoint_samples 14000 joint_samples 42 [986822, 1046657] processed_samples 14000 unjoint_samples 14000 joint_samples 42 [1039071, 1038964] processed_samples 14000 unjoint_samples 14000 joint_samples 42 [1039071, 1038964] processed_samples 14001 unjoint_samples 14000 joint_samples 42 [74987, 1043461] processed_samples 14001 unjoint_samples 14000 joint_samples 42 [74987, 1043461] processed_samples 14001 unjoint_samples 14000 joint_samples 42 [986822, 1046657] [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure processed_samples 14002 unjoint_samples 14000 joint_samples 42 [1040355, 761504] processed_samples 14002 unjoint_samples 14000 joint_samples 42 [1040355, 761504] [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559618395d00] [h264 @ 0x5617ec447000] mmco: unref short failure mmco: unref short failure [h264 @ 0x5617ec447000] [h264 @ 0x559618395d00] mmco: unref short failure mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure processed_samples 14100 unjoint_samples 14100 joint_samples 42 [86872, 1047013] processed_samples 14100 unjoint_samples 14100 joint_samples 42 [86872, 1047013] processed_samples 14100 unjoint_samples 14100 joint_samples 42 [548590, 1046531] processed_samples 14100 unjoint_samples 14100 joint_samples 42 [548590, 1046531] processed_samples 14100 unjoint_samples 14100 joint_samples 42 [927486, 1040381] processed_samples 14100 unjoint_samples 14100 joint_samples 42 [927486, 1040381] [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure processed_samples 14101 unjoint_samples 14100 joint_samples 43 [1044829, 251905] processed_samples 14101 unjoint_samples 14100 joint_samples 43 [1044829, 251905] processed_samples 14100 unjoint_samples 14100 joint_samples 43 [256924, 1046082] processed_samples 14100 unjoint_samples 14100 joint_samples 43 [256924, 1046082] processed_samples 14100 unjoint_samples 14100 joint_samples 42 [659894, 1039116] processed_samples 14100 unjoint_samples 14100 joint_samples 42 [659894, 1039116] processed_samples 14101 unjoint_samples 14100 joint_samples 42 [427070, 1043461] processed_samples 14102 unjoint_samples 14100 joint_samples 43 [29465, 1047234] processed_samples 14102 unjoint_samples 14100 joint_samples 43 [29465, 1047234] processed_samples 14101 unjoint_samples 14100 joint_samples 42 [427070, 1043461] [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x559618393880] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure processed_samples 14200 unjoint_samples 14200 joint_samples 42 [450446, 1047013] processed_samples 14200 unjoint_samples 14200 joint_samples 43 [568692, 1046082] processed_samples 14200 unjoint_samples 14200 joint_samples 43 [1048102, 124200] processed_samples 14200 unjoint_samples 14200 joint_samples 42 [450446, 1047013] processed_samples 14200 unjoint_samples 14200 joint_samples 43 [568692, 1046082] processed_samples 14200 unjoint_samples 14200 joint_samples 43 [1048102, 124200] [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure processed_samples 14201 unjoint_samples 14200 joint_samples 43 [1044829, 576908] processed_samples 14201 unjoint_samples 14200 joint_samples 43 [1044829, 576908] processed_samples 14200 unjoint_samples 14200 joint_samples 42 [941611, 1046531] processed_samples 14200 unjoint_samples 14200 joint_samples 42 [941611, 1046531] processed_samples 14200 unjoint_samples 14200 joint_samples 43 [1042502, 31054] processed_samples 14200 unjoint_samples 14200 joint_samples 43 [1042502, 31054] processed_samples 14201 unjoint_samples 14200 joint_samples 42 [775376, 1043461] processed_samples 14201 unjoint_samples 14200 joint_samples 42 [775376, 1043461] processed_samples 14202 unjoint_samples 14200 joint_samples 43 [353152, 1047234] processed_samples 14202 unjoint_samples 14200 joint_samples 43 [353152, 1047234] [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure processed_samples 14300 unjoint_samples 14300 joint_samples 43 [1042502, 411055] processed_samples 14300 unjoint_samples 14300 joint_samples 43 [218592, 1046589] processed_samples 14300 unjoint_samples 14300 joint_samples 42 [722825, 1047013] processed_samples 14300 unjoint_samples 14300 joint_samples 43 [1048102, 461346] processed_samples 14300 unjoint_samples 14300 joint_samples 43 [1042502, 411055] processed_samples 14300 unjoint_samples 14300 joint_samples 43 [218592, 1046589] processed_samples 14300 unjoint_samples 14300 joint_samples 42 [722825, 1047013] processed_samples 14300 unjoint_samples 14300 joint_samples 43 [899151, 1046082] processed_samples 14301 unjoint_samples 14300 joint_samples 42 [1044688, 1045564] processed_samples 14300 unjoint_samples 14300 joint_samples 43 [1048102, 461346] processed_samples 14302 unjoint_samples 14300 joint_samples 43 [672522, 1047234] processed_samples 14301 unjoint_samples 14300 joint_samples 43 [1044829, 862791] processed_samples 14302 unjoint_samples 14300 joint_samples 43 [672522, 1047234] processed_samples 14300 unjoint_samples 14300 joint_samples 43 [899151, 1046082] processed_samples 14301 unjoint_samples 14300 joint_samples 43 [1044829, 862791] [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure processed_samples 14301 unjoint_samples 14300 joint_samples 42 [1044688, 1045564] [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e9741400] mmco: unref short failure [h264 @ 0x5617e9741400] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure processed_samples 14400 unjoint_samples 14400 joint_samples 43 [1046704, 44848] processed_samples 14400 unjoint_samples 14400 joint_samples 43 [1046704, 44848] processed_samples 14400 unjoint_samples 14400 joint_samples 43 [614090, 1046589] processed_samples 14400 unjoint_samples 14400 joint_samples 43 [614090, 1046589] [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure processed_samples 14401 unjoint_samples 14400 joint_samples 44 [149264, 1037076] processed_samples 14401 unjoint_samples 14400 joint_samples 44 [149264, 1037076] processed_samples 14400 unjoint_samples 14400 joint_samples 44 [1047243, 159173] processed_samples 14400 unjoint_samples 14400 joint_samples 44 [1047243, 159173] processed_samples 14401 unjoint_samples 14400 joint_samples 43 [285762, 1046516] processed_samples 14401 unjoint_samples 14400 joint_samples 43 [285762, 1046516] processed_samples 14400 unjoint_samples 14400 joint_samples 43 [1048102, 686639] processed_samples 14400 unjoint_samples 14400 joint_samples 43 [1048102, 686639] processed_samples 14400 unjoint_samples 14400 joint_samples 43 [1042502, 758418] processed_samples 14400 unjoint_samples 14400 joint_samples 43 [1042502, 758418] [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure processed_samples 14402 unjoint_samples 14400 joint_samples 43 [941912, 1047234] processed_samples 14402 unjoint_samples 14400 joint_samples 43 [941912, 1047234] [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x55961adc42c0] mmco: unref short failure [h264 @ 0x55961adc42c0] mmco: unref short failure processed_samples 14500 unjoint_samples 14500 joint_samples 44 [87615, 1025751] [h264 @ 0x55961adc42c0] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure processed_samples 14500 unjoint_samples 14500 joint_samples 44 [87615, 1025751] processed_samples 14500 unjoint_samples 14500 joint_samples 43 [1046704, 362893] processed_samples 14501 unjoint_samples 14500 joint_samples 44 [483205, 1037076] processed_samples 14500 unjoint_samples 14500 joint_samples 43 [919901, 1046589] processed_samples 14501 unjoint_samples 14500 joint_samples 43 [640490, 1046516] [h264 @ 0x559618693e80] mmco: unref short failure processed_samples 14500 unjoint_samples 14500 joint_samples 44 [1047243, 572141] processed_samples 14500 unjoint_samples 14500 joint_samples 43 [1048102, 923783] processed_samples 14502 unjoint_samples 14500 joint_samples 44 [1047665, 227152] processed_samples 14500 unjoint_samples 14500 joint_samples 43 [1046704, 362893] processed_samples 14501 unjoint_samples 14500 joint_samples 44 [483205, 1037076] processed_samples 14500 unjoint_samples 14500 joint_samples 44 [1047243, 572141] processed_samples 14500 unjoint_samples 14500 joint_samples 43 [919901, 1046589] processed_samples 14501 unjoint_samples 14500 joint_samples 43 [640490, 1046516] processed_samples 14500 unjoint_samples 14500 joint_samples 43 [1048102, 923783] processed_samples 14502 unjoint_samples 14500 joint_samples 44 [1047665, 227152] [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure processed_samples 14600 unjoint_samples 14600 joint_samples 43 [1046704, 756518] processed_samples 14600 unjoint_samples 14600 joint_samples 44 [240363, 1037647] [h264 @ 0x559613728d00] mmco: unref short failure processed_samples 14600 unjoint_samples 14600 joint_samples 44 [128321, 1046589] processed_samples 14600 unjoint_samples 14600 joint_samples 44 [240363, 1037647] [h264 @ 0x5617e9493900] mmco: unref short failure processed_samples 14600 unjoint_samples 14600 joint_samples 44 [364597, 1025751] processed_samples 14600 unjoint_samples 14600 joint_samples 44 [128321, 1046589] processed_samples 14600 unjoint_samples 14600 joint_samples 44 [364597, 1025751] processed_samples 14600 unjoint_samples 14600 joint_samples 43 [1046704, 756518] processed_samples 14601 unjoint_samples 14600 joint_samples 43 [937915, 1046516] processed_samples 14600 unjoint_samples 14600 joint_samples 44 [1047243, 880662] [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure processed_samples 14602 unjoint_samples 14600 joint_samples 44 [1047665, 461009] processed_samples 14601 unjoint_samples 14600 joint_samples 44 [887680, 1037076] processed_samples 14601 unjoint_samples 14600 joint_samples 44 [887680, 1037076] processed_samples 14602 unjoint_samples 14600 joint_samples 44 [1047665, 461009] processed_samples 14601 unjoint_samples 14600 joint_samples 43 [937915, 1046516] processed_samples 14600 unjoint_samples 14600 joint_samples 44 [1047243, 880662] [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure processed_samples 14700 unjoint_samples 14700 joint_samples 44 [1046704, 10841] processed_samples 14700 unjoint_samples 14700 joint_samples 44 [1046704, 10841] [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure processed_samples 14700 unjoint_samples 14700 joint_samples 45 [1047243, 141163] processed_samples 14700 unjoint_samples 14700 joint_samples 45 [1047243, 141163] processed_samples 14701 unjoint_samples 14700 joint_samples 45 [1042469, 133612] processed_samples 14701 unjoint_samples 14700 joint_samples 45 [1042469, 133612] processed_samples 14700 unjoint_samples 14700 joint_samples 44 [474367, 1037647] processed_samples 14700 unjoint_samples 14700 joint_samples 44 [474367, 1037647] processed_samples 14701 unjoint_samples 14700 joint_samples 44 [1039774, 144255] processed_samples 14701 unjoint_samples 14700 joint_samples 44 [1039774, 144255] processed_samples 14700 unjoint_samples 14700 joint_samples 44 [508766, 1046589] processed_samples 14700 unjoint_samples 14700 joint_samples 44 [508766, 1046589] processed_samples 14700 unjoint_samples 14700 joint_samples 44 [652485, 1025751] processed_samples 14700 unjoint_samples 14700 joint_samples 44 [652485, 1025751] [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure processed_samples 14702 unjoint_samples 14700 joint_samples 44 [1047665, 840460] processed_samples 14702 unjoint_samples 14700 joint_samples 44 [1047665, 840460] [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x55961d3b8040] [h264 @ 0x5617e829eac0] mmco: unref short failure mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x559618411940] mmco: unref short failure [h264 @ 0x559618411940] mmco: unref short failure [h264 @ 0x5617eeefb080] mmco: unref short failure [h264 @ 0x5617eeefb080] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure processed_samples 14800 unjoint_samples 14800 joint_samples 45 [1023383, 131564] processed_samples 14800 unjoint_samples 14800 joint_samples 45 [1023383, 131564] processed_samples 14800 unjoint_samples 14800 joint_samples 44 [1046704, 249875] processed_samples 14800 unjoint_samples 14800 joint_samples 44 [1046704, 249875] processed_samples 14801 unjoint_samples 14800 joint_samples 44 [1039774, 477601] processed_samples 14801 unjoint_samples 14800 joint_samples 44 [1039774, 477601] [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure processed_samples 14800 unjoint_samples 14800 joint_samples 44 [742837, 1037647] [h264 @ 0x5617ec251b40] mmco: unref short failure processed_samples 14800 unjoint_samples 14800 joint_samples 44 [742837, 1037647] [h264 @ 0x559614220440] mmco: unref short failure processed_samples 14801 unjoint_samples 14800 joint_samples 45 [1042469, 457446] processed_samples 14801 unjoint_samples 14800 joint_samples 45 [1042469, 457446] processed_samples 14800 unjoint_samples 14800 joint_samples 45 [1047243, 447654] processed_samples 14800 unjoint_samples 14800 joint_samples 45 [1047243, 447654] [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure processed_samples 14800 unjoint_samples 14800 joint_samples 44 [815860, 1046589] processed_samples 14800 unjoint_samples 14800 joint_samples 44 [815860, 1046589] processed_samples 14802 unjoint_samples 14800 joint_samples 45 [1047665, 177665] processed_samples 14802 unjoint_samples 14800 joint_samples 45 [1047665, 177665] [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure processed_samples 14900 unjoint_samples 14900 joint_samples 45 [1003629, 99727] [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure processed_samples 14900 unjoint_samples 14900 joint_samples 45 [1003629, 99727] processed_samples 14900 unjoint_samples 14900 joint_samples 44 [1046704, 537011] processed_samples 14900 unjoint_samples 14900 joint_samples 44 [1046704, 537011] processed_samples 14900 unjoint_samples 14900 joint_samples 45 [1023383, 460547] processed_samples 14900 unjoint_samples 14900 joint_samples 45 [1023383, 460547] processed_samples 14900 unjoint_samples 14900 joint_samples 45 [1047243, 812021] processed_samples 14901 unjoint_samples 14900 joint_samples 45 [1042469, 797589] processed_samples 14900 unjoint_samples 14900 joint_samples 45 [1047243, 812021] processed_samples 14901 unjoint_samples 14900 joint_samples 45 [1042469, 797589] processed_samples 14900 unjoint_samples 14900 joint_samples 44 [1022567, 1037647] processed_samples 14902 unjoint_samples 14900 joint_samples 45 [1047665, 576883] processed_samples 14901 unjoint_samples 14900 joint_samples 44 [1039774, 873131] processed_samples 14901 unjoint_samples 14900 joint_samples 44 [1039774, 873131] processed_samples 14900 unjoint_samples 14900 joint_samples 44 [1022567, 1037647] processed_samples 14902 unjoint_samples 14900 joint_samples 45 [1047665, 576883] [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x55961342b000] [h264 @ 0x5617ee580e80] mmco: unref short failure mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure processed_samples 15000 unjoint_samples 15000 joint_samples 46 [1047956, 143474] processed_samples 15000 unjoint_samples 15000 joint_samples 45 [1003629, 398899] processed_samples 15000 unjoint_samples 15000 joint_samples 46 [1047956, 143474] processed_samples 15000 unjoint_samples 15000 joint_samples 45 [1037537, 538334] processed_samples 15000 unjoint_samples 15000 joint_samples 45 [1003629, 398899] processed_samples 15001 unjoint_samples 15000 joint_samples 45 [95734, 1045614] processed_samples 15001 unjoint_samples 15000 joint_samples 46 [260222, 1036416] processed_samples 15000 unjoint_samples 15000 joint_samples 45 [1037537, 538334] processed_samples 15000 unjoint_samples 15000 joint_samples 44 [1046704, 767828] processed_samples 15000 unjoint_samples 15000 joint_samples 45 [1023383, 997577] [h264 @ 0x5617e9213b80] mmco: unref short failure processed_samples 15001 unjoint_samples 15000 joint_samples 45 [95734, 1045614] processed_samples 15001 unjoint_samples 15000 joint_samples 46 [260222, 1036416] processed_samples 15000 unjoint_samples 15000 joint_samples 45 [1023383, 997577] [h264 @ 0x559614220440] mmco: unref short failure processed_samples 15000 unjoint_samples 15000 joint_samples 44 [1046704, 767828] processed_samples 15002 unjoint_samples 15000 joint_samples 45 [1047665, 995635] [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure processed_samples 15002 unjoint_samples 15000 joint_samples 45 [1047665, 995635] [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure processed_samples 15100 unjoint_samples 15100 joint_samples 45 [1046704, 157225] processed_samples 15100 unjoint_samples 15100 joint_samples 45 [1046704, 157225] [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure processed_samples 15100 unjoint_samples 15100 joint_samples 46 [1031365, 346748] processed_samples 15100 unjoint_samples 15100 joint_samples 46 [1031365, 346748] processed_samples 15100 unjoint_samples 15100 joint_samples 45 [1003629, 687210] processed_samples 15101 unjoint_samples 15100 joint_samples 45 [395632, 1045614] processed_samples 15101 unjoint_samples 15100 joint_samples 46 [626247, 1036416] processed_samples 15100 unjoint_samples 15100 joint_samples 45 [1037537, 819554] processed_samples 15101 unjoint_samples 15100 joint_samples 45 [395632, 1045614] processed_samples 15100 unjoint_samples 15100 joint_samples 45 [1037537, 819554] processed_samples 15100 unjoint_samples 15100 joint_samples 46 [1047956, 481841] processed_samples 15101 unjoint_samples 15100 joint_samples 46 [626247, 1036416] processed_samples 15100 unjoint_samples 15100 joint_samples 45 [1003629, 687210] processed_samples 15102 unjoint_samples 15100 joint_samples 46 [259889, 1045545] processed_samples 15100 unjoint_samples 15100 joint_samples 46 [1047956, 481841] processed_samples 15102 unjoint_samples 15100 joint_samples 46 [259889, 1045545] [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617e8e4ff40] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure processed_samples 15200 unjoint_samples 15200 joint_samples 45 [1046704, 474767] processed_samples 15200 unjoint_samples 15200 joint_samples 45 [1046704, 474767] processed_samples 15200 unjoint_samples 15200 joint_samples 46 [83364, 1047500] processed_samples 15200 unjoint_samples 15200 joint_samples 46 [83364, 1047500] processed_samples 15201 unjoint_samples 15200 joint_samples 45 [697590, 1045614] processed_samples 15201 unjoint_samples 15200 joint_samples 46 [923122, 1036416] processed_samples 15200 unjoint_samples 15200 joint_samples 46 [1031365, 778238] processed_samples 15201 unjoint_samples 15200 joint_samples 45 [697590, 1045614] processed_samples 15200 unjoint_samples 15200 joint_samples 46 [1047956, 721572] processed_samples 15200 unjoint_samples 15200 joint_samples 46 [1047956, 721572] processed_samples 15200 unjoint_samples 15200 joint_samples 46 [1031365, 778238] processed_samples 15200 unjoint_samples 15200 joint_samples 45 [1003629, 962960] processed_samples 15200 unjoint_samples 15200 joint_samples 45 [1003629, 962960] processed_samples 15201 unjoint_samples 15200 joint_samples 46 [923122, 1036416] processed_samples 15202 unjoint_samples 15200 joint_samples 46 [637475, 1045545] processed_samples 15202 unjoint_samples 15200 joint_samples 46 [637475, 1045545] [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure processed_samples 15300 unjoint_samples 15300 joint_samples 45 [1046704, 797591] [h264 @ 0x5617eeca6a40] mmco: unref short failure processed_samples 15300 unjoint_samples 15300 joint_samples 45 [1046704, 797591] processed_samples 15300 unjoint_samples 15300 joint_samples 46 [101479, 1046408] [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure processed_samples 15300 unjoint_samples 15300 joint_samples 46 [101479, 1046408] processed_samples 15300 unjoint_samples 15300 joint_samples 47 [1043531, 196050] processed_samples 15300 unjoint_samples 15300 joint_samples 47 [1043531, 196050] processed_samples 15300 unjoint_samples 15300 joint_samples 46 [498681, 1047500] processed_samples 15300 unjoint_samples 15300 joint_samples 46 [498681, 1047500] processed_samples 15301 unjoint_samples 15300 joint_samples 47 [1047331, 167413] processed_samples 15301 unjoint_samples 15300 joint_samples 47 [1047331, 167413] processed_samples 15301 unjoint_samples 15300 joint_samples 46 [1035655, 64451] processed_samples 15301 unjoint_samples 15300 joint_samples 46 [1035655, 64451] processed_samples 15300 unjoint_samples 15300 joint_samples 47 [3482, 1047927] processed_samples 15300 unjoint_samples 15300 joint_samples 47 [3482, 1047927] processed_samples 15302 unjoint_samples 15300 joint_samples 46 [977531, 1045545] processed_samples 15302 unjoint_samples 15300 joint_samples 46 [977531, 1045545] [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x55961adc42c0] mmco: unref short failure [h264 @ 0x55961adc42c0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x55961adc42c0] mmco: unref short failure [h264 @ 0x55961adc42c0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e8e0d640] mmco: unref short failure [h264 @ 0x5617e8e0d640] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure processed_samples 15400 unjoint_samples 15400 joint_samples 47 [304962, 1047927] processed_samples 15400 unjoint_samples 15400 joint_samples 46 [441576, 1046408] processed_samples 15400 unjoint_samples 15400 joint_samples 47 [304962, 1047927] processed_samples 15400 unjoint_samples 15400 joint_samples 46 [441576, 1046408] [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure processed_samples 15400 unjoint_samples 15400 joint_samples 47 [1043531, 492170] processed_samples 15400 unjoint_samples 15400 joint_samples 47 [1043531, 492170] processed_samples 15401 unjoint_samples 15400 joint_samples 47 [1047331, 440444] processed_samples 15401 unjoint_samples 15400 joint_samples 47 [1047331, 440444] processed_samples 15401 unjoint_samples 15400 joint_samples 46 [1035655, 338002] processed_samples 15401 unjoint_samples 15400 joint_samples 46 [1035655, 338002] processed_samples 15400 unjoint_samples 15400 joint_samples 45 [1048072, 1048205] processed_samples 15400 unjoint_samples 15400 joint_samples 45 [1048072, 1048205] processed_samples 15400 unjoint_samples 15400 joint_samples 46 [969578, 1047500] processed_samples 15402 unjoint_samples 15400 joint_samples 47 [979792, 379437] processed_samples 15400 unjoint_samples 15400 joint_samples 46 [969578, 1047500] processed_samples 15402 unjoint_samples 15400 joint_samples 47 [979792, 379437] [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure processed_samples 15500 unjoint_samples 15500 joint_samples 46 [1048072, 330183] processed_samples 15500 unjoint_samples 15500 joint_samples 47 [1035565, 209131] processed_samples 15500 unjoint_samples 15500 joint_samples 47 [1043531, 763288] processed_samples 15500 unjoint_samples 15500 joint_samples 47 [1035565, 209131] processed_samples 15500 unjoint_samples 15500 joint_samples 46 [1048072, 330183] processed_samples 15500 unjoint_samples 15500 joint_samples 47 [1043531, 763288] processed_samples 15500 unjoint_samples 15500 joint_samples 47 [649061, 1047927] [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure processed_samples 15501 unjoint_samples 15500 joint_samples 47 [1047331, 803421] processed_samples 15501 unjoint_samples 15500 joint_samples 46 [1035655, 612634] [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure processed_samples 15500 unjoint_samples 15500 joint_samples 47 [649061, 1047927] processed_samples 15500 unjoint_samples 15500 joint_samples 46 [767952, 1046408] processed_samples 15502 unjoint_samples 15500 joint_samples 47 [979792, 786920] processed_samples 15501 unjoint_samples 15500 joint_samples 46 [1035655, 612634] processed_samples 15501 unjoint_samples 15500 joint_samples 47 [1047331, 803421] processed_samples 15500 unjoint_samples 15500 joint_samples 46 [767952, 1046408] processed_samples 15502 unjoint_samples 15500 joint_samples 47 [979792, 786920] [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617ea25da40] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure processed_samples 15600 unjoint_samples 15600 joint_samples 46 [1048072, 692221] processed_samples 15600 unjoint_samples 15600 joint_samples 47 [1028465, 248504] processed_samples 15600 unjoint_samples 15600 joint_samples 47 [1035565, 619741] [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure processed_samples 15601 unjoint_samples 15600 joint_samples 48 [1047331, 101391] processed_samples 15600 unjoint_samples 15600 joint_samples 47 [1043531, 1044127] processed_samples 15600 unjoint_samples 15600 joint_samples 47 [954882, 1047927] processed_samples 15601 unjoint_samples 15600 joint_samples 46 [1035655, 911028] processed_samples 15600 unjoint_samples 15600 joint_samples 46 [1048072, 692221] [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure processed_samples 15600 unjoint_samples 15600 joint_samples 47 [1028465, 248504] processed_samples 15601 unjoint_samples 15600 joint_samples 48 [1047331, 101391] processed_samples 15600 unjoint_samples 15600 joint_samples 47 [1035565, 619741] processed_samples 15602 unjoint_samples 15600 joint_samples 47 [1036173, 1035027] processed_samples 15600 unjoint_samples 15600 joint_samples 47 [1043531, 1044127] processed_samples 15600 unjoint_samples 15600 joint_samples 47 [954882, 1047927] processed_samples 15601 unjoint_samples 15600 joint_samples 46 [1035655, 911028] processed_samples 15602 unjoint_samples 15600 joint_samples 47 [1036173, 1035027] [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617e7e34fc0] mmco: unref short failure [h264 @ 0x5617e7e34fc0] mmco: unref short failure [h264 @ 0x5617e7e34fc0] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e8298040] mmco: unref short failure [h264 @ 0x5617e8298040] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure processed_samples 15700 unjoint_samples 15700 joint_samples 47 [1028465, 675501] processed_samples 15700 unjoint_samples 15700 joint_samples 46 [1048072, 994167] [h264 @ 0x5617ee8ce7c0] mmco: unref short failure processed_samples 15700 unjoint_samples 15700 joint_samples 47 [1028465, 675501] processed_samples 15700 unjoint_samples 15700 joint_samples 46 [1048072, 994167] processed_samples 15700 unjoint_samples 15700 joint_samples 48 [217887, 928088] processed_samples 15700 unjoint_samples 15700 joint_samples 48 [217887, 928088] [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure processed_samples 15700 unjoint_samples 15700 joint_samples 48 [180488, 1047927] processed_samples 15700 unjoint_samples 15700 joint_samples 48 [180488, 1047927] processed_samples 15701 unjoint_samples 15700 joint_samples 47 [1041396, 132187] processed_samples 15701 unjoint_samples 15700 joint_samples 48 [1047331, 412996] processed_samples 15701 unjoint_samples 15700 joint_samples 48 [1047331, 412996] processed_samples 15701 unjoint_samples 15700 joint_samples 47 [1041396, 132187] processed_samples 15700 unjoint_samples 15700 joint_samples 48 [312632, 1047857] processed_samples 15702 unjoint_samples 15700 joint_samples 48 [1046816, 251458] processed_samples 15700 unjoint_samples 15700 joint_samples 48 [312632, 1047857] processed_samples 15702 unjoint_samples 15700 joint_samples 48 [1046816, 251458] [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure processed_samples 15800 unjoint_samples 15800 joint_samples 47 [214803, 1048068] processed_samples 15800 unjoint_samples 15800 joint_samples 47 [214803, 1048068] processed_samples 15800 unjoint_samples 15800 joint_samples 47 [1028465, 937118] processed_samples 15801 unjoint_samples 15800 joint_samples 48 [1047331, 779182] processed_samples 15800 unjoint_samples 15800 joint_samples 48 [610485, 1047857] processed_samples 15801 unjoint_samples 15800 joint_samples 47 [1041396, 381237] processed_samples 15800 unjoint_samples 15800 joint_samples 48 [498402, 928088] processed_samples 15800 unjoint_samples 15800 joint_samples 47 [1028465, 937118] processed_samples 15802 unjoint_samples 15800 joint_samples 48 [1046816, 523284] processed_samples 15800 unjoint_samples 15800 joint_samples 48 [498402, 928088] processed_samples 15801 unjoint_samples 15800 joint_samples 48 [1047331, 779182] processed_samples 15800 unjoint_samples 15800 joint_samples 48 [610485, 1047857] processed_samples 15801 unjoint_samples 15800 joint_samples 47 [1041396, 381237] processed_samples 15800 unjoint_samples 15800 joint_samples 48 [464932, 1047927] processed_samples 15802 unjoint_samples 15800 joint_samples 48 [1046816, 523284] [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure processed_samples 15800 unjoint_samples 15800 joint_samples 48 [464932, 1047927] [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x559617100c80] illegal short term buffer state detected [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e7f6a140] illegal short term buffer state detected [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure processed_samples 15900 unjoint_samples 15900 joint_samples 48 [797566, 928088] processed_samples 15900 unjoint_samples 15900 joint_samples 47 [548351, 1048068] processed_samples 15900 unjoint_samples 15900 joint_samples 48 [925101, 1047857] processed_samples 15900 unjoint_samples 15900 joint_samples 48 [185223, 1026357] processed_samples 15901 unjoint_samples 15900 joint_samples 48 [1047331, 1014742] [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure processed_samples 15900 unjoint_samples 15900 joint_samples 48 [797566, 928088] processed_samples 15900 unjoint_samples 15900 joint_samples 48 [185223, 1026357] processed_samples 15900 unjoint_samples 15900 joint_samples 47 [548351, 1048068] processed_samples 15901 unjoint_samples 15900 joint_samples 47 [1041396, 765493] [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure processed_samples 15901 unjoint_samples 15900 joint_samples 48 [1047331, 1014742] [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure processed_samples 15900 unjoint_samples 15900 joint_samples 48 [925101, 1047857] [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure processed_samples 15901 unjoint_samples 15900 joint_samples 47 [1041396, 765493] processed_samples 15900 unjoint_samples 15900 joint_samples 48 [884294, 1047927] [h264 @ 0x5617e7fbb200] mmco: unref short failure processed_samples 15902 unjoint_samples 15900 joint_samples 48 [1046816, 852943] processed_samples 15900 unjoint_samples 15900 joint_samples 48 [884294, 1047927] [h264 @ 0x5596136e0d80] mmco: unref short failure processed_samples 15902 unjoint_samples 15900 joint_samples 48 [1046816, 852943] [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5596136e0d80] mmco: unref short failure [h264 @ 0x5596136e0d80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure processed_samples 16000 unjoint_samples 16000 joint_samples 49 [992818, 397226] processed_samples 16000 unjoint_samples 16000 joint_samples 49 [99858, 1047927] processed_samples 16000 unjoint_samples 16000 joint_samples 48 [591170, 1026357] [h264 @ 0x55961609e040] mmco: unref short failure processed_samples 16001 unjoint_samples 16000 joint_samples 49 [1047935, 282623] processed_samples 16000 unjoint_samples 16000 joint_samples 47 [818602, 1048068] processed_samples 16000 unjoint_samples 16000 joint_samples 48 [968023, 965916] processed_samples 16002 unjoint_samples 16000 joint_samples 49 [1046816, 139230] processed_samples 16001 unjoint_samples 16000 joint_samples 47 [1041396, 1021544] [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure processed_samples 16000 unjoint_samples 16000 joint_samples 48 [591170, 1026357] processed_samples 16000 unjoint_samples 16000 joint_samples 49 [99858, 1047927] processed_samples 16000 unjoint_samples 16000 joint_samples 49 [992818, 397226] processed_samples 16001 unjoint_samples 16000 joint_samples 49 [1047935, 282623] processed_samples 16000 unjoint_samples 16000 joint_samples 47 [818602, 1048068] processed_samples 16001 unjoint_samples 16000 joint_samples 47 [1041396, 1021544] [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure processed_samples 16000 unjoint_samples 16000 joint_samples 48 [968023, 965916] processed_samples 16002 unjoint_samples 16000 joint_samples 49 [1046816, 139230] [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x559618471cc0] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure FileNotFoundError datasets/LMM/lmms-lab/LLaVA-Video-178K/liwei_youtube_videos/videos/youtube_video_2024/ytb_GqeRnxSuLFI.mp4 FileNotFoundError datasets/LMM/lmms-lab/LLaVA-Video-178K/liwei_youtube_videos/videos/youtube_video_2024/ytb_GqeRnxSuLFI.mp4 [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure processed_samples 16100 unjoint_samples 16100 joint_samples 48 [1047083, 41223] processed_samples 16100 unjoint_samples 16100 joint_samples 48 [1047083, 41223] processed_samples 16100 unjoint_samples 16100 joint_samples 49 [361861, 1047927] processed_samples 16100 unjoint_samples 16100 joint_samples 49 [361861, 1047927] processed_samples 16100 unjoint_samples 16100 joint_samples 49 [992818, 625250] processed_samples 16101 unjoint_samples 16100 joint_samples 49 [281594, 1014403] processed_samples 16100 unjoint_samples 16100 joint_samples 49 [992818, 625250] processed_samples 16101 unjoint_samples 16100 joint_samples 49 [281594, 1014403] processed_samples 16101 unjoint_samples 16100 joint_samples 48 [1045913, 336679] processed_samples 16100 unjoint_samples 16100 joint_samples 48 [821004, 1026357] processed_samples 16100 unjoint_samples 16100 joint_samples 48 [821004, 1026357] processed_samples 16101 unjoint_samples 16100 joint_samples 48 [1045913, 336679] processed_samples 16101 unjoint_samples 16100 joint_samples 49 [1047935, 554793] processed_samples 16101 unjoint_samples 16100 joint_samples 49 [1047935, 554793] processed_samples 16102 unjoint_samples 16100 joint_samples 49 [1046816, 470860] processed_samples 16102 unjoint_samples 16100 joint_samples 49 [1046816, 470860] [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure processed_samples 16200 unjoint_samples 16200 joint_samples 49 [1036778, 177169] processed_samples 16200 unjoint_samples 16200 joint_samples 48 [1047083, 340705] processed_samples 16200 unjoint_samples 16200 joint_samples 49 [1002024, 1000351] [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure processed_samples 16201 unjoint_samples 16200 joint_samples 48 [1045913, 624408] processed_samples 16200 unjoint_samples 16200 joint_samples 49 [604934, 1047927] processed_samples 16201 unjoint_samples 16200 joint_samples 49 [545290, 1014403] processed_samples 16202 unjoint_samples 16200 joint_samples 49 [1046816, 704379] processed_samples 16201 unjoint_samples 16200 joint_samples 49 [1047935, 922280] processed_samples 16200 unjoint_samples 16200 joint_samples 49 [1036778, 177169] processed_samples 16200 unjoint_samples 16200 joint_samples 48 [1047083, 340705] processed_samples 16200 unjoint_samples 16200 joint_samples 49 [1002024, 1000351] processed_samples 16200 unjoint_samples 16200 joint_samples 49 [604934, 1047927] processed_samples 16201 unjoint_samples 16200 joint_samples 49 [545290, 1014403] processed_samples 16201 unjoint_samples 16200 joint_samples 48 [1045913, 624408] processed_samples 16201 unjoint_samples 16200 joint_samples 49 [1047935, 922280] processed_samples 16202 unjoint_samples 16200 joint_samples 49 [1046816, 704379] [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure processed_samples 16300 unjoint_samples 16300 joint_samples 49 [1036778, 471457] processed_samples 16300 unjoint_samples 16300 joint_samples 50 [1045248, 218514] processed_samples 16300 unjoint_samples 16300 joint_samples 48 [1047083, 723752] processed_samples 16301 unjoint_samples 16300 joint_samples 50 [1047935, 149062] processed_samples 16300 unjoint_samples 16300 joint_samples 49 [940881, 1047927] [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure processed_samples 16301 unjoint_samples 16300 joint_samples 48 [1045913, 911901] processed_samples 16301 unjoint_samples 16300 joint_samples 49 [849024, 1014403] [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure processed_samples 16302 unjoint_samples 16300 joint_samples 49 [1046816, 977824] processed_samples 16300 unjoint_samples 16300 joint_samples 49 [1036778, 471457] processed_samples 16300 unjoint_samples 16300 joint_samples 50 [1045248, 218514] processed_samples 16300 unjoint_samples 16300 joint_samples 48 [1047083, 723752] processed_samples 16301 unjoint_samples 16300 joint_samples 50 [1047935, 149062] processed_samples 16300 unjoint_samples 16300 joint_samples 49 [940881, 1047927] [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure processed_samples 16301 unjoint_samples 16300 joint_samples 48 [1045913, 911901] [h264 @ 0x5596134bb5c0] mmco: unref short failure processed_samples 16301 unjoint_samples 16300 joint_samples 49 [849024, 1014403] [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure processed_samples 16302 unjoint_samples 16300 joint_samples 49 [1046816, 977824] [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559616fdd040] mmco: unref short failure [h264 @ 0x559616fdd040] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x559616fdd040] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x559616fdd040] mmco: unref short failure [h264 @ 0x559616fdd040] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure processed_samples 16400 unjoint_samples 16400 joint_samples 50 [1045248, 456340] processed_samples 16400 unjoint_samples 16400 joint_samples 50 [1044932, 218295] processed_samples 16400 unjoint_samples 16400 joint_samples 50 [1045248, 456340] processed_samples 16400 unjoint_samples 16400 joint_samples 50 [1044932, 218295] processed_samples 16401 unjoint_samples 16400 joint_samples 49 [175635, 1045453] processed_samples 16401 unjoint_samples 16400 joint_samples 49 [175635, 1045453] [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure processed_samples 16400 unjoint_samples 16400 joint_samples 49 [1036778, 766099] processed_samples 16400 unjoint_samples 16400 joint_samples 49 [1036778, 766099] processed_samples 16400 unjoint_samples 16400 joint_samples 48 [1047083, 1000009] processed_samples 16400 unjoint_samples 16400 joint_samples 48 [1047083, 1000009] processed_samples 16401 unjoint_samples 16400 joint_samples 50 [109523, 1041155] processed_samples 16401 unjoint_samples 16400 joint_samples 50 [109523, 1041155] processed_samples 16402 unjoint_samples 16400 joint_samples 50 [1046816, 198883] processed_samples 16402 unjoint_samples 16400 joint_samples 50 [1046816, 198883] processed_samples 16401 unjoint_samples 16400 joint_samples 50 [1047935, 478351] processed_samples 16401 unjoint_samples 16400 joint_samples 50 [1047935, 478351] [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ed3f3040] mmco: unref short failure [h264 @ 0x5617ed3f3040] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559617afc780] mmco: unref short failure [h264 @ 0x559617afc780] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure processed_samples 16500 unjoint_samples 16500 joint_samples 49 [1047083, 219901] [h264 @ 0x5617e7c5b300] mmco: unref short failure processed_samples 16500 unjoint_samples 16500 joint_samples 50 [1044932, 658728] [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure processed_samples 16501 unjoint_samples 16500 joint_samples 50 [1047935, 757637] processed_samples 16500 unjoint_samples 16500 joint_samples 50 [1045248, 893812] processed_samples 16501 unjoint_samples 16500 joint_samples 49 [492538, 1045453] processed_samples 16501 unjoint_samples 16500 joint_samples 50 [548333, 1041155] processed_samples 16500 unjoint_samples 16500 joint_samples 49 [1047083, 219901] processed_samples 16500 unjoint_samples 16500 joint_samples 50 [11261, 1040595] [h264 @ 0x559618693e80] mmco: unref short failure processed_samples 16500 unjoint_samples 16500 joint_samples 50 [1044932, 658728] processed_samples 16501 unjoint_samples 16500 joint_samples 50 [1047935, 757637] processed_samples 16500 unjoint_samples 16500 joint_samples 50 [1045248, 893812] processed_samples 16501 unjoint_samples 16500 joint_samples 49 [492538, 1045453] processed_samples 16501 unjoint_samples 16500 joint_samples 50 [548333, 1041155] processed_samples 16500 unjoint_samples 16500 joint_samples 50 [11261, 1040595] [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure processed_samples 16502 unjoint_samples 16500 joint_samples 50 [1046816, 526802] processed_samples 16502 unjoint_samples 16500 joint_samples 50 [1046816, 526802] [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559616fdd040] mmco: unref short failure [h264 @ 0x559616fdd040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e8e0d640] mmco: unref short failure [h264 @ 0x5617e8e0d640] mmco: unref short failure [h264 @ 0x5617e8e0d640] mmco: unref short failure [h264 @ 0x5617e8e0d640] mmco: unref short failure [h264 @ 0x5617e8e0d640] mmco: unref short failure [h264 @ 0x5617e8e0d640] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure processed_samples 16600 unjoint_samples 16600 joint_samples 51 [1047406, 108175] processed_samples 16600 unjoint_samples 16600 joint_samples 50 [244924, 1040595] processed_samples 16600 unjoint_samples 16600 joint_samples 49 [1047083, 552269] [h264 @ 0x559613a0c140] mmco: unref short failure processed_samples 16601 unjoint_samples 16600 joint_samples 49 [763926, 1045453] processed_samples 16600 unjoint_samples 16600 joint_samples 50 [1044932, 925191] processed_samples 16601 unjoint_samples 16600 joint_samples 50 [790605, 1041155] processed_samples 16601 unjoint_samples 16600 joint_samples 51 [92864, 1026792] processed_samples 16600 unjoint_samples 16600 joint_samples 51 [1047406, 108175] processed_samples 16600 unjoint_samples 16600 joint_samples 50 [244924, 1040595] processed_samples 16600 unjoint_samples 16600 joint_samples 49 [1047083, 552269] processed_samples 16601 unjoint_samples 16600 joint_samples 49 [763926, 1045453] processed_samples 16600 unjoint_samples 16600 joint_samples 50 [1044932, 925191] processed_samples 16601 unjoint_samples 16600 joint_samples 50 [790605, 1041155] processed_samples 16601 unjoint_samples 16600 joint_samples 51 [92864, 1026792] [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure processed_samples 16602 unjoint_samples 16600 joint_samples 50 [1046816, 887278] processed_samples 16602 unjoint_samples 16600 joint_samples 50 [1046816, 887278] [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617e9262740] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure processed_samples 16700 unjoint_samples 16700 joint_samples 51 [145659, 1047672] processed_samples 16700 unjoint_samples 16700 joint_samples 51 [145659, 1047672] processed_samples 16700 unjoint_samples 16700 joint_samples 51 [1047406, 499973] processed_samples 16700 unjoint_samples 16700 joint_samples 51 [1047406, 499973] processed_samples 16701 unjoint_samples 16700 joint_samples 50 [1046992, 28580] processed_samples 16701 unjoint_samples 16700 joint_samples 50 [1046992, 28580] processed_samples 16701 unjoint_samples 16700 joint_samples 51 [392601, 1026792] processed_samples 16701 unjoint_samples 16700 joint_samples 51 [392601, 1026792] processed_samples 16701 unjoint_samples 16700 joint_samples 51 [1046714, 24431] processed_samples 16701 unjoint_samples 16700 joint_samples 51 [1046714, 24431] processed_samples 16700 unjoint_samples 16700 joint_samples 49 [1047083, 820106] processed_samples 16700 unjoint_samples 16700 joint_samples 49 [1047083, 820106] processed_samples 16700 unjoint_samples 16700 joint_samples 50 [648674, 1040595] processed_samples 16700 unjoint_samples 16700 joint_samples 50 [648674, 1040595] processed_samples 16702 unjoint_samples 16700 joint_samples 51 [1046816, 123529] processed_samples 16702 unjoint_samples 16700 joint_samples 51 [1046816, 123529] [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559617ae1f00] mmco: unref short failure [h264 @ 0x559617ae1f00] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e7cb59c0] mmco: unref short failure [h264 @ 0x5617e7cb59c0] mmco: unref short failure [h264 @ 0x5617e7cb59c0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure processed_samples 16800 unjoint_samples 16800 joint_samples 50 [1047083, 133896] processed_samples 16800 unjoint_samples 16800 joint_samples 50 [1047083, 133896] processed_samples 16800 unjoint_samples 16800 joint_samples 51 [379736, 1047672] [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure processed_samples 16801 unjoint_samples 16800 joint_samples 50 [1046992, 395880] processed_samples 16800 unjoint_samples 16800 joint_samples 50 [892582, 1040595] processed_samples 16800 unjoint_samples 16800 joint_samples 51 [379736, 1047672] processed_samples 16800 unjoint_samples 16800 joint_samples 51 [1047406, 820738] processed_samples 16800 unjoint_samples 16800 joint_samples 51 [1047406, 820738] processed_samples 16801 unjoint_samples 16800 joint_samples 51 [1046714, 278988] processed_samples 16801 unjoint_samples 16800 joint_samples 50 [1046992, 395880] processed_samples 16800 unjoint_samples 16800 joint_samples 50 [892582, 1040595] processed_samples 16802 unjoint_samples 16800 joint_samples 51 [1046816, 458250] processed_samples 16801 unjoint_samples 16800 joint_samples 51 [688615, 1026792] processed_samples 16802 unjoint_samples 16800 joint_samples 51 [1046816, 458250] processed_samples 16801 unjoint_samples 16800 joint_samples 51 [1046714, 278988] processed_samples 16801 unjoint_samples 16800 joint_samples 51 [688615, 1026792] [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617e8298040] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617e8298040] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure processed_samples 16900 unjoint_samples 16900 joint_samples 51 [1044496, 123571] processed_samples 16900 unjoint_samples 16900 joint_samples 52 [185349, 1036982] processed_samples 16900 unjoint_samples 16900 joint_samples 50 [1047083, 624763] processed_samples 16900 unjoint_samples 16900 joint_samples 50 [1047083, 624763] [h264 @ 0x55961609e040] mmco: unref short failure processed_samples 16901 unjoint_samples 16900 joint_samples 51 [1046714, 557791] processed_samples 16900 unjoint_samples 16900 joint_samples 51 [1044496, 123571] processed_samples 16900 unjoint_samples 16900 joint_samples 52 [185349, 1036982] processed_samples 16900 unjoint_samples 16900 joint_samples 51 [738808, 1047672] processed_samples 16901 unjoint_samples 16900 joint_samples 51 [1046635, 1047195] processed_samples 16902 unjoint_samples 16900 joint_samples 51 [1046816, 765999] processed_samples 16900 unjoint_samples 16900 joint_samples 51 [738808, 1047672] processed_samples 16901 unjoint_samples 16900 joint_samples 50 [1046992, 697410] processed_samples 16901 unjoint_samples 16900 joint_samples 50 [1046992, 697410] processed_samples 16901 unjoint_samples 16900 joint_samples 51 [1046714, 557791] processed_samples 16902 unjoint_samples 16900 joint_samples 51 [1046816, 765999] processed_samples 16901 unjoint_samples 16900 joint_samples 51 [1046635, 1047195] [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure processed_samples 17000 unjoint_samples 17000 joint_samples 51 [1044496, 382607] processed_samples 17000 unjoint_samples 17000 joint_samples 51 [1044496, 382607] processed_samples 17000 unjoint_samples 17000 joint_samples 50 [1047083, 937355] processed_samples 17000 unjoint_samples 17000 joint_samples 50 [1047083, 937355] processed_samples 17000 unjoint_samples 17000 joint_samples 52 [461884, 1036982] processed_samples 17000 unjoint_samples 17000 joint_samples 52 [461884, 1036982] processed_samples 17001 unjoint_samples 17000 joint_samples 51 [1046714, 824294] processed_samples 17001 unjoint_samples 17000 joint_samples 51 [1046714, 824294] [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure processed_samples 17001 unjoint_samples 17000 joint_samples 51 [1046992, 8618] processed_samples 17001 unjoint_samples 17000 joint_samples 51 [1046992, 8618] processed_samples 17002 unjoint_samples 17000 joint_samples 52 [1047062, 28000] processed_samples 17002 unjoint_samples 17000 joint_samples 52 [1047062, 28000] processed_samples 17000 unjoint_samples 17000 joint_samples 52 [50245, 1047672] processed_samples 17000 unjoint_samples 17000 joint_samples 52 [50245, 1047672] processed_samples 17001 unjoint_samples 17000 joint_samples 52 [1046635, 381163] processed_samples 17001 unjoint_samples 17000 joint_samples 52 [1046635, 381163] [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559618471cc0] mmco: unref short failure [h264 @ 0x559618471cc0] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure processed_samples 17101 unjoint_samples 17100 joint_samples 52 [136546, 1041294] processed_samples 17100 unjoint_samples 17100 joint_samples 51 [240579, 1046421] processed_samples 17100 unjoint_samples 17100 joint_samples 51 [1044496, 684982] processed_samples 17100 unjoint_samples 17100 joint_samples 51 [240579, 1046421] processed_samples 17101 unjoint_samples 17100 joint_samples 52 [136546, 1041294] processed_samples 17101 unjoint_samples 17100 joint_samples 51 [1046992, 345714] processed_samples 17100 unjoint_samples 17100 joint_samples 52 [352177, 1047672] processed_samples 17101 unjoint_samples 17100 joint_samples 51 [1046992, 345714] processed_samples 17100 unjoint_samples 17100 joint_samples 52 [768395, 1036982] processed_samples 17101 unjoint_samples 17100 joint_samples 52 [1046635, 653406] processed_samples 17100 unjoint_samples 17100 joint_samples 51 [1044496, 684982] processed_samples 17100 unjoint_samples 17100 joint_samples 52 [768395, 1036982] processed_samples 17100 unjoint_samples 17100 joint_samples 52 [352177, 1047672] processed_samples 17101 unjoint_samples 17100 joint_samples 52 [1046635, 653406] processed_samples 17102 unjoint_samples 17100 joint_samples 52 [1047062, 349756] processed_samples 17102 unjoint_samples 17100 joint_samples 52 [1047062, 349756] [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617ea25da40] mmco: unref short failure [h264 @ 0x5617ea25da40] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure processed_samples 17200 unjoint_samples 17200 joint_samples 51 [547829, 1046421] processed_samples 17200 unjoint_samples 17200 joint_samples 51 [547829, 1046421] processed_samples 17200 unjoint_samples 17200 joint_samples 52 [1046651, 11964] processed_samples 17200 unjoint_samples 17200 joint_samples 52 [1046651, 11964] processed_samples 17201 unjoint_samples 17200 joint_samples 52 [497388, 1041294] processed_samples 17200 unjoint_samples 17200 joint_samples 52 [728568, 1047672] processed_samples 17200 unjoint_samples 17200 joint_samples 52 [728568, 1047672] processed_samples 17201 unjoint_samples 17200 joint_samples 52 [497388, 1041294] processed_samples 17201 unjoint_samples 17200 joint_samples 51 [1046992, 633927] processed_samples 17201 unjoint_samples 17200 joint_samples 51 [1046992, 633927] processed_samples 17200 unjoint_samples 17200 joint_samples 52 [1016813, 1036982] [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure processed_samples 17200 unjoint_samples 17200 joint_samples 52 [1016813, 1036982] [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure processed_samples 17201 unjoint_samples 17200 joint_samples 52 [1046635, 980360] processed_samples 17201 unjoint_samples 17200 joint_samples 52 [1046635, 980360] processed_samples 17202 unjoint_samples 17200 joint_samples 52 [1047062, 615800] processed_samples 17202 unjoint_samples 17200 joint_samples 52 [1047062, 615800] [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure processed_samples 17300 unjoint_samples 17300 joint_samples 52 [1046651, 369511] processed_samples 17300 unjoint_samples 17300 joint_samples 52 [1046651, 369511] processed_samples 17300 unjoint_samples 17300 joint_samples 53 [1048499, 326192] processed_samples 17300 unjoint_samples 17300 joint_samples 53 [1048499, 326192] processed_samples 17300 unjoint_samples 17300 joint_samples 51 [872797, 1046421] [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure processed_samples 17300 unjoint_samples 17300 joint_samples 51 [872797, 1046421] processed_samples 17300 unjoint_samples 17300 joint_samples 53 [961210, 148781] [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure processed_samples 17301 unjoint_samples 17300 joint_samples 53 [1046635, 221184] processed_samples 17300 unjoint_samples 17300 joint_samples 53 [961210, 148781] [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure processed_samples 17301 unjoint_samples 17300 joint_samples 53 [1046635, 221184] processed_samples 17301 unjoint_samples 17300 joint_samples 52 [843106, 1041294] processed_samples 17301 unjoint_samples 17300 joint_samples 52 [843106, 1041294] processed_samples 17301 unjoint_samples 17300 joint_samples 51 [1046992, 872336] processed_samples 17302 unjoint_samples 17300 joint_samples 52 [1047062, 980438] [h264 @ 0x55961845fd00] mmco: unref short failure processed_samples 17301 unjoint_samples 17300 joint_samples 51 [1046992, 872336] [h264 @ 0x5617e9274140] mmco: unref short failure processed_samples 17302 unjoint_samples 17300 joint_samples 52 [1047062, 980438] [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ed3f3040] mmco: unref short failure [h264 @ 0x5617ed3f3040] mmco: unref short failure [h264 @ 0x559618471cc0] mmco: unref short failure [h264 @ 0x559618471cc0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure processed_samples 17400 unjoint_samples 17400 joint_samples 52 [1027912, 191642] processed_samples 17400 unjoint_samples 17400 joint_samples 52 [1027912, 191642] processed_samples 17400 unjoint_samples 17400 joint_samples 52 [1046651, 792021] processed_samples 17400 unjoint_samples 17400 joint_samples 52 [1046651, 792021] [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure processed_samples 17401 unjoint_samples 17400 joint_samples 52 [117209, 1046487] processed_samples 17401 unjoint_samples 17400 joint_samples 53 [1046635, 489580] processed_samples 17401 unjoint_samples 17400 joint_samples 53 [166943, 1045652] processed_samples 17401 unjoint_samples 17400 joint_samples 52 [117209, 1046487] processed_samples 17401 unjoint_samples 17400 joint_samples 53 [166943, 1045652] processed_samples 17400 unjoint_samples 17400 joint_samples 53 [1048499, 740100] processed_samples 17400 unjoint_samples 17400 joint_samples 53 [961210, 484954] processed_samples 17400 unjoint_samples 17400 joint_samples 53 [961210, 484954] processed_samples 17400 unjoint_samples 17400 joint_samples 53 [1048499, 740100] processed_samples 17401 unjoint_samples 17400 joint_samples 53 [1046635, 489580] processed_samples 17402 unjoint_samples 17400 joint_samples 53 [260868, 1044620] processed_samples 17402 unjoint_samples 17400 joint_samples 53 [260868, 1044620] [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e803bd40] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617e803bd40] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure processed_samples 17500 unjoint_samples 17500 joint_samples 53 [24508, 1041150] processed_samples 17500 unjoint_samples 17500 joint_samples 53 [24508, 1041150] processed_samples 17500 unjoint_samples 17500 joint_samples 54 [41001, 1039022] processed_samples 17500 unjoint_samples 17500 joint_samples 54 [41001, 1039022] processed_samples 17500 unjoint_samples 17500 joint_samples 52 [1027912, 635881] processed_samples 17500 unjoint_samples 17500 joint_samples 52 [1027912, 635881] processed_samples 17501 unjoint_samples 17500 joint_samples 53 [434362, 1045652] processed_samples 17500 unjoint_samples 17500 joint_samples 53 [961210, 797384] processed_samples 17500 unjoint_samples 17500 joint_samples 53 [961210, 797384] processed_samples 17501 unjoint_samples 17500 joint_samples 53 [434362, 1045652] processed_samples 17501 unjoint_samples 17500 joint_samples 52 [430304, 1046487] processed_samples 17501 unjoint_samples 17500 joint_samples 52 [430304, 1046487] processed_samples 17501 unjoint_samples 17500 joint_samples 53 [1046635, 858072] processed_samples 17501 unjoint_samples 17500 joint_samples 53 [1046635, 858072] processed_samples 17502 unjoint_samples 17500 joint_samples 53 [462250, 1044620] processed_samples 17502 unjoint_samples 17500 joint_samples 53 [462250, 1044620] [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617e7e34fc0] mmco: unref short failure [h264 @ 0x5617e7e34fc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e7e34fc0] mmco: unref short failure [h264 @ 0x5617e7e34fc0] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure processed_samples 17600 unjoint_samples 17600 joint_samples 53 [360580, 1041150] processed_samples 17600 unjoint_samples 17600 joint_samples 54 [318265, 1039022] processed_samples 17600 unjoint_samples 17600 joint_samples 53 [997513, 1005936] processed_samples 17600 unjoint_samples 17600 joint_samples 52 [1027912, 997443] processed_samples 17601 unjoint_samples 17600 joint_samples 54 [71505, 1039435] processed_samples 17601 unjoint_samples 17600 joint_samples 52 [719273, 1046487] processed_samples 17602 unjoint_samples 17600 joint_samples 53 [828286, 1044620] processed_samples 17601 unjoint_samples 17600 joint_samples 53 [818273, 1045652] processed_samples 17600 unjoint_samples 17600 joint_samples 54 [318265, 1039022] processed_samples 17600 unjoint_samples 17600 joint_samples 53 [360580, 1041150] [h264 @ 0x5617e8b75440] mmco: unref short failure processed_samples 17600 unjoint_samples 17600 joint_samples 53 [997513, 1005936] processed_samples 17600 unjoint_samples 17600 joint_samples 52 [1027912, 997443] processed_samples 17601 unjoint_samples 17600 joint_samples 54 [71505, 1039435] processed_samples 17601 unjoint_samples 17600 joint_samples 52 [719273, 1046487] processed_samples 17602 unjoint_samples 17600 joint_samples 53 [828286, 1044620] [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure processed_samples 17601 unjoint_samples 17600 joint_samples 53 [818273, 1045652] [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure processed_samples 17700 unjoint_samples 17700 joint_samples 53 [1045077, 217125] processed_samples 17700 unjoint_samples 17700 joint_samples 53 [1045077, 217125] processed_samples 17701 unjoint_samples 17700 joint_samples 54 [372245, 1039435] processed_samples 17701 unjoint_samples 17700 joint_samples 54 [372245, 1039435] processed_samples 17700 unjoint_samples 17700 joint_samples 54 [1045394, 194274] processed_samples 17701 unjoint_samples 17700 joint_samples 54 [132311, 1046616] processed_samples 17700 unjoint_samples 17700 joint_samples 54 [1045394, 194274] processed_samples 17701 unjoint_samples 17700 joint_samples 54 [132311, 1046616] processed_samples 17700 unjoint_samples 17700 joint_samples 53 [773857, 1041150] [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure processed_samples 17700 unjoint_samples 17700 joint_samples 53 [773857, 1041150] [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure processed_samples 17700 unjoint_samples 17700 joint_samples 54 [649180, 1039022] processed_samples 17702 unjoint_samples 17700 joint_samples 54 [140713, 1046563] processed_samples 17700 unjoint_samples 17700 joint_samples 54 [649180, 1039022] processed_samples 17702 unjoint_samples 17700 joint_samples 54 [140713, 1046563] processed_samples 17701 unjoint_samples 17700 joint_samples 52 [1043044, 1046487] processed_samples 17701 unjoint_samples 17700 joint_samples 52 [1043044, 1046487] [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure processed_samples 17800 unjoint_samples 17800 joint_samples 53 [1045077, 490190] processed_samples 17800 unjoint_samples 17800 joint_samples 54 [1031055, 39511] processed_samples 17800 unjoint_samples 17800 joint_samples 54 [1045394, 580643] processed_samples 17800 unjoint_samples 17800 joint_samples 54 [1010285, 1039022] processed_samples 17801 unjoint_samples 17800 joint_samples 54 [614867, 1039435] processed_samples 17800 unjoint_samples 17800 joint_samples 53 [1045077, 490190] processed_samples 17800 unjoint_samples 17800 joint_samples 54 [1045394, 580643] processed_samples 17800 unjoint_samples 17800 joint_samples 54 [1010285, 1039022] processed_samples 17800 unjoint_samples 17800 joint_samples 54 [1031055, 39511] processed_samples 17801 unjoint_samples 17800 joint_samples 54 [614867, 1039435] processed_samples 17801 unjoint_samples 17800 joint_samples 53 [355536, 1046487] processed_samples 17802 unjoint_samples 17800 joint_samples 54 [397147, 1046563] processed_samples 17801 unjoint_samples 17800 joint_samples 53 [355536, 1046487] processed_samples 17802 unjoint_samples 17800 joint_samples 54 [397147, 1046563] processed_samples 17801 unjoint_samples 17800 joint_samples 54 [446765, 1046616] processed_samples 17801 unjoint_samples 17800 joint_samples 54 [446765, 1046616] [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure processed_samples 17900 unjoint_samples 17900 joint_samples 54 [1031055, 306440] [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure processed_samples 17900 unjoint_samples 17900 joint_samples 53 [1045077, 768145] [h264 @ 0x5617e90f0680] mmco: unref short failure processed_samples 17900 unjoint_samples 17900 joint_samples 55 [1036237, 282416] processed_samples 17901 unjoint_samples 17900 joint_samples 54 [978782, 1039435] processed_samples 17900 unjoint_samples 17900 joint_samples 54 [1045394, 992673] processed_samples 17901 unjoint_samples 17900 joint_samples 53 [682192, 1046487] processed_samples 17902 unjoint_samples 17900 joint_samples 54 [699009, 1046563] processed_samples 17901 unjoint_samples 17900 joint_samples 54 [722315, 1046616] processed_samples 17900 unjoint_samples 17900 joint_samples 54 [1031055, 306440] [h264 @ 0x5617ec342580] mmco: unref short failure processed_samples 17900 unjoint_samples 17900 joint_samples 55 [1036237, 282416] processed_samples 17901 unjoint_samples 17900 joint_samples 54 [978782, 1039435] processed_samples 17900 unjoint_samples 17900 joint_samples 53 [1045077, 768145] [h264 @ 0x55961672f480] mmco: unref short failure processed_samples 17900 unjoint_samples 17900 joint_samples 54 [1045394, 992673] processed_samples 17901 unjoint_samples 17900 joint_samples 53 [682192, 1046487] processed_samples 17902 unjoint_samples 17900 joint_samples 54 [699009, 1046563] processed_samples 17901 unjoint_samples 17900 joint_samples 54 [722315, 1046616] [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559609414580] mmco: unref short failure [h264 @ 0x559609414580] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure processed_samples 18000 unjoint_samples 18000 joint_samples 54 [1031055, 539036] processed_samples 18000 unjoint_samples 18000 joint_samples 55 [1036237, 577629] processed_samples 18000 unjoint_samples 18000 joint_samples 55 [1045394, 267829] processed_samples 18000 unjoint_samples 18000 joint_samples 53 [1045077, 1042129] processed_samples 18001 unjoint_samples 18000 joint_samples 55 [1035612, 247435] [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure processed_samples 18001 unjoint_samples 18000 joint_samples 54 [995506, 1046616] processed_samples 18001 unjoint_samples 18000 joint_samples 53 [951016, 1046487] processed_samples 18002 unjoint_samples 18000 joint_samples 54 [989846, 1046563] [h264 @ 0x5596139c8400] mmco: unref short failure processed_samples 18000 unjoint_samples 18000 joint_samples 55 [1045394, 267829] processed_samples 18000 unjoint_samples 18000 joint_samples 54 [1031055, 539036] processed_samples 18000 unjoint_samples 18000 joint_samples 55 [1036237, 577629] processed_samples 18001 unjoint_samples 18000 joint_samples 55 [1035612, 247435] [h264 @ 0x5617eeca6a40] mmco: unref short failure processed_samples 18000 unjoint_samples 18000 joint_samples 53 [1045077, 1042129] processed_samples 18001 unjoint_samples 18000 joint_samples 53 [951016, 1046487] processed_samples 18001 unjoint_samples 18000 joint_samples 54 [995506, 1046616] [h264 @ 0x5617e955dbc0] mmco: unref short failure processed_samples 18002 unjoint_samples 18000 joint_samples 54 [989846, 1046563] [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617ed3f3040] mmco: unref short failure [h264 @ 0x5617ed3f3040] mmco: unref short failure [h264 @ 0x55961adc42c0] mmco: unref short failure [h264 @ 0x55961adc42c0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure processed_samples 18100 unjoint_samples 18100 joint_samples 54 [1045077, 265775] processed_samples 18101 unjoint_samples 18100 joint_samples 55 [1035612, 544165] processed_samples 18100 unjoint_samples 18100 joint_samples 55 [1045394, 567325] processed_samples 18100 unjoint_samples 18100 joint_samples 55 [1036237, 885977] processed_samples 18100 unjoint_samples 18100 joint_samples 54 [1045077, 265775] processed_samples 18101 unjoint_samples 18100 joint_samples 54 [125568, 1046487] processed_samples 18101 unjoint_samples 18100 joint_samples 55 [1042868, 366761] processed_samples 18100 unjoint_samples 18100 joint_samples 54 [1031055, 810878] processed_samples 18100 unjoint_samples 18100 joint_samples 55 [1045394, 567325] processed_samples 18102 unjoint_samples 18100 joint_samples 55 [353922, 1046563] processed_samples 18101 unjoint_samples 18100 joint_samples 55 [1035612, 544165] processed_samples 18100 unjoint_samples 18100 joint_samples 55 [1036237, 885977] [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure processed_samples 18101 unjoint_samples 18100 joint_samples 54 [125568, 1046487] processed_samples 18101 unjoint_samples 18100 joint_samples 55 [1042868, 366761] processed_samples 18100 unjoint_samples 18100 joint_samples 54 [1031055, 810878] processed_samples 18102 unjoint_samples 18100 joint_samples 55 [353922, 1046563] [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559616d35c80] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x55961adc42c0] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x559615d28940] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure processed_samples 18200 unjoint_samples 18200 joint_samples 54 [1045077, 530769] [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure processed_samples 18200 unjoint_samples 18200 joint_samples 56 [295084, 1048251] processed_samples 18200 unjoint_samples 18200 joint_samples 55 [1045394, 902449] processed_samples 18201 unjoint_samples 18200 joint_samples 55 [1035612, 838327] processed_samples 18201 unjoint_samples 18200 joint_samples 55 [1042868, 697207] processed_samples 18201 unjoint_samples 18200 joint_samples 54 [485707, 1046487] processed_samples 18200 unjoint_samples 18200 joint_samples 54 [1031055, 1024074] [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure processed_samples 18202 unjoint_samples 18200 joint_samples 55 [678922, 1046563] [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure processed_samples 18200 unjoint_samples 18200 joint_samples 54 [1045077, 530769] [h264 @ 0x55961c6febc0] mmco: unref short failure processed_samples 18200 unjoint_samples 18200 joint_samples 56 [295084, 1048251] processed_samples 18200 unjoint_samples 18200 joint_samples 55 [1045394, 902449] processed_samples 18201 unjoint_samples 18200 joint_samples 55 [1035612, 838327] processed_samples 18201 unjoint_samples 18200 joint_samples 55 [1042868, 697207] processed_samples 18201 unjoint_samples 18200 joint_samples 54 [485707, 1046487] processed_samples 18200 unjoint_samples 18200 joint_samples 54 [1031055, 1024074] [h264 @ 0x559617100c80] mmco: unref short failure processed_samples 18202 unjoint_samples 18200 joint_samples 55 [678922, 1046563] [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5596176fb900] mmco: unref short failure [h264 @ 0x5596176fb900] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559616d35c80] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure processed_samples 18300 unjoint_samples 18300 joint_samples 55 [278838, 1043840] processed_samples 18300 unjoint_samples 18300 joint_samples 55 [278838, 1043840] processed_samples 18300 unjoint_samples 18300 joint_samples 56 [180826, 1046974] processed_samples 18300 unjoint_samples 18300 joint_samples 56 [180826, 1046974] processed_samples 18300 unjoint_samples 18300 joint_samples 56 [539045, 1048251] processed_samples 18300 unjoint_samples 18300 joint_samples 56 [539045, 1048251] processed_samples 18301 unjoint_samples 18300 joint_samples 56 [1046020, 32741] processed_samples 18301 unjoint_samples 18300 joint_samples 56 [1046020, 32741] processed_samples 18300 unjoint_samples 18300 joint_samples 54 [1045077, 765991] processed_samples 18300 unjoint_samples 18300 joint_samples 54 [1045077, 765991] processed_samples 18301 unjoint_samples 18300 joint_samples 55 [1042868, 1029804] [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure processed_samples 18301 unjoint_samples 18300 joint_samples 55 [1042868, 1029804] processed_samples 18301 unjoint_samples 18300 joint_samples 54 [805036, 1046487] processed_samples 18301 unjoint_samples 18300 joint_samples 54 [805036, 1046487] [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure processed_samples 18302 unjoint_samples 18300 joint_samples 55 [947653, 1046563] processed_samples 18302 unjoint_samples 18300 joint_samples 55 [947653, 1046563] [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e9262740] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x559618a89040] mmco: unref short failure [h264 @ 0x559618a89040] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617e7e34fc0] mmco: unref short failure [h264 @ 0x5617e7e34fc0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596177086c0] mmco: unref short failure [h264 @ 0x5596177086c0] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure processed_samples 18400 unjoint_samples 18400 joint_samples 55 [94559, 1028525] processed_samples 18400 unjoint_samples 18400 joint_samples 55 [94559, 1028525] processed_samples 18400 unjoint_samples 18400 joint_samples 55 [633201, 1043840] processed_samples 18401 unjoint_samples 18400 joint_samples 56 [282866, 1031866] processed_samples 18401 unjoint_samples 18400 joint_samples 55 [19418, 1046487] processed_samples 18401 unjoint_samples 18400 joint_samples 56 [282866, 1031866] processed_samples 18400 unjoint_samples 18400 joint_samples 56 [484986, 1046974] processed_samples 18400 unjoint_samples 18400 joint_samples 55 [633201, 1043840] processed_samples 18400 unjoint_samples 18400 joint_samples 56 [484986, 1046974] processed_samples 18401 unjoint_samples 18400 joint_samples 56 [1046020, 379079] processed_samples 18401 unjoint_samples 18400 joint_samples 56 [1046020, 379079] processed_samples 18401 unjoint_samples 18400 joint_samples 55 [19418, 1046487] processed_samples 18400 unjoint_samples 18400 joint_samples 56 [859293, 1048251] [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure processed_samples 18402 unjoint_samples 18400 joint_samples 56 [167709, 1046563] processed_samples 18400 unjoint_samples 18400 joint_samples 56 [859293, 1048251] [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure processed_samples 18402 unjoint_samples 18400 joint_samples 56 [167709, 1046563] [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x55961cd41c40] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617e8298040] mmco: unref short failure [h264 @ 0x5617e8298040] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961adc42c0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure processed_samples 18500 unjoint_samples 18500 joint_samples 55 [476833, 1028525] processed_samples 18500 unjoint_samples 18500 joint_samples 57 [1046985, 218037] [h264 @ 0x5617e8298040] mmco: unref short failure processed_samples 18501 unjoint_samples 18500 joint_samples 56 [1046020, 655329] processed_samples 18501 unjoint_samples 18500 joint_samples 55 [402905, 1046487] processed_samples 18501 unjoint_samples 18500 joint_samples 56 [547490, 1031866] [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure processed_samples 18500 unjoint_samples 18500 joint_samples 56 [811595, 1046974] processed_samples 18500 unjoint_samples 18500 joint_samples 56 [1027579, 27914] processed_samples 18500 unjoint_samples 18500 joint_samples 57 [1046985, 218037] processed_samples 18500 unjoint_samples 18500 joint_samples 55 [476833, 1028525] processed_samples 18501 unjoint_samples 18500 joint_samples 55 [402905, 1046487] processed_samples 18501 unjoint_samples 18500 joint_samples 56 [547490, 1031866] processed_samples 18501 unjoint_samples 18500 joint_samples 56 [1046020, 655329] processed_samples 18500 unjoint_samples 18500 joint_samples 56 [1027579, 27914] processed_samples 18500 unjoint_samples 18500 joint_samples 56 [811595, 1046974] processed_samples 18502 unjoint_samples 18500 joint_samples 56 [619340, 1046563] processed_samples 18502 unjoint_samples 18500 joint_samples 56 [619340, 1046563] [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e8298040] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure processed_samples 18600 unjoint_samples 18600 joint_samples 55 [754809, 1028525] [h264 @ 0x559614220440] mmco: unref short failure processed_samples 18600 unjoint_samples 18600 joint_samples 56 [1027579, 343031] processed_samples 18600 unjoint_samples 18600 joint_samples 57 [1046985, 633655] processed_samples 18600 unjoint_samples 18600 joint_samples 57 [1019691, 152289] [h264 @ 0x559613520f80] mmco: unref short failure processed_samples 18601 unjoint_samples 18600 joint_samples 55 [700859, 1046487] processed_samples 18601 unjoint_samples 18600 joint_samples 56 [1046020, 893900] processed_samples 18602 unjoint_samples 18600 joint_samples 56 [952148, 1046563] processed_samples 18600 unjoint_samples 18600 joint_samples 55 [754809, 1028525] processed_samples 18601 unjoint_samples 18600 joint_samples 56 [801729, 1031866] [h264 @ 0x5617e87a0740] mmco: unref short failure processed_samples 18600 unjoint_samples 18600 joint_samples 57 [1046985, 633655] processed_samples 18600 unjoint_samples 18600 joint_samples 56 [1027579, 343031] [h264 @ 0x5617ec0115c0] mmco: unref short failure processed_samples 18600 unjoint_samples 18600 joint_samples 57 [1019691, 152289] [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure processed_samples 18601 unjoint_samples 18600 joint_samples 55 [700859, 1046487] processed_samples 18601 unjoint_samples 18600 joint_samples 56 [801729, 1031866] processed_samples 18602 unjoint_samples 18600 joint_samples 56 [952148, 1046563] processed_samples 18601 unjoint_samples 18600 joint_samples 56 [1046020, 893900] [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure processed_samples 18700 unjoint_samples 18700 joint_samples 57 [1019691, 403440] processed_samples 18700 unjoint_samples 18700 joint_samples 57 [1019691, 403440] processed_samples 18700 unjoint_samples 18700 joint_samples 56 [1027579, 641948] processed_samples 18700 unjoint_samples 18700 joint_samples 58 [3409, 1046154] processed_samples 18700 unjoint_samples 18700 joint_samples 56 [1027579, 641948] processed_samples 18700 unjoint_samples 18700 joint_samples 58 [3409, 1046154] processed_samples 18700 unjoint_samples 18700 joint_samples 56 [1046854, 4373] processed_samples 18700 unjoint_samples 18700 joint_samples 56 [1046854, 4373] processed_samples 18701 unjoint_samples 18700 joint_samples 57 [238648, 1042004] processed_samples 18701 unjoint_samples 18700 joint_samples 57 [238648, 1042004] processed_samples 18701 unjoint_samples 18700 joint_samples 56 [1042235, 18237] processed_samples 18701 unjoint_samples 18700 joint_samples 56 [1042235, 18237] processed_samples 18701 unjoint_samples 18700 joint_samples 57 [956052, 463061] processed_samples 18701 unjoint_samples 18700 joint_samples 57 [956052, 463061] processed_samples 18702 unjoint_samples 18700 joint_samples 57 [170566, 1046563] processed_samples 18702 unjoint_samples 18700 joint_samples 57 [170566, 1046563] [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x55961b8e3e40] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x559618a89040] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure processed_samples 18800 unjoint_samples 18800 joint_samples 56 [1046854, 304297] processed_samples 18800 unjoint_samples 18800 joint_samples 56 [1046854, 304297] processed_samples 18800 unjoint_samples 18800 joint_samples 58 [280019, 1046154] processed_samples 18800 unjoint_samples 18800 joint_samples 56 [1027579, 901131] processed_samples 18800 unjoint_samples 18800 joint_samples 56 [1027579, 901131] processed_samples 18800 unjoint_samples 18800 joint_samples 58 [280019, 1046154] processed_samples 18801 unjoint_samples 18800 joint_samples 57 [523438, 1042004] processed_samples 18801 unjoint_samples 18800 joint_samples 56 [1042235, 327537] processed_samples 18801 unjoint_samples 18800 joint_samples 57 [523438, 1042004] processed_samples 18800 unjoint_samples 18800 joint_samples 57 [1019691, 694700] processed_samples 18800 unjoint_samples 18800 joint_samples 57 [1019691, 694700] processed_samples 18801 unjoint_samples 18800 joint_samples 57 [956052, 813763] processed_samples 18801 unjoint_samples 18800 joint_samples 56 [1042235, 327537] processed_samples 18802 unjoint_samples 18800 joint_samples 57 [422810, 1046563] processed_samples 18802 unjoint_samples 18800 joint_samples 57 [422810, 1046563] processed_samples 18801 unjoint_samples 18800 joint_samples 57 [956052, 813763] [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617ee580e80] Invalid NAL unit size (1113487007 > 108878). [h264 @ 0x5617ee580e80] Error splitting the input into NAL units. [h264 @ 0x55961342b000] Invalid NAL unit size (1113487007 > 108878). [h264 @ 0x55961342b000] Error splitting the input into NAL units. processed_samples 18900 unjoint_samples 18900 joint_samples 57 [346267, 1011104] processed_samples 18901 unjoint_samples 18900 joint_samples 58 [1041822, 111277] processed_samples 18900 unjoint_samples 18900 joint_samples 56 [1046854, 628969] processed_samples 18901 unjoint_samples 18900 joint_samples 56 [1042235, 583724] processed_samples 18901 unjoint_samples 18900 joint_samples 58 [551080, 1046154] processed_samples 18900 unjoint_samples 18900 joint_samples 57 [1019691, 921402] processed_samples 18901 unjoint_samples 18900 joint_samples 57 [870588, 1042004] processed_samples 18900 unjoint_samples 18900 joint_samples 57 [346267, 1011104] processed_samples 18902 unjoint_samples 18900 joint_samples 57 [818989, 1046563] processed_samples 18901 unjoint_samples 18900 joint_samples 58 [1041822, 111277] processed_samples 18900 unjoint_samples 18900 joint_samples 56 [1046854, 628969] processed_samples 18901 unjoint_samples 18900 joint_samples 56 [1042235, 583724] processed_samples 18900 unjoint_samples 18900 joint_samples 57 [1019691, 921402] processed_samples 18901 unjoint_samples 18900 joint_samples 57 [870588, 1042004] [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure processed_samples 18901 unjoint_samples 18900 joint_samples 58 [551080, 1046154] processed_samples 18902 unjoint_samples 18900 joint_samples 57 [818989, 1046563] [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5596187a1900] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617ea25da40] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e8e19000] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure processed_samples 19000 unjoint_samples 19000 joint_samples 57 [581563, 1011104] processed_samples 19001 unjoint_samples 19000 joint_samples 58 [1048140, 185649] processed_samples 19001 unjoint_samples 19000 joint_samples 58 [794067, 1046154] processed_samples 19000 unjoint_samples 19000 joint_samples 56 [1046854, 936939] processed_samples 19000 unjoint_samples 19000 joint_samples 57 [581563, 1011104] processed_samples 19002 unjoint_samples 19000 joint_samples 58 [1047707, 133472] processed_samples 19001 unjoint_samples 19000 joint_samples 58 [1041822, 434258] processed_samples 19000 unjoint_samples 19000 joint_samples 58 [1043877, 81220] processed_samples 19001 unjoint_samples 19000 joint_samples 58 [794067, 1046154] processed_samples 19000 unjoint_samples 19000 joint_samples 56 [1046854, 936939] processed_samples 19001 unjoint_samples 19000 joint_samples 58 [1048140, 185649] processed_samples 19001 unjoint_samples 19000 joint_samples 58 [1041822, 434258] [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure processed_samples 19002 unjoint_samples 19000 joint_samples 58 [1047707, 133472] processed_samples 19000 unjoint_samples 19000 joint_samples 58 [1043877, 81220] processed_samples 19001 unjoint_samples 19000 joint_samples 56 [1042235, 1002515] [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure processed_samples 19001 unjoint_samples 19000 joint_samples 56 [1042235, 1002515] [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure processed_samples 19100 unjoint_samples 19100 joint_samples 57 [326822, 1007324] processed_samples 19100 unjoint_samples 19100 joint_samples 57 [326822, 1007324] processed_samples 19100 unjoint_samples 19100 joint_samples 58 [1043877, 403731] processed_samples 19100 unjoint_samples 19100 joint_samples 57 [811755, 1011104] processed_samples 19100 unjoint_samples 19100 joint_samples 57 [811755, 1011104] processed_samples 19101 unjoint_samples 19100 joint_samples 58 [1048140, 423568] processed_samples 19101 unjoint_samples 19100 joint_samples 57 [1046770, 300516] processed_samples 19101 unjoint_samples 19100 joint_samples 58 [1041822, 738019] processed_samples 19100 unjoint_samples 19100 joint_samples 58 [1043877, 403731] processed_samples 19101 unjoint_samples 19100 joint_samples 58 [1041822, 738019] processed_samples 19101 unjoint_samples 19100 joint_samples 58 [1045028, 1046154] [h264 @ 0x559618059540] mmco: unref short failure processed_samples 19101 unjoint_samples 19100 joint_samples 58 [1048140, 423568] processed_samples 19101 unjoint_samples 19100 joint_samples 57 [1046770, 300516] processed_samples 19102 unjoint_samples 19100 joint_samples 58 [1047707, 403813] [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure processed_samples 19101 unjoint_samples 19100 joint_samples 58 [1045028, 1046154] [h264 @ 0x5617e8e19000] mmco: unref short failure processed_samples 19102 unjoint_samples 19100 joint_samples 58 [1047707, 403813] [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e8e0d640] mmco: unref short failure [h264 @ 0x5617e8e0d640] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure processed_samples 19200 unjoint_samples 19200 joint_samples 58 [36813, 1046434] processed_samples 19200 unjoint_samples 19200 joint_samples 58 [36813, 1046434] processed_samples 19200 unjoint_samples 19200 joint_samples 58 [1043877, 702219] processed_samples 19200 unjoint_samples 19200 joint_samples 58 [1043877, 702219] processed_samples 19201 unjoint_samples 19200 joint_samples 58 [1048140, 831663] processed_samples 19201 unjoint_samples 19200 joint_samples 58 [1048140, 831663] processed_samples 19201 unjoint_samples 19200 joint_samples 57 [1046770, 555869] processed_samples 19201 unjoint_samples 19200 joint_samples 57 [1046770, 555869] processed_samples 19201 unjoint_samples 19200 joint_samples 59 [1045028, 228162] processed_samples 19201 unjoint_samples 19200 joint_samples 59 [1045028, 228162] processed_samples 19200 unjoint_samples 19200 joint_samples 57 [640429, 1007324] processed_samples 19200 unjoint_samples 19200 joint_samples 57 [640429, 1007324] [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure processed_samples 19202 unjoint_samples 19200 joint_samples 58 [1047707, 732216] processed_samples 19201 unjoint_samples 19200 joint_samples 58 [1041822, 1000341] processed_samples 19201 unjoint_samples 19200 joint_samples 58 [1041822, 1000341] [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure processed_samples 19202 unjoint_samples 19200 joint_samples 58 [1047707, 732216] [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e9262740] mmco: unref short failure [h264 @ 0x5617e9262740] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559617afc780] mmco: unref short failure [h264 @ 0x559617afc780] mmco: unref short failure [h264 @ 0x5617e803bd40] mmco: unref short failure [h264 @ 0x5617e803bd40] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure processed_samples 19300 unjoint_samples 19300 joint_samples 58 [305717, 1046434] processed_samples 19300 unjoint_samples 19300 joint_samples 57 [874887, 1007324] [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure processed_samples 19301 unjoint_samples 19300 joint_samples 59 [226818, 985579] processed_samples 19301 unjoint_samples 19300 joint_samples 59 [1045331, 314194] [h264 @ 0x5617e9493900] mmco: unref short failure processed_samples 19301 unjoint_samples 19300 joint_samples 57 [1046770, 828561] processed_samples 19301 unjoint_samples 19300 joint_samples 59 [1045028, 530692] [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure processed_samples 19300 unjoint_samples 19300 joint_samples 58 [1043877, 987680] [h264 @ 0x55961563b940] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure processed_samples 19301 unjoint_samples 19300 joint_samples 57 [1046770, 828561] processed_samples 19302 unjoint_samples 19300 joint_samples 58 [1047707, 1001728] processed_samples 19300 unjoint_samples 19300 joint_samples 58 [305717, 1046434] processed_samples 19301 unjoint_samples 19300 joint_samples 59 [226818, 985579] processed_samples 19301 unjoint_samples 19300 joint_samples 59 [1045331, 314194] processed_samples 19300 unjoint_samples 19300 joint_samples 57 [874887, 1007324] [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure processed_samples 19301 unjoint_samples 19300 joint_samples 59 [1045028, 530692] processed_samples 19300 unjoint_samples 19300 joint_samples 58 [1043877, 987680] processed_samples 19302 unjoint_samples 19300 joint_samples 58 [1047707, 1001728] [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5596186f0980] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x5617ebbed500] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5596177086c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5596141221c0] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x55961adc42c0] mmco: unref short failure processed_samples 19400 unjoint_samples 19400 joint_samples 58 [23056, 1047360] processed_samples 19400 unjoint_samples 19400 joint_samples 58 [23056, 1047360] processed_samples 19400 unjoint_samples 19400 joint_samples 59 [1046296, 162383] processed_samples 19400 unjoint_samples 19400 joint_samples 59 [1046296, 162383] processed_samples 19401 unjoint_samples 19400 joint_samples 59 [530235, 985579] processed_samples 19401 unjoint_samples 19400 joint_samples 58 [171831, 1040073] processed_samples 19401 unjoint_samples 19400 joint_samples 59 [530235, 985579] processed_samples 19401 unjoint_samples 19400 joint_samples 59 [1045028, 920692] processed_samples 19400 unjoint_samples 19400 joint_samples 58 [634176, 1046434] processed_samples 19400 unjoint_samples 19400 joint_samples 58 [634176, 1046434] processed_samples 19402 unjoint_samples 19400 joint_samples 59 [287775, 1037213] processed_samples 19401 unjoint_samples 19400 joint_samples 59 [1045331, 646605] processed_samples 19401 unjoint_samples 19400 joint_samples 58 [171831, 1040073] processed_samples 19401 unjoint_samples 19400 joint_samples 59 [1045331, 646605] processed_samples 19401 unjoint_samples 19400 joint_samples 59 [1045028, 920692] processed_samples 19402 unjoint_samples 19400 joint_samples 59 [287775, 1037213] [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x559617361700] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure processed_samples 19500 unjoint_samples 19500 joint_samples 58 [540790, 1047360] processed_samples 19500 unjoint_samples 19500 joint_samples 58 [540790, 1047360] [h264 @ 0x5617e8298040] mmco: unref short failure [h264 @ 0x5617e8298040] mmco: unref short failure [h264 @ 0x559616fdd040] mmco: unref short failure [h264 @ 0x559616fdd040] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure processed_samples 19501 unjoint_samples 19500 joint_samples 59 [760212, 985579] processed_samples 19500 unjoint_samples 19500 joint_samples 59 [1046296, 586179] processed_samples 19501 unjoint_samples 19500 joint_samples 59 [760212, 985579] processed_samples 19500 unjoint_samples 19500 joint_samples 59 [1046296, 586179] processed_samples 19501 unjoint_samples 19500 joint_samples 58 [478213, 1040073] processed_samples 19501 unjoint_samples 19500 joint_samples 60 [108338, 1046902] processed_samples 19501 unjoint_samples 19500 joint_samples 58 [478213, 1040073] processed_samples 19501 unjoint_samples 19500 joint_samples 60 [108338, 1046902] processed_samples 19501 unjoint_samples 19500 joint_samples 60 [119090, 984007] processed_samples 19501 unjoint_samples 19500 joint_samples 60 [119090, 984007] processed_samples 19500 unjoint_samples 19500 joint_samples 58 [968038, 1046434] processed_samples 19500 unjoint_samples 19500 joint_samples 58 [968038, 1046434] processed_samples 19502 unjoint_samples 19500 joint_samples 59 [649086, 1037213] processed_samples 19502 unjoint_samples 19500 joint_samples 59 [649086, 1037213] [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559616fdd040] mmco: unref short failure [h264 @ 0x5617e8298040] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure [h264 @ 0x5617e8ce2d40] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617e8e0d640] mmco: unref short failure [h264 @ 0x55961d3b8040] [h264 @ 0x5617e8f78980] mmco: unref short failure mmco: unref short failure [h264 @ 0x55961d3b8040] [h264 @ 0x5617e8f78980] mmco: unref short failure mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure processed_samples 19600 unjoint_samples 19600 joint_samples 59 [1037992, 307482] processed_samples 19600 unjoint_samples 19600 joint_samples 59 [1037992, 307482] processed_samples 19600 unjoint_samples 19600 joint_samples 58 [852221, 1047360] processed_samples 19600 unjoint_samples 19600 joint_samples 58 [852221, 1047360] processed_samples 19601 unjoint_samples 19600 joint_samples 60 [33823, 1039783] processed_samples 19601 unjoint_samples 19600 joint_samples 60 [33823, 1039783] processed_samples 19601 unjoint_samples 19600 joint_samples 60 [412442, 1046902] processed_samples 19601 unjoint_samples 19600 joint_samples 60 [412442, 1046902] processed_samples 19601 unjoint_samples 19600 joint_samples 60 [412175, 984007] processed_samples 19601 unjoint_samples 19600 joint_samples 60 [412175, 984007] processed_samples 19601 unjoint_samples 19600 joint_samples 58 [854412, 1040073] processed_samples 19601 unjoint_samples 19600 joint_samples 58 [854412, 1040073] [h264 @ 0x5596172068c0] mmco: unref short failure processed_samples 19600 unjoint_samples 19600 joint_samples 59 [1046296, 902036] [h264 @ 0x5617ebbed500] mmco: unref short failure processed_samples 19602 unjoint_samples 19600 joint_samples 59 [989226, 1037213] processed_samples 19600 unjoint_samples 19600 joint_samples 59 [1046296, 902036] processed_samples 19602 unjoint_samples 19600 joint_samples 59 [989226, 1037213] [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x559619420380] mmco: unref short failure [h264 @ 0x559619420380] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure processed_samples 19700 unjoint_samples 19700 joint_samples 59 [1041358, 224317] processed_samples 19700 unjoint_samples 19700 joint_samples 59 [1041358, 224317] processed_samples 19701 unjoint_samples 19700 joint_samples 60 [398673, 1039783] processed_samples 19701 unjoint_samples 19700 joint_samples 60 [398673, 1039783] processed_samples 19700 unjoint_samples 19700 joint_samples 59 [1037992, 664777] processed_samples 19700 unjoint_samples 19700 joint_samples 60 [247011, 1044578] processed_samples 19700 unjoint_samples 19700 joint_samples 60 [247011, 1044578] processed_samples 19701 unjoint_samples 19700 joint_samples 59 [1032832, 154133] processed_samples 19701 unjoint_samples 19700 joint_samples 59 [1032832, 154133] processed_samples 19700 unjoint_samples 19700 joint_samples 59 [1037992, 664777] processed_samples 19702 unjoint_samples 19700 joint_samples 60 [210414, 1047464] processed_samples 19702 unjoint_samples 19700 joint_samples 60 [210414, 1047464] processed_samples 19701 unjoint_samples 19700 joint_samples 60 [688449, 984007] processed_samples 19701 unjoint_samples 19700 joint_samples 60 [688449, 984007] processed_samples 19701 unjoint_samples 19700 joint_samples 60 [716120, 1046902] processed_samples 19701 unjoint_samples 19700 joint_samples 60 [716120, 1046902] [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x559618395d00] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559616fe4f40] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617efb834c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure processed_samples 19800 unjoint_samples 19800 joint_samples 59 [1037992, 1007121] [h264 @ 0x5596156ec940] mmco: unref short failure processed_samples 19800 unjoint_samples 19800 joint_samples 59 [1041358, 682090] processed_samples 19800 unjoint_samples 19800 joint_samples 60 [559708, 1044578] processed_samples 19801 unjoint_samples 19800 joint_samples 60 [685825, 1039783] processed_samples 19801 unjoint_samples 19800 joint_samples 61 [1038396, 81983] processed_samples 19801 unjoint_samples 19800 joint_samples 59 [1032832, 510582] processed_samples 19802 unjoint_samples 19800 joint_samples 60 [513610, 1047464] processed_samples 19801 unjoint_samples 19800 joint_samples 60 [937841, 984007] processed_samples 19800 unjoint_samples 19800 joint_samples 59 [1037992, 1007121] [h264 @ 0x5617ec342580] mmco: unref short failure processed_samples 19800 unjoint_samples 19800 joint_samples 59 [1041358, 682090] [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure processed_samples 19800 unjoint_samples 19800 joint_samples 60 [559708, 1044578] processed_samples 19801 unjoint_samples 19800 joint_samples 60 [685825, 1039783] processed_samples 19801 unjoint_samples 19800 joint_samples 61 [1038396, 81983] processed_samples 19801 unjoint_samples 19800 joint_samples 59 [1032832, 510582] [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure processed_samples 19802 unjoint_samples 19800 joint_samples 60 [513610, 1047464] processed_samples 19801 unjoint_samples 19800 joint_samples 60 [937841, 984007] [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x55961361a140] mmco: unref short failure [h264 @ 0x559616e6e540] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x55961adc42c0] mmco: unref short failure [h264 @ 0x5617e7fbb200] mmco: unref short failure [h264 @ 0x5617eb0faac0] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617ee580e80] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559613520f80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure processed_samples 19900 unjoint_samples 19900 joint_samples 59 [1041358, 992713] processed_samples 19900 unjoint_samples 19900 joint_samples 60 [1039951, 329295] processed_samples 19901 unjoint_samples 19900 joint_samples 61 [94321, 1040201] processed_samples 19901 unjoint_samples 19900 joint_samples 61 [144304, 1046676] processed_samples 19901 unjoint_samples 19900 joint_samples 61 [1038396, 362589] processed_samples 19900 unjoint_samples 19900 joint_samples 60 [997215, 1044578] [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5595f1f30fc0] mmco: unref short failure [h264 @ 0x5617e8b75440] mmco: unref short failure processed_samples 19902 unjoint_samples 19900 joint_samples 60 [759446, 1047464] [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure processed_samples 19901 unjoint_samples 19900 joint_samples 59 [1032832, 818269] processed_samples 19900 unjoint_samples 19900 joint_samples 59 [1041358, 992713] processed_samples 19900 unjoint_samples 19900 joint_samples 60 [1039951, 329295] processed_samples 19901 unjoint_samples 19900 joint_samples 61 [94321, 1040201] processed_samples 19901 unjoint_samples 19900 joint_samples 61 [144304, 1046676] processed_samples 19901 unjoint_samples 19900 joint_samples 61 [1038396, 362589] processed_samples 19900 unjoint_samples 19900 joint_samples 60 [997215, 1044578] [h264 @ 0x5617e9157440] mmco: unref short failure [h264 @ 0x5617e9157440] mmco: unref short failure processed_samples 19902 unjoint_samples 19900 joint_samples 60 [759446, 1047464] [h264 @ 0x5617ec447000] mmco: unref short failure processed_samples 19901 unjoint_samples 19900 joint_samples 59 [1032832, 818269] [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5596180e8840] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x5617eba55600] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5617e8c6f380] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x5617e9531f40] mmco: unref short failure [h264 @ 0x559617e83980] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ed0d5800] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617eccc9b40] mmco: unref short failure processed_samples 20000 unjoint_samples 20000 joint_samples 60 [1039951, 672828] [h264 @ 0x559614220440] mmco: unref short failure processed_samples 20001 unjoint_samples 20000 joint_samples 61 [398494, 1046676] processed_samples 20000 unjoint_samples 20000 joint_samples 60 [327738, 1028062] processed_samples 20000 unjoint_samples 20000 joint_samples 61 [1045568, 310744] processed_samples 20001 unjoint_samples 20000 joint_samples 61 [367770, 1040201] processed_samples 20001 unjoint_samples 20000 joint_samples 60 [1047800, 121080] processed_samples 20001 unjoint_samples 20000 joint_samples 61 [1038396, 745269] processed_samples 20000 unjoint_samples 20000 joint_samples 60 [1039951, 672828] [h264 @ 0x5617ec5992c0] mmco: unref short failure processed_samples 20001 unjoint_samples 20000 joint_samples 61 [398494, 1046676] processed_samples 20001 unjoint_samples 20000 joint_samples 61 [367770, 1040201] processed_samples 20000 unjoint_samples 20000 joint_samples 60 [327738, 1028062] processed_samples 20000 unjoint_samples 20000 joint_samples 61 [1045568, 310744] processed_samples 20001 unjoint_samples 20000 joint_samples 60 [1047800, 121080] processed_samples 20002 unjoint_samples 20000 joint_samples 60 [956974, 1047464] processed_samples 20001 unjoint_samples 20000 joint_samples 61 [1038396, 745269] [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure processed_samples 20002 unjoint_samples 20000 joint_samples 60 [956974, 1047464] [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x5617e8087580] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure processed_samples 20101 unjoint_samples 20100 joint_samples 61 [677278, 1040201] processed_samples 20100 unjoint_samples 20100 joint_samples 60 [1039951, 1025781] processed_samples 20101 unjoint_samples 20100 joint_samples 60 [1047800, 378053] processed_samples 20100 unjoint_samples 20100 joint_samples 60 [734390, 1028062] processed_samples 20101 unjoint_samples 20100 joint_samples 61 [724018, 1046676] processed_samples 20100 unjoint_samples 20100 joint_samples 61 [1045568, 658825] processed_samples 20101 unjoint_samples 20100 joint_samples 61 [1038396, 1016192] processed_samples 20102 unjoint_samples 20100 joint_samples 61 [1037508, 291286] [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure processed_samples 20100 unjoint_samples 20100 joint_samples 60 [734390, 1028062] processed_samples 20101 unjoint_samples 20100 joint_samples 61 [677278, 1040201] processed_samples 20101 unjoint_samples 20100 joint_samples 60 [1047800, 378053] processed_samples 20101 unjoint_samples 20100 joint_samples 61 [1038396, 1016192] processed_samples 20100 unjoint_samples 20100 joint_samples 60 [1039951, 1025781] processed_samples 20102 unjoint_samples 20100 joint_samples 61 [1037508, 291286] processed_samples 20100 unjoint_samples 20100 joint_samples 61 [1045568, 658825] processed_samples 20101 unjoint_samples 20100 joint_samples 61 [724018, 1046676] [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x559616fb2ec0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ebfdebc0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x559616c4eb00] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617e84806c0] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617eee843c0] mmco: unref short failure [h264 @ 0x55961342b000] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x559618461980] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x55961c6febc0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x559613a0c140] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5617e9b3b480] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure processed_samples 20200 unjoint_samples 20200 joint_samples 61 [950189, 210565] [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure processed_samples 20200 unjoint_samples 20200 joint_samples 61 [313789, 1027880] processed_samples 20200 unjoint_samples 20200 joint_samples 61 [1045568, 994246] processed_samples 20201 unjoint_samples 20200 joint_samples 62 [1043385, 321013] processed_samples 20201 unjoint_samples 20200 joint_samples 61 [1017726, 1046676] processed_samples 20201 unjoint_samples 20200 joint_samples 60 [1047800, 720092] processed_samples 20200 unjoint_samples 20200 joint_samples 61 [950189, 210565] processed_samples 20202 unjoint_samples 20200 joint_samples 61 [1037508, 655150] [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure processed_samples 20201 unjoint_samples 20200 joint_samples 61 [996599, 1040201] [h264 @ 0x559618693e80] mmco: unref short failure [h264 @ 0x559618693e80] mmco: unref short failure processed_samples 20200 unjoint_samples 20200 joint_samples 61 [313789, 1027880] [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure processed_samples 20201 unjoint_samples 20200 joint_samples 62 [1043385, 321013] processed_samples 20200 unjoint_samples 20200 joint_samples 61 [1045568, 994246] processed_samples 20202 unjoint_samples 20200 joint_samples 61 [1037508, 655150] [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure processed_samples 20201 unjoint_samples 20200 joint_samples 60 [1047800, 720092] processed_samples 20201 unjoint_samples 20200 joint_samples 61 [996599, 1040201] processed_samples 20201 unjoint_samples 20200 joint_samples 61 [1017726, 1046676] [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617ec447000] mmco: unref short failure [h264 @ 0x5617f0b8e140] mmco: unref short failure [h264 @ 0x5617ebd1da00] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ec251b40] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x559616d0c780] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961adc42c0] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x559618471cc0] mmco: unref short failure [h264 @ 0x559618471cc0] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5617ec31cc40] mmco: unref short failure [h264 @ 0x5596134bb5c0] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x559617212e80] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x559617100c80] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x559614221480] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ebab32c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5617ea1fdc80] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559613eda640] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure processed_samples 20300 unjoint_samples 20300 joint_samples 61 [950189, 556171] processed_samples 20300 unjoint_samples 20300 joint_samples 62 [1047837, 180425] processed_samples 20300 unjoint_samples 20300 joint_samples 61 [542148, 1027880] processed_samples 20301 unjoint_samples 20300 joint_samples 62 [1043385, 625549] processed_samples 20301 unjoint_samples 20300 joint_samples 62 [226869, 1046676] processed_samples 20301 unjoint_samples 20300 joint_samples 62 [261397, 1047824] processed_samples 20301 unjoint_samples 20300 joint_samples 60 [1047800, 964274] [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure [h264 @ 0x55961acdfdc0] mmco: unref short failure processed_samples 20300 unjoint_samples 20300 joint_samples 62 [1047837, 180425] processed_samples 20300 unjoint_samples 20300 joint_samples 61 [950189, 556171] processed_samples 20300 unjoint_samples 20300 joint_samples 61 [542148, 1027880] processed_samples 20301 unjoint_samples 20300 joint_samples 62 [226869, 1046676] processed_samples 20301 unjoint_samples 20300 joint_samples 62 [261397, 1047824] processed_samples 20301 unjoint_samples 20300 joint_samples 60 [1047800, 964274] processed_samples 20301 unjoint_samples 20300 joint_samples 62 [1043385, 625549] [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure processed_samples 20302 unjoint_samples 20300 joint_samples 61 [1037508, 982054] [h264 @ 0x5617ec3749c0] mmco: unref short failure [h264 @ 0x5617ec3749c0] mmco: unref short failure processed_samples 20302 unjoint_samples 20300 joint_samples 61 [1037508, 982054] [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x559617cf4f80] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x5617e9493900] mmco: unref short failure [h264 @ 0x559616fdd040] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5617e803bd40] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x559618059540] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x5617ec417900] mmco: unref short failure [h264 @ 0x55961845fd00] mmco: unref short failure [h264 @ 0x5617ec27fb00] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x5617e829eac0] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x559617ba6840] mmco: unref short failure [h264 @ 0x5617e803bd40] mmco: unref short failure [h264 @ 0x559615fe6e40] mmco: unref short failure FileNotFoundError datasets/LMM/lmms-lab/LLaVA-Video-178K/liwei_youtube_videos/videos/youtube_video_2024/ytb_GqeRnxSuLFI.mp4 FileNotFoundError datasets/LMM/lmms-lab/LLaVA-Video-178K/liwei_youtube_videos/videos/youtube_video_2024/ytb_GqeRnxSuLFI.mp4 [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x559618131fc0] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x55961c1ba780] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617eeca6a40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617f06c03c0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x559616359900] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5617e7f6a140] mmco: unref short failure [h264 @ 0x5596139c8400] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ee5b7480] mmco: unref short failure [h264 @ 0x559616baef00] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure processed_samples 20400 unjoint_samples 20400 joint_samples 61 [827729, 1027880] processed_samples 20400 unjoint_samples 20400 joint_samples 62 [1047837, 524364] processed_samples 20401 unjoint_samples 20400 joint_samples 62 [1043385, 897576] processed_samples 20401 unjoint_samples 20400 joint_samples 61 [355281, 1033836] processed_samples 20401 unjoint_samples 20400 joint_samples 62 [600883, 1047824] processed_samples 20401 unjoint_samples 20400 joint_samples 62 [518512, 1046676] [h264 @ 0x5596196a8b80] mmco: unref short failure [h264 @ 0x5596196a8b80] mmco: unref short failure processed_samples 20402 unjoint_samples 20400 joint_samples 62 [1046883, 179112] processed_samples 20401 unjoint_samples 20400 joint_samples 61 [957049, 956655] processed_samples 20400 unjoint_samples 20400 joint_samples 62 [1047837, 524364] processed_samples 20401 unjoint_samples 20400 joint_samples 62 [518512, 1046676] processed_samples 20401 unjoint_samples 20400 joint_samples 61 [957049, 956655] [h264 @ 0x559615fe6e40] mmco: unref short failure processed_samples 20401 unjoint_samples 20400 joint_samples 62 [1043385, 897576] processed_samples 20401 unjoint_samples 20400 joint_samples 61 [355281, 1033836] processed_samples 20401 unjoint_samples 20400 joint_samples 62 [600883, 1047824] processed_samples 20400 unjoint_samples 20400 joint_samples 61 [827729, 1027880] [h264 @ 0x559616fdd040] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure processed_samples 20402 unjoint_samples 20400 joint_samples 62 [1046883, 179112] [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5617e90f0680] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e9213b80] mmco: unref short failure [h264 @ 0x5596137952c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x55961741a940] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x5617e9100f80] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x5617e7d0f940] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x559616fe59c0] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x5617ebb77d00] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617e7c5b300] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5617e8713340] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596194f09c0] mmco: unref short failure [h264 @ 0x5596170d5600] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x559613728d00] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x55961864b640] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x5617e8c73f00] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x55961609a680] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure processed_samples 20500 unjoint_samples 20500 joint_samples 62 [1047605, 23832] processed_samples 20500 unjoint_samples 20500 joint_samples 62 [1047605, 23832] processed_samples 20501 unjoint_samples 20500 joint_samples 62 [1040996, 120038] processed_samples 20501 unjoint_samples 20500 joint_samples 62 [1040996, 120038] processed_samples 20501 unjoint_samples 20500 joint_samples 62 [766692, 1046676] processed_samples 20500 unjoint_samples 20500 joint_samples 62 [1047837, 836423] processed_samples 20501 unjoint_samples 20500 joint_samples 61 [668258, 1033836] processed_samples 20501 unjoint_samples 20500 joint_samples 61 [668258, 1033836] processed_samples 20501 unjoint_samples 20500 joint_samples 62 [766692, 1046676] processed_samples 20500 unjoint_samples 20500 joint_samples 62 [1047837, 836423] processed_samples 20501 unjoint_samples 20500 joint_samples 63 [1045666, 134221] processed_samples 20501 unjoint_samples 20500 joint_samples 63 [1045666, 134221] processed_samples 20502 unjoint_samples 20500 joint_samples 62 [1046883, 433773] processed_samples 20502 unjoint_samples 20500 joint_samples 62 [1046883, 433773] processed_samples 20501 unjoint_samples 20500 joint_samples 62 [853367, 1047824] processed_samples 20501 unjoint_samples 20500 joint_samples 62 [853367, 1047824] [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5617e9274140] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x5596172068c0] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x5617e8c99800] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x559617077b40] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x5617e87a0740] mmco: unref short failure [h264 @ 0x55961b920980] mmco: unref short failure [h264 @ 0x5617e9426340] mmco: unref short failure [h264 @ 0x5596161e2ac0] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x559615069280] mmco: unref short failure [h264 @ 0x5617ebf51f00] mmco: unref short failure [h264 @ 0x5596183f9540] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617ea4c1040] mmco: unref short failure [h264 @ 0x5617e7db17c0] mmco: unref short failure [h264 @ 0x5596152d1b80] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x559618127c00] mmco: unref short failure [h264 @ 0x5617e8001bc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x5617e955dbc0] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5596170a2940] mmco: unref short failure [h264 @ 0x5617e8f78980] mmco: unref short failure [h264 @ 0x5596156ec940] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e803bd40] mmco: unref short failure [h264 @ 0x5617e803bd40] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x55961d500c40] mmco: unref short failure [h264 @ 0x5617ec0115c0] mmco: unref short failure [h264 @ 0x5617ee8ce7c0] mmco: unref short failure [h264 @ 0x5596183cd8c0] mmco: unref short failure [h264 @ 0x559614220440] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5617ebde1580] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5596136b5540] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5617e8742140] mmco: unref short failure [h264 @ 0x5596135e0fc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ec5992c0] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961602ec80] mmco: unref short failure [h264 @ 0x55961609e040] mmco: unref short failure [h264 @ 0x55961d3b8040] mmco: unref short failure [h264 @ 0x559615eea1c0] mmco: unref short failure [h264 @ 0x5617e9256180] mmco: unref short failure [h264 @ 0x5617e8995380] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure [h264 @ 0x5617ec342580] mmco: unref short failure processed_samples 20600 unjoint_samples 20600 joint_samples 62 [1047605, 376022] processed_samples 20600 unjoint_samples 20600 joint_samples 63 [52047, 1038519] processed_samples 20601 unjoint_samples 20600 joint_samples 62 [1040996, 451524] [h264 @ 0x55961672f480] mmco: unref short failure [h264 @ 0x55961672f480] mmco: unref short failure processed_samples 20600 unjoint_samples 20600 joint_samples 63 [52047, 1038519] processed_samples 20600 unjoint_samples 20600 joint_samples 62 [1047605, 376022] processed_samples 20601 unjoint_samples 20600 joint_samples 62 [1040996, 451524] processed_samples 20601 unjoint_samples 20600 joint_samples 63 [6394, 1047281] processed_samples 20601 unjoint_samples 20600 joint_samples 63 [6394, 1047281] processed_samples 20601 unjoint_samples 20600 joint_samples 63 [1027021, 244279] processed_samples 20601 unjoint_samples 20600 joint_samples 63 [1027021, 244279] processed_samples 20601 unjoint_samples 20600 joint_samples 63 [1045666, 476811] processed_samples 20601 unjoint_samples 20600 joint_samples 63 [1045666, 476811] processed_samples 20602 unjoint_samples 20600 joint_samples 62 [1046883, 765766] processed_samples 20602 unjoint_samples 20600 joint_samples 62 [1046883, 765766] processed_samples 20601 unjoint_samples 20600 joint_samples 61 [1040562, 1040039] processed_samples 20601 unjoint_samples 20600 joint_samples 61 [1040562, 1040039] [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x5617ebf1ebc0] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x55961d016580] mmco: unref short failure [h264 @ 0x5617e9f7e780] mmco: unref short failure [h264 @ 0x559619462b80] mmco: unref short failure [2024-12-01 22:44:35,935] torch.distributed.elastic.agent.server.api: [ERROR] Error waiting on exit barrier. Elapsed: 300.04437160491943 seconds + set +x